Find Tutor Online
Certi cate in Quantitative Finance
Final Project Brief
JUNE 2017 COHORT
This document outlines each available topic together with submission requirements for the project report. By-step instructions o er a structure not a limit to what you can implement. A separate source, Q&A on Robust Modelling discusses the most frequent issues and relevant detail of numerical methods.
This Brief, Project Workshops and Q&A are your primary guidance. Please make sure you have reviewed the workshops and topical lectures relevant to your chosen topic.
| |Portfolio Construction, Time Series, HJM/LMM/SABR, Copula Method |
| | |
| |Credit Analytics, CVA |
| | |
| |Local Volatility, PDEs, One-factor rate models |
| | |
| |Table 1: Support Allocation |
1
1 Instructions
To complete the project, you must implement one topic, plus CVA component, as given and described in this current Brief. Each topic covers speci c pricing or allocation model, but numerical techniques can vary and include advances from Time Series Analysis, Data Analytics and Machine Learning. Final project mark is entered as Module Six mark.
All. CVA Calculation for Interest Rate Swap. This is a mandatory addition as it balances the exposure to the quant modelling in rates and credit. Can be implemented on an Excel spreadsheet (or CQF example spreadsheets reused).
1. Credit Portfolio Fair Spread and Sensitivity Analysis
2. Portfolio Construction with Robust Volatility
3. Data Analytics: Asset Allocation with Views Features (advanced)
4. Arbitrage Trading on Cointegration with Backtest
5. Local Volatility in Interest Rates
6. LIBOR and OIS Rates: Market Volatility (advanced)
Each of the topics is more speci c than elective because the topic focuses on a speci c pricing model (for derivatives), speci c portfolio construction approach (Black-Litterman), or speci c time series analysis (cointegration). Electives are usually broader and cover the subject area. Advanced topics are recommended to delegates with experience in respective areas.
1.1 Submission Requirements
Submit working code together with a well-written report and originality declaration.
Project report to have an exact topic title and content must correspond to it. Recom-mended length is 30-50 pages, excluding code.
CVA component would be a chapter in your report, 4-5 pages or more depending on your venturing to interest rate modelling and analysis of output.
Submissions to be uploaded to online portal only. Upload format: one written report (PDF), one zip archive with code and data les, and one scanned declaration (PDF).
Submission date is Monday, 8 January 2018, 23:59 GMT
Submissions must match the Brief. There is no extension to the Final Project.
2
1.2 CQF Electives
We ask you to indicate the choice of two Electives in order to preserve focus { content for all electives will be available later for your advancement. Electives are not a condition to Final Project. You will be using one/a few techniques (but not all) covered in electives on Data Analytics, Python Applications, Computational Methods for Black-Scholes Pricing, and Counterparty Credit Risk.
Risk Management and Machine Learning with Python are recommended as stand-alone intro-ductions to these professional areas.
|Algorithmic Trading |An example of strategy tting the reversion to the Log-Periodic Power |
| |Law (LPPL), a kind of statistical arbitrage. Discusses optimisation in |
| |context of trading strategies (Slide 51 onwards). Multicharts software. |
| |Notes on trader's psychology. Outcome: relevant to Arbitrage Trading |
| | |
| | |
| |of SDEs. Dynamic programming solves optimisation over a stochastic |
| |variable (eg, asset price, interest rate, risk premium). Outcome: this |
| | |
| |optimal allocations by Merton index of satisfaction, Kelly criterion and |
| |venture to Risk-Sensitive Asset Management. |
| | | | | | | | |
|Behavioural Finance |Heuristics, Biases and Framing. Excursion to game theory and proba- |
| |bility to formalise the psychology of choice. Outcome: the elective is |
| | |
| |Data Analytics topic. |
| | | | | | |
|Risk Management |The base you need to be a risk manager (read Coleman guide) and |
| |tools for Basel regulation: (a) weakness of risk-weighted assets, (b) |
| |extreme value theory for ES and capital requirement and (c) adjoint |
| |automatic di erentiation to compute formal sensitivities (derivatives). |
| |Other numerical methods covered the same as Counterparty Risk Elec- |
| |tive. Outcome: this elective is best taken for your own advancement. |
| | | | | | |
|Counterparty Credit |CDS, survival probabilities and hazard rates reviewed. Three key nu- |
|Risk |merical methods for quant nance pricing (Monte-Carlo, Binomial Trees, |
| |Finite Di erence). Monte Carlo for simple LMM. Review of Module Five |
| |on Credit with a touch on the copula method. Outcome: covers CVA |
| | | | |
| |Computation in simple and great detail, in-depth review to prepare for |
| |Credit Spread topic if necessary. |
| | | | |
|Modeling | | |sion) and their analytical solution: the main approach to solve stochastic |
| | | |volatility (Heston model) is via Fourier Transform. In-depth on integra- |
| | | |tion Outcome: Local Volatility topic o ers a classic pricing PDE, which |
| | | | |
| | | |
|ods | | |tion, root nding (Bisection, Newton), polynomial interpolation, and |
| | | |numerical integration (trapezium and Simpson rules), vector and ma- |
| | | |trix norms. Outcome: a refresher of methods that support modelling |
| | | | |
| | | | | | | | |
|Python Applications | |Reviews quant numerical techniques, now with Python examples |
| | | |and notes on computational e ciency: Normal from Uniform RN, |
| | | |linear equations and eigenvalues, numerical integration, root nd- |
| | | |ing (Bisection, Newton), random numbers with arrays and seeding, |
| | | |Binomial/Poisson/LogNormal distributions, SDE simulation (GBM, |
| | | |OU cases). Introduction to Jupyter Notebook, arrays and indexes. |
| | | |Outcome: relevant particularly for Monte-Carlo in Credit Spread and |
| | | | |
| | | | | | | |
|Data Analytics | |Level I on Python for quant nance, data structures (Dataframe), |
| | | |NumPy for Numerical Analysis, Pandas for Financial Time Series Anal- |
| | | |ysis, data visualization. Outcome: likely for Arbitrage Trading and |
| | | |Portfolio Construction | | |
| | | | | | |
|Machine |Learning | |Level II on Python for quant nance, capabilities of scikit-learn libary |
|with Python | | |(OLS, logistic), tensor ow library example of 'deep learning' (classi er) |
| | | |{ use the Github link below and explore. Outcome: Arbitrage Trading, |
| | | | | | |
| | | |returns modelling for allocations backtesting (extra to Portfolio Con- |
| | | |struction), and Least Squares Monte Carlo for pricing Bermudans in |
| | | |LMM framework. |
| | | | | | | | | | | | | |
Table 3: Mapping Electives vs. Final Project Topics (cont.)
Machine Learning with Python has computational material on Github
PyAlgoTrade documentation
4
1.3 Coding for Quant Finance
Programming environment must have appropriate strengths and facilities to implement the topic (pricing model) chosen. Common choices range from VBA to Python to C++, please exercise judgement as quants.
Use of R/Matlab/Mathematica/Matlab is encouraged where time series or presentation involved. CQF-supplied Excel spreadsheets can be used as a starting point and to validate results but coding of numerical techniques/use of industry code libraries is expected.
`Scripted solution' means the ready functionality from toolboxes and libraries is called, but the amount of own coding of numerical methods is minimal or non-existent. This particularly applies to Matlab/R as well as Excel spreadsheet functions (not robust).
The aim of the project is to enable you to code numerical methods and develop model pro-totypes in a production environment. Spreadsheets-only or scripted solutions are below the expected standard for completion of the project.
To answer the question, \What should I code?" Delegates are expected to re-code numer-ical methods that are central to the model and exercise judgement in identifying them. Balanced use of libraries is allowed at the delegate's own discretion and subject to a de-scription of limitations for ready functions/borrowed code (in the report).
It is up to delegates to develop their own test cases, sensibility checks and validation. It is normal to observe irregularities when the model is implemented on real life data. If in doubt, re ect on the issue in the project report.
The code must be thoroughly tested and well-documented: each function must be de-scribed, and comments must be used. Provide instructions on how to run the code.
1.4 Project Report
The main purpose of the report is to facilitate access to numerical methods' implementation (the code) and pricing results.
The report must contain a su cient description of the mathematical model, numerical methods and their properties. In-depth study is welcome but must be relevant.
If you coded a numerical method please feature that in the report/appendix.
Please give due attention and space for presentation and discussion of your pricing results. Have an explicit sensitivity/risk analysis.
Use charts, test cases and comparison to research results where available.
Mathematical sections of the report can be prepared using LaTeX or Equation Editor (Word). For Mathematica and Python notebooks, make sure they are presentable.
5
This page intentionally left blank.
6
CVA Calculation for an Interest Rate Swap
Summary
To recognise the importance of credit and counterparty risk adjustments to the derivatives business we introduce this mandatory component which must be implemented with each topic. One-o spreadsheet CVA computation is acceptable, but better quality work will have MtM out-put from Monte-Carlo.
Calculate the credit valuation adjustment, taken by Counterparty A, for an interest rate swap instrument using credit spreads for Counterparty B. Plot MtM values and produce (a) Expected Exposure pro le. While EE is de ned as max (MtM ; 0)+, simulated curves allow presenting exposure distribution at each tenor and computing Potential Future Exposure at (b) the median of positive exposure and (c) 97:5th percentile. Use the output of HJM/LMM models, take a ready implementation preferably calibrated to the recent data.
Provide a brief discussion of your observations, e.g., exposure over time, location of maximum exposure, impact of very small or negative rates. The advanced sensitivity analysis will illustrate the concept of the wrong-way risk.
Step-By-Step
The inputs for IRS valuation are Forward LIBORs and discounting factors. CVA requires default probabilities: bootstrap or make reasonable assumptions to supplement the data (e.g., at credit spreads to 5Y tenor).
Probability of default is bootstrapped from credit spreads for a reference name (any rea-sonable set values) in 6M increment. Linear interpolation over spreads and use of ready PD bootstrapping spreadsheet are acceptable, RR = 40%. CVA LGD { own choice.
Assume the swap is written on 6M LIBOR over 5Y. Notional N = 1 and = 0:5.
To simulate the future values of L6M at times T1; T2; T3; : : : take either (a) HJM MC spreadsheet, calibration using 2 years of recent data preferred, (b) full calibrated LMM, or
(c) ready/own calibrated one-factor model for r(t). Naive Vasicek with constant parame-ters not recommended, use Hull & White or CIR++ which has the elasticity of variance.
De ne MtM position as Floating Leg Fixed Leg = (L6M K) appropriately discounted. Depending on your inputs, choose xed leg (rate) K to have a positive exposure.
Discounting factors for a static case to be taken from the OIS curve. Alternative is to use LOIS spread on simulated curves (see below).
7
CQF Lecture Valuation Adjustment - Implementation has a simple spreadsheet with three plots: yield curve, MtM, and Expected Exposure computation for an IRS. It can be used as a starting point for your CVA calculation and analysis of multiple simulated exposures.
CQF Lecture Valuation Adjustment - Theory lecture relies on xVA exercises from the textbook by Jon Gregory, as on Portal. Spreadsheet10.1 presents Expected Exposure for a swaption instrument; Spreadsheet10.2 evolves Vasicek process for r(t) (constant parameters case) and presents EE/PFE for an IRS.
If performing PCA analysis to update HJM simulation to the recent rates, please refer to the simpli ed technical note PCA: Application to Yield Curves by Richard Diamond from CQF Lecture HJM Model.
Forward LIBOR. OIS Discounting in Models
While based on historic volatility, the HJM o ers a simple SDE, Gaussian simulation of interest rates and a ne calibration with the PCA. HJM re-calibration has to be on 2-3 years of recent data of the inst. forward rates from the BLC curve (Bank of England data).
HJM and LMM output simulated forward curves. You might wish to convert from inst. forward to annualised LIBOR but that is not essential. LMM output is iterative, term structure in columns, the last column only has one terminal rate L(Tn 1), and the LIBOR curve result is taken from the diagonal. LMM SDE is example of terminal measure Q(Tn), whereas HJM operates under simple risk-neutral measure `today' Q.
Ideally discounting factors are taken from the same models (simulated curves) to match the measure. Assume constant LOIS spread and simply subtract from each simulated LIBOR curve to obtain the matching discounting curve. In practice, it might be necessary to operate with di erent LOIS spread for each tenor 0:5; 1; 1:5; : : :, the plot of such spread is referred to as a tenor basis curve. Note. OIS Discounting regime poses modelling challenges: (a) there are no caps traded on the OIS underlying to calibrate LMM model and (b) if undertaken from historical Forward OIS data, the second HJM calibration is best on di erences fs = fF wd;t fF wdOIS;t, representing `the stochastic basis'.
One-Factor Models for r(t)
Calibration of one-factor interest rate models is covered in all common textbooks: t time-dependent parameters, such as [Ti 1; Ti] with multiple instruments (bond prices). r(t) models are good for quick risk management and demonstration, however, the preference for CQF Final Project is on full curve modelling.
r(t) simulation drawbacks: inability to evolve the entire curve (even match the curve `today'), calibration with constant volatility parameter prevents from matching any term structure of market volatility. One remedy to enable the use of r(t) models for pricing purposes, is simulating with the stripped market volatility term structure: e.e., 3M caplet stripped volatility tted to the a,b,c,d function).1
1 (t) is the function of (Ti 1 t), where the tenor Ti 1 comes in as a constant and function is computed piecewise for t < Ti 1.
8
This page intentionally left blank.
9
Credit Portfolio Fair Spread and Sensitivity Analysis
Summary
Price a fair spread for a portfolio of CDS for 5 reference names (Basket CDS), as an expec-tation over the joint distribution of default times. The distribution is unknown analytically and so, co-dependent uniform variables are sampled from a copula and then converted to default times using a marginal term structure of hazard rates (separately for each name). Copula is calibrated by estimating the appropriate default correlation (historical data of CDS di erences is natural candidate but poses market noise issue). Initial results are histograms (uniformity checks) and scatter plots (co-dependence checks). Substantial result is sensitivity analysis by repricing.
A successful project will implement sampling from both, Gaussian and t copulae, and price all k-th to default instruments (1st to 5th). Spread convergence can require the low discrepancy sequences (e.g., Halton, Sobol) when sampling. Sensitivity analysis wrt inputs is required.
Data Requirements
Two separate datasets required, together with matching discounting curve data for each.
1. A snapshot of credit curves on a particular day. A debt issuer likely to have a
USD/EUR CDS curve { from which a term structure of hazard rates is bootstrapped and utilised to obtain exact default times, ui ! i. In absence of data, spread values for each tenor can be assumed or stripped visually from the plots in nancial media. The typical credit curve is concave (positive slope), monotonically increasing for 1Y; 2Y; : : : ; 5Y tenors.
2. Historical credit spreads time series taken at the most liquid tenor 5Y for each ref-
erence name. Therefore, for ve names, one computes 5 5 default correlation matrix. Choosing corporate names, it is much easier to compute correlation matrix from equity returns.
Corporate credit spreads are unlikely to be in open access; they can be obtained from
Bloomberg or Reuters terminals (via your rm or a colleague). For sovereign credit spreads, time series of ready bootstrapped PD5Y were available from DB Research, however, the open access varies. Explore data sources such as and . Even if CDS5Y and PD5Y series are available with daily frequency, the co-movement of daily changes is market noise more than correlation of default events, which are rare to observe. Weekly/monthly changes give more appropriate input for default correlation, however that entails using 2-3 years of historical data given that we need at least 100 data points to estimate correlation with the degree of signi cance.
If access to historical credit spreads poses a problem remember, default correlation matrix can be estimated from historic equity returns or debt yields.
10
Step-by-Step Instructions
1. For each reference name, bootstrap implied default probabilities from quoted CDS and
^ ^
convert them to a term structure of hazard rates, Exp( 1Y ; : : : ; 5Y ).
2. Estimate default correlation matrices (near and rank) and d.f. parameter (ie, calibrate copul ). You will need to implement pricing by Gaussian and t copul separately.
3. Using sampling form copula algorithm, repeat the following routine (simulation):
a) Generate a vector of correlated uniform random variable.
b) For each reference name, use its term structure of hazard rates to calculate exact time of default (or use semi-annual accrual).
c) Calculate the discounted values of premium and default legs for every instrument from 1st to 5th-to-default. Conduct MC separately or use one big simulated dataset.
4. Average premium and default legs across simulations separately. Calculate the fair spread.
Model Validation
The fair spread for kth-to-default Basket CDS should be less than k-1 to default. Why?
Project Report on this topic should have a section on Risk and Sensitivity Analysis of the fair spread w.r.t.
1. default correlation among reference names: either stress-test by constant high/low correlation or percentage change in correlation from the actual estimated levels.
2. credit quality of each individual name (change in credit spread, credit delta) as well as recovery rate.
Make sure you discuss and compare sensitivities for all ve instruments.
Ensure that you explain historical sampling of default correlation matrix and copula t (uniformity of pseudo-samples) { that is, Correlations Experiment and Distribution Fitting Experiment as will be described at the Project Workshop. Use histograms.
Copula, CDF and Tails for Market Risk
The recent practical tutorial on copula tting and market risk is o ered at the link below. Semi-parametric CDF tting gives us percentile values, Extreme Value Theory comes as the tail exceedances correction for an Empirical CDF. Generalised Pareto Distribution applied to model the tails after percentile thresholds, while the CDF interior remains Gaussian kernel smoothed.
11
Portfolio Construction with Robust Volatility
Summary
Construct a diversi ed portfolio, compute at least two kinds of optimisation, and study robustness of allocations by varying the inputs. Within each optimisation, utilise the Black-Litterman model to update allocations with absolute and relative views. Compute optimal allocations for three common levels of risk aversion (Trustee/Market/Kelly Investor).
Choosing optimisation kind and constraints, devising views, stabilising historical covariance estimate, and searching for maximum decorrelation among assets and optimal bets { all are methods for robust portfolio construction. Focus on improving the covariance matrix by obtain-ing the e cient volatility estimator. Use EGARCH on asset volatilities and rank correlation methods to construct a stabilised covariance matrix.2 Present the information content of each covariance matrix by plotting its eigenvalues.
The e ective implementation will provide explanation of each optimisation problem, choos-ing from mean-variance, mean-VaR, Tracking Error, Maximum Sharpe Ratio or other suitable index of satisfaction. A successful project will have matrix form calculations and numerical techniques coded (rather than spreadsheet). A well-illustrated study of allocations robustness is a requirement. Also, compare allocations to a market benchmark and discuss.
Portfolio Choice and Data
The main idea behind portfolio choice is to come up with an assets sets that gives an optimal diversi cation. For example, selecting large caps from S&P500 is fully exposed to one factor, the market index itself. (a) If constructing a specialised portfolio, that focuses on an industry, emerging market, credit assets, use 6-8 names and introduce 1-2 most uncorrelated assets wrt your portfolio choice such as commodity or VIX. (b) if constructing a factor or smart beta portfolio { value, momentum, small cap factors { this based on rebalancing (time diversi cation) of long/short position.3
Replication of an index makes sourcing the returns time series easier, and adjusted index weights provide market equilibrium weights. However, index-following is not optimal bets diversi cation which would seek to invest in a few assets and obtain returns similar to ones of the index. For this choice, you have to conduct Tracking Error optimisation.
Portfolio based on explicit factors (value, momentum, small cap) and long/short position requires development of reallocation scheme (eg, weekly, monthly; mean-reversion) and P&L backtesting.
2PCA Multivariate GARCH, aka Dynamic Correlation, reconstructs the covariance matrix from top principal components is a possible but non-a ne approach: we do not know what principal components mean.
3 Statistically, a factor is a time series of returns of a long-short portfolio.
12
Tactical Allocation is a common trend, concentrated on 1-2 macro variables, and driven by data as well as beliefs of fund managers. It is best to formalise such beliefs into BL views. BL as well as any regressions used for prediction are the tools of GTAA.
Further Notes. Markowitz mean-variance optimisation is speci ed for excess simple returns (not log-returns). Risk-free rate can be assumed constant over the periods, 3M US Treasury Bill is relevant.
Recommended historical time period for daily returns data is 2-3 years (> 500 observations). Since portfolio rebalancing is not daily, consider weekly or monthly returns but think of measures to counter autocorrelation in returns of on these time scales, they will not be Normal.
Shorter periods of 1-3 months for variance estimation with EMWA from daily returns. Earning from portfolio investment is not the same as hedging { you are not protected from spikes in correlation and volatility. Those hedges have to be constructed separately; testing portfolio performance over stressed periods is an exercise that is separate from portfolio choice.
A starting source for historical daily close prices of US equities and ETFs is Yahoo!Finance. Today's coding environments have libraries to access Quandl, Bloomberg, Reuters and other.
If benchmark index not available, equilibrium weights computed from the market cap (dollar value).
Step-by-Step Instructions
Part I: Robust Inputs
1. Implement Portfolio Choice based on your approach to optimal diversi cation: introduce an exogenous asset, choose the lesser correlated assets, long/short, implement rebalancing.
2. In addition to a sample covariance matrix, preserve correlations and construct the covari-ance using the appropriate EWMA/GARCH model. GARCH volatility is a forecast, so consider weekly returns. What would be the problem with monthly returns?
Part II: Black-Litterman Model
1. Construct the prior (reference distribution): equilibrium returns can come from a bench-mark index, while covariance is estimated with robustness improvements.
2. De ne input views of both kinds, relative and absolute. [Portfolio Construction utilising predictive algorithms to generate views { please see next topic].
3. Compute the posterior distribution of excess returns using the computational BL.
Part III: Optimisation
1. In addition to variance minimisation, choose, describe analytically, and perform optimi-sation of at least two kinds. Robustness is considerably improved by using sensible constraints, eg, `no short positions in bonds', `no short positions'. Compute allocations for three levels of risk aversion and discuss.
2. Loop back to the study of allocations robustness wrt the covariance input: na•ve historical sample-based covariance vs. EGARCH. Consider advances of Random Matrix Theory, such as (a) eigenvalues analysis and/or (b) shrinkage of the covariance matrix.
13
Data Analytics: Asset Allocation with Views Features
Summary
This topic builds on Portfolio Construction and you will still perform the optimisation within the Black-Litterman framework { guidance presented above remain valid. Except under this topic, you do not need to experiment with the robust covariance input.
Within each optimisation (at least two kinds), utilise the Black-Litterman model to update allocations with absolute and relative views. Identify sensible ways of obtaining and formalising views input for the Black-Litterman model { that can be (a) quantitatively-generated recommen-dations, (b) analysts recommendations or even (c) predicted/inferred analyst recommendations! Another challenge is sourcing of equilibrium market allocations (for the prior).
BL Views
The simple quanti cation of BL views would be to select from `bullish' (+1 SD change), `very bullish' (+2 SD change), and alike for bearish. Your pick or predictive algorithm gives the label, asset standard deviation (SD) gives the number, e.g., asset expected to outperform by 10%.
Quantitative views input to BL can evolve into utilising ready predictive algorithms (clas-si ers). The steps are the following: choosing an algorithm suitable for industry data, de ning predictive features (indicators), and training the algorithm to assign (very) bullish/bearish labels to the current state of indicators. You can use but are not limited to at least two classi ers,
(1) Random Forest, (2) Support Vector Machine, and (3) Na•ve Bayes.
Consider which industry data to feed into the classi ers. Imposing too many views is counter-productive (they will stretch EF allocations in multiple ways), so you can devise two scenarios of 2-4 views each.
Quantitative Diversi cation. Instead of using simplistic approaches to diversi cation (eg, selecting bonds vs. equities, doing tactical allocation), consider formal quantitative approaches to build (or analyse) diversi cation of the chosen portfolio ex ante as follows:-
Rank Correlation between assets.
Decorrelated Factors based on PCA. OPTIONAL Utilise the novel approach of E ective Number of Minimum-Torsion Bets.
Marginal contributions to risk (to portfolio VaR or portfolio volatility).
Compare your allocations to market benchmark weights { check if being over/underweight in a partic-ular asset is a desired result, eg, impact of the view. Check for common pitfalls such as `corner solutions'. Formulate and answer questions such as: Are you over or underweight compared to a market? Is there a common theme, such as shorting low volatility assets and/or concentrated positions in high volatility assets? Are the allocations artefacts of the modelling choice made?
14
Arbitrage Trading on Cointegration with Backtest
Summary
The aim here is the estimation and analysis of an arbitrage relationship between two or more nancial time series. Identifying and backtesting a robust cointegrated relationship means exposing a factor that drives two (several) asset prices. The factor is traded by entering a long-short position given by cointegrating weights.
Through implementation you will have a hands-on experience with model-free multivariate regression with Vector Autoregression (for returns) and Error Correction (for price levels). The task is not to perform econometric forecasting. Backtesting techniques and recipes are spe-ci c to systematic strategy selected. This topic focuses on generating an optimal trading signal from mean-reversion strategy. Cointegrating factor discussed above is represented by a mean-reverting spread, for trading on which optimal bounds can be computed.
A successful project requires coding from the rst principles: matrix form regression esti-mation, Engle-Granger Procedure (or Johansen Procedure), ADF test for stationarity. After the cointegrated (mean-reverting) relationship estimated, ready optimisation and backtesting libraries can be used.4
A project that solely runs pre-programmed statistical tests and procedures on data is insu cient.
Signal Generation and Backtesting
Search for inventive applications of cointegration beyond equity pairs. Consider commodity futures, interest rates, and aggregated indices.
The strategy is realised by using cointegrating coe cients Coint as allocations w. That creates a long-short portfolio that generates a mean-reverting spread. All project designs should include optimal trading signal generation and backtesting. Present optimisation results for entry/exit bounds.
Does cumulative P&L behave as expected for a cointegration trade? Is P&L coming from a few or many trades, what is half-life? Maximum Drawdown and behaviour of volatility/VaR?
Backtest should include bid/ask spread and impact of transaction costs.
Common backtest is done by computing the rolling SR and beta against S&P500 and three factors (returns from value, momentum, and small cap strategies). Use a ready software library, such as pyfolio in Python.
Use the environment with facilities for matrix and time series manipulation (R, Matlab) or code in Python/C++ with the use of quant libraries. VBA will be cumbersome for time series analysis.
15
Step-by-Step Instructions
A starting source for historical daily close prices of US equities and ETFs is Yahoo!Finance.
Today's environments have libraries to access Quandl, Bloomberg, Reuters and others.
`Learning' and Cointegration in Pairs
An understanding-level design can use the ready speci cation tests, but matrix form regres-sion estimation must be re-coded. The project can rely on the Engle-Granger procedure for cointegration testing among pairs but multivariate exploration is encouraged.
1. Implement concise matrix form estimation for multivariate regression and conduct model speci cation tests for (a) identifying optimal lag p with AIC BIC tests and (b) stability check with eigenvectors of the autoregression system. Here, it is choice whether to code these tests or use ready functions.
2. Implement Engle-Granger procedure and explore several cointegrated pairs. Estimate relationships both ways to select the appropriate lead variable. ADF test for unit root must be coded and used.
3. Ensure robust estimation: one recipe to do so is to shift time window (2-3 months shift) and
apply LR testing for di erence in time series' means between samples. As an alternative,
develop the adaptive estimation of cointegrating coe cients Yt Coint0 = et.
4. Decide on strategy trading rules (common approach is to enter on bounds, exit on et touch-ing the mean value e). Use optimisation to compute optimal bounds Zopt eq. Produce
appropriate plots for Drawdown, Rolling SR and backtesting against factors for the P&L. eq, speed of mean-reversion, and half-life between trades are obtained by tting of et to the OU process.
OPTIONAL Multivariate Cointegration
It is recommended to validate results for Johansen Procedure against existing R/Matlab libraries. E cient implementation steps for the procedure are outlined in Jang & Osaki (2001).
New 2. Apply Maximum Likelihood Estimation (Johansen Procedure) for multivariate cointegra-tion on prices data. Test speci cation will involve the in-calibrated constant trend inside the cointegrating residual et 1 in the Yt error-correction equation.
There are ve possible kinds of deterministic trend in et 1, however rarely utilised in practice because that leads to over tting of cointegrated relation to particular sample. Particularly for time-dependent trend (ie, do not make your time series function of time as an independent variable of regression).
Present analysis for Maximum Eigenvalue and Trace statistical tests, both are based on Likelihood Ratio principle, on how you decided the number of cointegrated relationships.
16
Local Volatility in Interest Rates
Summary
LIBOR Local Volatility Model is made possible by introducing a spot-like risk neutral mea-sure for the xed tenor rolling LIBOR. Dupire-type arbitrage-free model for the volatility smile in interest rates was not available. LMM was not suitable to deal with volatility smile (across strikes) without modi cation. Also, the pricing of IR derivatives is limited to Monte-Carlo method for most multi-factor models. Interest rates smile calibration required iterative multi-dimensional numerical optimisation, such as least-squares on top of Monte-Carlo (the Longsta - Schwartz method).
Calibration of IR derivatives' pricing models is done on caps and European swaptions. Volatility stripping from those market option prices is always involved and one has to be im-plemented for LVM. Begin with re-pricing market-quoted caps (cash ow) in volatility terms by the special pricing fomulae for spot process (analytical). Those formulae require the drift, fully computable from today's curve analytically. Then, implied volatility of the rolling LIBOR can be converted into the local volatility.5
The report has to include a discussion of the measure change technique to the rolling LIBOR under the spot measure.
Data for this project to be taken from market sources (eg, Bloomberg) but can be simulated from a pre-calibrated HJM Model. Simulate cap prices from a calibrated HJM Model and converting them to implied volatility terms using the Black formula. Please see LIBOR and OIS rates topic for data sources.
Step-by-Step Instructions (all analytical)
1. Obtain the quoted caps from market data or HJM simulation { must be cash ows. Com-pute cap prices (in volatility terms) by special pricing fomulae for the spot process.
Those formulae require the drift [T ] computable and fully known from today's curve analytically.
Result is an implied volatility of the rolling LIBOR imp(T; K). Interpolation to 3M granularity can be done using linear variance imp2T .
2. European swaptions also have fully analytical pricing formula to obtain imp(T; K).
3. Implied volatility of rolling LIBOR to be converted into local volatility using the solution for a Dupire-type forward PDE (to be provided at the Workshop).
Local volatility can be used to price path-dependent exotics in IR derivatives space. Backward PDE (very much like Black-Scholes) is available for pricing.
17
LIBOR and OIS Rates - Market Volatility
Summary
This advanced topic is for delegates with experience in interest rates who would like a chal-lenge. The topic does require access to caplet and/or swaption data, an alternative will be to make up the data by simulating cap prices from a pre-calibrated HJM Model and converting them to implied volatility terms using the Black formula.
Discretised LMM SDE requires stripping of 3M caplet volatility from market cap prices and converting into instantaneous volatility is tted to the a,b,c,d function. Be prepared if the tted inst. volatility does not give good t to the market caps and oors, and basis curve is required. Pricing of caps and swaptions is done using the Monte-Carlo. The outcomes are a. sensitivity analysis of caplet pricing wrt the market risks (ie, bumping the curve) and b. dis-cussion of interest rate swaps and pricing of vanilla swaptions. Pricing of Bermudan swaptions is not a requirement. Joint caplet/swaption calibration is not a requirement but advantageous.
If you decide to look into implementing SABR model, you may do so on advice of a tutor. SABR has the explicit analytical solutions for volatility tting and arbitrage-free advantage that comes from matching the market volatility skew. Shifted SABR allows to deal with negative rates.
Multi-curve modelling many incarnations: LMM with stochastic basis, displaced di u-sion, credit driven OIS-LIBOR basis (LOIS). The concept of tenor basis is simple: a typical tenor basis swap has two oating legs, for example, L3M into L6M. The nuance is: one leg pays a spread above an index rate.6
Ad-hoc calibrated `basis curve' { you might end up introducing this curve for your cali-bration of LMM, etc. to match the market prices.
fsj = fFj wd;t fFj wdOIS;t from the historical data (BOE) gives tenor basis vs OIS for each tenor j.
\With stochastic basis" modelling invites three choices: modelling joint evolution of two
rates, or rate and a spread; from the spread paid over the indexed rate S(t) = L(t) F (t).
The explicit link between tenor basis and credit spread is given by `spot credit spread' and `fwd credit spread' computable using CDS-implied probabilities of default, see Castagna et al. (2015).
LIBOR or funding spread already includes credit risk. The tenor basis comes from the optionality of rolling credit in shorter tenors, this is the second order e ect of credit.
Tenor basis swap (money market basis) is a market instrument with maturity up to 25Y or 30Y point. So, the basis swap of 3M vs 6M has values for all lengths of nancing from 6 months to 30 years.
18
Remember, there is always a simplifying assumption of the constant (piecewise) LIBOR-OIS spread (LOIS). Simply subtract the constant from each simulated LIBOR curve to obtain a matching discounting curve.
Data Sources: Interest Rates
Caplet or swaption data is usually maintained by trading desks and interdealer brokers. Data for certain markets is available from Thomson Reuters and Bloomberg. The simple input, a strip of market cap volatility prices (1M, 2M, 3M, 6M, 9M, 12M, 2Y, 3Y, etc.) can be taken from a textbook/research paper.
LMM can also be calibrated to swaption volatility data (Option Maturity 1Y, 2Y, 3Y etc. vs. Swap Length 1Y, 2Y, 3Y, etc. that begins after maturity) { all at-the-money swaption volatilities (strike is equal to the rate on Forward Swap). That calibration is achieved via optimisation and called the Rebonato Method.
For each tenor, you will need a discount factor and should make a decision on whether to use dual-curve discounting where DF is coming from the OIS curve. You will also need Forward LIBOR rates given deterministically by the forward curve today.
Below are the links for Pound Sterling Bank Liability Curve (BLC) from LIBOR-linked instruments, OIS spot rates, Euro area curve (Government Liability) and FRB H.15 release (Treasuries and Interest Rate Swaps, each instrument gives its own spot curve):
19
Step-by-Step Instructions
Part I: Data
1. You will need market price data { cap price and discount factor (two columns). Caps are quoted in terms of implied volatility cap(t; T ) with T = 1Y; 2Y; 3Y; : : :.
a) For pre-simulated caplet data (ie, from the HJM model) the Black formula is conven-tional means of converting the caplet's cash ow to the implied volatility number.
2. The second set of data to which model tting can be done is swaptions, for which the deliverable asset is an interest rate swap.
Part II: Calibration (Volatility Stripping)
3. Strip 3M caplet volatilities cap(Ti; Ti+3M ) or market traded caps traded in one-year in-crement using simplifying assumptions, e.g., at volatility and pre-interpolated volatilities cap(t; T + 3M) using de nition of a cap:
cap(t; 1Y ) = caplet(t; T3M ; T6M ) + caplet(t; T6M ; T9M ) + caplet(t; T9M ; T12M ):
To use Black formula you will need from today's curve, forward LIBOR fi and strike K
a) The rst step is to determine a strike for each caplet as a forward swap rate S(t; Ti 3M ; Ti).
b) The second step is to strip caplet volatilities.
4. Alternatively, volatilities can be calibrated from vanilla swaptions (European options on forward-starting swaps) where Rebonato method makes the Black formula suitable.
5. Fitting the abcd instantaneous volatility function is de ned for each tenor as i(t).
Coe cients abcd are estimated by optimisation that can be joint wrt caplet implied volatil-ities icap and swaption implied volatilities i. Goal is to minimise the squared di erences between two implied volatilities (for the same tenor enumerated i).
a) Use a parametric function for correlation structure ij.
Part II: Pricing and Sensitivity Analysis
Pricing of swaptions has already been done in the process of calibration (stripping) of caplet volatilities because forward swap rates S(t; Ti 3M ; Ti) were calculated. Pricing of path-dependent options, such as Bermudans that give exercise exibility on some or all payment dates Ti; Ti+1; :::; Tm, would require the modi ed Least Squares Monte-Carlo simulation. Make sure that calibrated LMM model returns cap prices similar to the input data. Sensitivity Analysis means modifying the input forward curve (today) and/or discounting factors to evaluate the impact on derivatives pricing.
Another question is to explore how the LIBOR Market Model has to be modi ed with the introduction of OIS discounting. The change of numeraire to the discounting factor given by the OIS curve leads to an adjustment to the drift term of the LMM SDE.
20
Resources
Reading List: Rates Volatility
Review Methods for Constructing a Yield Curve by Pat Hagan and Graeme West { best to start with version published in WILMOTT (May-June 2008).
The LIBOR Market Model in Practice specialised textbook by Gatarek, et al. (2006) gives technical detail on cailbration from caplets and swaptions (Chapters 7 and 9 respectively) that will be useful to those working with LIBOR derivatives. (Please email the tutor.)
On Local Volatility in Interest Rates see CQFI Talk by Dong Qu and Chapter 12 (partic-ularly pp. 319-331) from his Manufacturing and Managing Customer-Driven Derivatives.
Reading List: Credit Portfolio
Very likely you will need to re-visit CDO & Copula Lecture material, particularly slides 48-52 that illustrate Elliptical copula densities and discuss Cholesky factorisation.
Sampling from copula algorithm is given Project Workshop and Monte Carlo Methods in Finance textbook by Peter Jaekel (2002) { see Chapter 5.
Rank correlation coe cients are introduced Correlation Sensitivity Lecture and Jaekel (2002) as well. Project Q&A document gives the clari ed formulae and explanations.
Reading List: Portfolio Construction
CQF Lecture on Fundamentals of Optimization and Application to Portfolio Selection
A Step-by-step Guide to The Black-Litterman Model by Thomas Idzorek, 2002 tells the basics of what you need to implement.
The Black-Litterman Approach: Original Model and Extensions Attilio Meucci, 2010.
Risk Budgeting and Diversi cation Based on Optimized Uncorrelated Factors is an advanced piece by Meucci, et al. (2015) node/599 { see Section 7.
Reading List: Time Series
CQF Lectures on Cointegration and Statistical Methods for PD.
21
................
................
In order to avoid copyright disputes, this page is only a partial summary.
To fulfill the demand for quickly locating and searching documents.
It is intelligent file search solution for home and business.
Related download
- find tutor online
- faculty of engineering and mathematical sciences
- international college of economics and finance
- anthropology 5305 home committee on curricula and
- introduction e cab e commerce association of
- univerzita komenskÉho v bratislave
- information and software technology ist year 9 2005
- ajmal mian associate professor uwa
- doc to help small manual
Related searches
- word find games online free
- find jobs online in jamaica
- tutor for free online chat
- free math tutor online chat
- best free typing tutor download windows 10
- typing tutor free download for windows 10
- typing tutor download windows 10
- typing tutor app download
- how to find addresses online free
- tabe tutor 11 12
- live math tutor free
- find someone online for free