The Hidden Correlation of
Collateralized Debt Obligations
N. N. Kellogg College University of Oxford
A thesis submitted in partial fulfillment of the MSc in Mathematical Finance
April 13, 2009 Acknowledgments
I would like to thank my supervisor Dr Alonso Pe˜nafor his advice and help and for encouraging me to follow the approaches taken in this thesis. I want to thank my wife Blanca for her direct (by reading and correcting) and indirect (by motivating me) support of this work.
I would like to thank my former employer d-fine for the possibility to attend the MSc Programme in Mathematical Finance at the University of Oxford. For their engage- ment in this programme, I particularly would like to thank Dr Christoph Reisinger, the course director, and Prof Dr Sam Howison. Thank also goes to the other lecturers of the Mathematical Finance Programme. Finally, I thank my friend Ari Pankiewicz for the effort to read and correct this thesis.
ii Abstract
We propose a model for the correlation structure of reference portfolios of collater- alized debt obligations. The model is capable of exhibiting typical characteristics of the implied correlation smile (skew, respectively) observed in the market. Moreover, it features a simple economic interpretation and is computationally inexpensive as it naturally integrates into the factor model framework.
iii Contents
List of Figures v
List of Tables vi
1 Introduction to Collateralized Debt Obligations 1 1.1 The CDO Market ...... 1 1.2 Valuation of STCDOs ...... 4 1.3 Outline ...... 5
2 Modeling of Multivariate Default Risk 7 2.1 Structural Models ...... 8 2.2 Copula Functions ...... 9 2.3 Factor Models ...... 11
3 Default Correlation 18 3.1 Implied Correlation ...... 19 3.2 Correlation Smile ...... 22
4 Modeling the Correlation Matrix 25 4.1 Empirical Correlations and RMT ...... 26 4.2 Group Models ...... 28 4.3 Attainable Correlation Smiles ...... 31
5 Numerical Implementation 40 5.1 Integration Schemes ...... 40 5.2 Existence and Uniqueness of Correlation ...... 48 5.3 Performance ...... 49
Conclusions 51
iv CONTENTS CONTENTS
A Inverse Transform Sampling 53
B Correlation in Student-t Factor Model 54
C Used Parameters 56
Bibliography 59
v List of Figures
3.1 Example of correlation smile/skew observed in the market...... 22 3.2 Loss distribution, un-skewed vs. skewed...... 23
4.1 Example of implied correlations for non-flat correlation matrix. . . . . 25 4.2 Implied corr. for varying µ in the two-layer hierarchical model. . . . . 32 4.3 Implied corr. for varying ∆ in the two-layer hierarchical model. . . . 32 4.4 Implied corr. for varying n in the two-layer hierarchical model. . . . . 33 4.5 Implied corr. for varying ν in the two-layer hierarchical model. . . . . 34
4.6 Implied corr. for varying ∆n in the two-layer hierarchical model. . . . 35
4.7 Implied corr. for varying µρ in the two-layer hierarchical model. . . . 36
4.8 Implied corr. for varying ∆ρ in the two-layer hierarchical model. . . . 36
4.9 Implied corr. for varying ∆n in the block model...... 37
4.10 Implied corr. for varying ∆ρ in the block model...... 38
5.1 Integration schemes for the 1-factor Gaussian copula...... 46 5.2 Integration schemes for the 2-factor Student-t copula...... 47 5.3 Integration schemes for the 2-factor Gaussian copula...... 48 5.4 Uniqueness and existence of implied compound correlations...... 49
vi List of Tables
4.1 Types of group models...... 29
5.1 Test cases for the validation of the developed program code...... 41 5.2 Tranche spreads: LHP vs. numerical results in factor model framework. 49 5.3 CPU times for CDO pricing in different parameter regimes...... 50
C.1 Used parameters for the validation of integration schemes...... 56 C.2 Used parameters for the implied correlations of group models. . . . . 57
vii Chapter 1
Introduction to Collateralized Debt Obligations
1.1 The CDO Market
After a rapid growth in the recent ten years, the market for collateralized debt obli- gations (CDOs) has seen drastic and devastating changes in the months prior and during the creation of this thesis. Being in the center of the subprime mortgage crisis, the value of existing CDO positions has sharply dropped –if a price can be assigned at all, given the almost non-existing market– and the issuing of new CDOs has slowed down by almost 90% in 2008 compared to 2007 [1]. For the riskiest CDO portions on subprime mortgages, the term ”toxic waste” has been coined and the underlying mortgage-backed securities (MBS) are the main focus of the US$ 700 bil- lion ”Troubled Assets Relief Program” (commonly referred to as ”bailout”) of the US government [2]. We will later come to the involvement of CDOs in these recent developments, but first lay out the basic ideas behind CDOs.
CDOs are structured finance products that transfer and distribute credit risk on a reference portfolio of assets into tranches of increasing seniority. The risk is transferred from the arranger of the CDO, who acts as buyer of protection, to the investor, who acts as seller of protection. The common feature of all CDOs is the splitting of the underlying reference portfolio into different tranches, which defines a sequential allocation of the losses that occur in the portfolio. The lowest-lying tranche (equity tranche) is the first one to absorb losses, with the losses resulting in a payment from the investor to the arranger (and a decrease of the premiums paid to the investor, see below), up to a coverage limit which defines the size of the tranche (e.g. 3% of the CDO notional). If the cumulative losses exceed this limit, the next tranche (mezzanine tranche, e.g. 3%-6%) is affected, and so on. The CDO investors thus
1 1.1 The CDO Market 1 Introduction to Collateralized Debt Obligations take on exposure to a particular tranche, where the higher risk of the lower (junior) tranches is naturally rewarded with a higher premium, whereas the higher (senior) tranches are protected by the junior tranches, hence yielding a much lower spread.
In essence, CDOs are based on the idea that the idiosyncratic risk of the names in the portfolio can be diversified out and that the diversified pool of risky assets has a relatively predictable loss pattern, where losses hardly ever touch the senior tranches. This allows the senior tranches to achieve a higher rating than the average rating of the underlying portfolio. However, the residual systematic risk crucially depends on the dependency structure (loosely termed ”correlation” in the following, see Chapters 2 and 3 for details) of the portfolio and gives rise to correlation risk and correlation trading. Importantly, the correlation risk varies among the tranches: a high correlation benefits equity investors, whereas the price of senior tranches drops as correlation rises, see Chapter 3.
Originally, the aim of the first CDO deals was the management of the balance sheet of the originating banks by making use of the possibility to transfer credit risk and free up regulatory capital (balance-sheet CDOs), with the underlying assets typically being cash assets like bank loans, bonds, or asset-backed securities (ABS, MBS). The ownership of these assets then needs to be transferred to a separate legal entity (SPV, special purpose entity) which acts as arranger of the CDO and issues the tranches (cash CDO). In contrast, synthetic CDOs are backed by credit default swaps (CDS); the arranger typically does not hold these CDS positions beforehand but enters into them only in order to set up the CDO. This is done to take advantage of a possible spread difference between the average yield of the underlying CDS portfolio and the spread paid on the tranches (arbitrage CDO) 1. The spread difference can be seen as service charge for bundling the portfolio and reflects the benefit for the investor, who otherwise would have to assemble the portfolio himself.
As further step in the product development of synthetic CDOs, single-tranche CDOs (STCDOs) entered the market in 2003 and in the same year already accounted for 90% of the issued synthetic CDOs [12]. In a STCDO, only one isolated tranche is sold to a investor, thereby carving out a piece of the loss profile. From the investor’s point of view, the popularity of STCDOs originates from the flexibility in the risk structure (the underlying credit portfolio, level of subordination, and tranche size can be chosen by the investor according to his needs) while providing a substantially higher yield than similarly rated investments, whereas the arrangers appreciate the comparably
1An arbitrage CDO can also be set up as cash CDO.
2 1 Introduction to Collateralized Debt Obligations 1.1 The CDO Market easy set-up of the transaction. However, as the arranger sells only part of the CDO structure, he exposes himself to the ”remaining” credit risk and needs to hedge that risk. STCDOs are often written on standardized CDS indices like CDX and iTraxx which facilitates this hedging for the arranger. Obviously, STCDOs are leveraged products with respect to the spreads of the underlying names: if the portfolio losses would increase from e.g. 3% to 6% due to a change in the credit spreads of the same order of magnitude, a 3%-6% STCDO would suffer a total loss. In the course of the product evolution, also more sophisticated products have appeared on the market of which we only want to give two examples: CDO2 are CDOs backed by tranches of other, underlying CDOs (and ABS, typically), where additional leverage is achieved by the overlap of names in the reference portfolios of the underlying CDOs (can be further enhanced to CDOn). Callable STCDOs are the equivalent to bermudan options in equity markets and are particularly difficult to price due to their loss path dependence (cp. footnote 2 in Chapter 2). The focus of this work lies on standard CDOs, though.
With respect to financial stability, the CDO market was believed to have a positive impact, as it facilitates the distribution of credit risk to non-bank investors and in- creases the degree of completeness of the credit market [12]. However, the narrowing of the corporate spreads in the years before the financial crisis (which, due to the leverage, had a multiplicative effect on the CDO spreads) increased the pressure to find higher yield/higher risk underlyings to keep engagement into CDOs profitable for investors and arrangers, bringing subprime mortgage loans into focus. However, this selective choice of loans also implied a higher correlation than anticipated from historical estimates: the subprime loans are more correlated because they are all un- derwritten with weak credit standards. This was not adequately reflected by both the credit rating agencies, which continued to assign first-class ratings to the senior CDO tranches, as well as by the market participants who continued to invest. As default rates on mortgage loans started to rise in 2007, the affected CDOs experienced severe losses and rating downgrades. For the case of synthetic CDOs, systemic risk also emerges from the fact that the number of liquid CDS names is limited and arrangers are forced to constantly draw from the same pool of names. Consequently, the same name is contained in many CDO deals, greatly amplifying the impact of a default. Also, an important point to consider is the back-action of the growth of the CDO issuance on other markets: as pointed out above, the arranger of a STCDO needs
3 1.2 Valuation of STCDOs 1 Introduction to Collateralized Debt Obligations to hedge the credit risk he is exposed to, namely by selling protection on the CDS market, thereby contributing to the narrowing of the spreads. Due to the leverage of the STCDOs, this has to be done with a multiple of the notional value of the sold tranches. Moreover, these developments were probably supported by the high complexity of the involved products, with models and techniques that were not always sufficiently tested and understood. Especially on the investor’s side, the lack of expert knowledge probably led to CDO investments without the ability to correctly assess and manage the associated risk.
1.2 Valuation of STCDOs
As pointed out above, the investor of a STCDO covers only the tranche losses of the tranche he is invested in, i.e. the losses in the interval [aL, aH ], where aL (aH ) is referred to as lower (upper) attachment point and the size (or notional) of the tranche is defined as aH − aL. Each loss within this tranche results in a payment equal to the loss amount from the investor to the arranger; in analogy to a credit default swap, this is called the protection leg of the STCDO. As compensation, the investor receives periodic payments (premium leg), usually paid quarterly in arrears 2, at a spread s on an outstanding principal which is initially equal to the notional of the tranche and stochastically declines with each loss by the size of the loss. Based on this, the STCDO can be valued as a swap contract, with the value –from the investor’s point of view– equal to the expected present value of the premium payments less the expected present value of the protection payments. The expected tranche loss at time t reads
+ E [Ltr(t)] = E (min(L(t), aH ) − aL) , (1.1) where the expectation is taken with respect to the risk neutral measure and L(t) is the portfolio loss, see Eq. (2.1). If the loss distribution F (x; t) ≡ P(L(t) ≤ x) is known (e.g. in the LHP approximation, see Sec. 2.3.5), the expected tranche loss can be written as Z I E [Ltr(t)] = (min(x, aH ) − aL)dF (x; t), (1.2) aL where I is the maximal loss, i.e. the initial portfolio amount.
2Except for the equity tranche, where a part of the premium is usually paid upfront.
4 1 Introduction to Collateralized Debt Obligations 1.3 Outline
For a portfolio of finite size, we use the (numerically obtained) discrete loss proba- bilities P (w; t) ≡ P(L(t) = wu) (see Sec. 2.3.3 for the meaning of the loss unit u),
X + E [Ltr(t)] = P (w; t) (min(wu, aH ) − aL) . (1.3) w With this, the value of the premium leg evaluates to
n X Vpremium = s D(ti)∆ti aH − aL − E [Ltr(ti)] , (1.4) i=1 where s is the spread, 0 < t1 < . . . < tn = T are the payment dates, D(ti) is the discount factor for time ti, and ∆ti ≈ ti − ti−1 is the accrual factor. The value of the protection leg reads
Z T 0 0 Vprotection = E D(t ) dLtr(t ) , (1.5) 0 however, one commonly assumes that the payments are made on the next periodic payment date after the default (otherwise, one would also have to compute the loss distribution for each banking day in between the payment dates), thereby essentially neglecting the interest that the payment would accrue since the default,
n X Vprotection = D(ti)(E [Ltr(ti)] − E [Ltr(ti−1)]) . (1.6) i=1
The fair (or par) spread sPar is the spread that sets the total value of the tranche to zero, Vprotection sPar = P . (1.7) i D(ti)∆ti aH − aL − E [Ltr(ti)]
Hence, central for the valuation of the STCDO is the determination of the portfolio loss distribution F (x; t), respectively P (w; t). This will be the main focus of Chapter 2.
1.3 Outline
From the discussions in Sec. 1.1 above, it should have become apparent that the cor- relation among the names in the reference portfolio of a CDO is of crucial importance. It is therefore the primary topic of this thesis. In Chapter 2, we will cover the basics of the modeling of multivariate default risk in order to derive the loss distribution. It will become clear how the correlation
5 1.3 Outline 1 Introduction to Collateralized Debt Obligations between the names in the underlying portfolio of the CDO enters into the valuation. The Gaussian copula framework with flat correlation is introduced as the standard market model for the valuation of CDOs. The different concepts of historical vs. implied correlation are pointed out in Chap- ter 3, with the emphasis on the two methodologies for implied correlation, namely compound and base correlation. The properties, advantages, and pitfalls of these two measures are discussed. The starting point for the investigations in this thesis is the empirical observation that the market reports a non-constant correlation across the different tranches of a CDO (called smile, or skew, respectively) that contradicts the assumption of a constant correlation in the Gaussian copula framework. In order to reproduce the correlation smile, a model for the correlation matrix is proposed in Chapter 4, based on the investigation of empirical correlation matrices of stock price returns and on economic intuition. The shapes of the implied correlation curves emerging from this model are analyzed in detail for different parameter regimes. Chapter 5 finally focuses on some numerical issues that one encounters when imple- menting a CDO valuation code.
6 Chapter 2
Modeling of Multivariate Default Risk
The aim of this chapter is the derivation of the portfolio loss distribution F (x; t) in Eq. (1.2) respectively the discrete loss probabilities 1 P (w; t) from Eq. (1.3). To this end, we employ different techniques from default risk modeling like structural models, copulas, and factor models. Let us introduce some notation: we consider a reference portfolio of N obligors subject to default risk with random default times τ1, τ2, . . . , τN . Modeling of multivariate default risk denotes the modeling of the joint distribution H(t1, t2,...) ≡ P(τ1 ≤ t1, τ2 ≤ t2,...) with marginal CDFs pi(t) ≡ P(τi ≤ t) = E [1τi≤t]. pi(t) is called the default probability of name i and is assumed to be known, calibrated from credit curves (e.g. risky bonds or credit default swaps) on name i. Thus, one only needs to specify the dependency function (copula) for a full characterization of H, see Sec. 2.2. 2
The default of obligor i results in a loss li = Ni(1 − Ri) (for recovery type CDOs), where Ni is the notional of the i-th credit and Ri ∈ [0, 1] is its recovery rate, i.e. the percentage of the notional that is expected to be recovered in case of default by selling assets of the obligor. The random portfolio loss is then given by
N X L(t) = li1τi≤t , (2.1) i=1
1We will henceforth use the term ”loss distribution” also for P (w; t). 2To be more precise, we point out that the valuation of a CDO doesn’t require knowledge about the full joint distribution of default times which also contains all inter-temporal correlations, but rather the portfolio loss distribution which is the marginal distribution of losses at a fixed point in time. As described in Ref. [5], for the CDO valuation only the loss distribution at the discrete set of payment dates is needed.
7 2.1 Structural Models 2 Modeling of Multivariate Default Risk with expectation value
N X Z t E [L(t)] = E [li|τi = s] dpi(s). (2.2) i=1 0
Note that no default dependence enters here, i.e. the expected loss over the whole portfolio (respectively tranche structure) depends only on the single-name default probabilities. The expected loss suffered by part of the structure –like a tranche– is significantly affected by the default dependence between the obligors, though. This makes the modeling of the default dependence in Sec. 2.2 a key issue for the valuation and risk management of CDOs. In a first step, we cover the modeling of univariate default risk, though. The factor model approach that is followed in Sec. 2.3 derives from Merton’s structural model approach [33] for univariate default risk, we will therefore introduce structural models in the following. 3
2.1 Structural Models
Structural models (also termed asset value models or firm value models) define the time of default of a firm as the moment when the firm’s assets fall below its liabilities.
The asset value is modeled as random process Xt and looking at a fixed time scale (maturity) t the default is given by
τ ≤ t ⇔ Xt(t) ≤ C, (2.3) where the barrier C stands for the liabilities. The Black-Cox approach [8] extends this model by allowing a default anytime before maturity at the first hitting time of a (possibly time varying) barrier,
τ = inf{t : Xt(t) ≤ C(t)}. (2.4)
We will focus on the simplified case of static models that suppress the time dependence of X by treating it as random variable; let C be an increasing function of time. This relates the random variables X and τ by X = C(τ).
3We will not cover intensity-based models [25] in this work. We will, however, sometimes use the notion of the hazard rate h, borrowed from the intensity framework. For our purposes, the default probability and the hazard rate are related by p(t) = 1 − e−ht.
8 2 Modeling of Multivariate Default Risk 2.2 Copula Functions
In order to calibrate the static model to the market implied default curve p(t), the level of the barrier is adjusted. We have
p(t) = P(X ≤ C(t)) = P(FX (X) ≤ FX (C(t))) = P(U ≤ FX (C(t))) = FX (C(t)), (2.5) where FX is the distribution of X, and U denotes a uniformly distributed random variable (we have used the fact that FX (X) is uniformly distributed, cp. Appendix A). The calibration equation for C(t) then reads
−1 C(t) = FX (p(t)). (2.6)
Structural models naturally extend to the modeling of multiple dependent defaults since a default dependence can easily be incorporated by a correlation between the individual asset processes (variables, respectively). This will be useful in the context of factor models below.
2.2 Copula Functions
The specification of a joint distribution of default times is central to the valuation of basket credit derivatives. It is economically plausible and empirically shown that default events of obligors are not independent from each other but exhibit positive dependence, i.e. due to economic cycles or firm interactions, defaults tend to cluster together. Consequently, the joint default probability cannot simply be modeled by the product of the individual default probabilities. Obviously, the marginal single-name default probabilities discussed in Sec. 2.1 don’t uniquely fix the joint distribution. A convenient and powerful approach to specify the missing part, i.e. the default dependence between the obligors, is the use of copula functions. In general, copulas allow for the separate treatment of marginal distributions and dependency structure by linking given marginal distributions to a joint distribution. In this section a very brief and informal introduction to copulas is given, see e.g. Ref. [14] for further reading and proofs. A copula is defined as a multivariate distribution function on the unit hypercube with uniformly distributed marginals,
C(u1, . . . , un) = P(U1 ≤ u1,...,Un ≤ un), (2.7) C(1,..., 1, ui, 1,..., 1) = ui , ui ∈ [0, 1] . The idea of the analysis of dependency with copula functions is expressed by Sklar’s n theorem [36]: if F (x1, . . . , xn): R → [0, 1] is a n-dimensional distribution function
9 2.2 Copula Functions 2 Modeling of Multivariate Default Risk
with marginal distributions F1(x1),...,Fn(xn), then there exists a copula function C n (C is unique if the Fi(xi) are all continuous) such that for all (x1, . . . , xn) ∈ R
F (x1, . . . , xn) = C(F1(x1),...,Fn(xn)) (2.8) holds, i.e. C joins the marginal distributions to the joint distribution.
In practice, if a joint distribution function with a suitable dependency structure has been identified, the copula containing the dependency structure can be extracted by
−1 −1 C(u1, . . . , un) = F (F1 (u1),...,Fn (un)). (2.9)
Then, a set of given marginal distributions (e.g. the individual credit curves pi(t) obtained from the market) is plugged into Eq. (2.8), resulting in a joint distribu- tion function with the given marginal distributions and the extracted dependency structure. For the popular multivariate Gaussian, the extracted Gaussian copula reads (Φ being the standard normal CDF)
−1 −1 Cρ(u1, . . . , un) = Φρ(Φ (u1),..., Φ (un)) −1 −1 1 Z Φ (u1) Z Φ (un) 1 = √ ... exp − xT ρ−1x dx. (2.10) n/2 (2π) detρ −∞ −∞ 2
For the special case of the Gaussian copula, the dependency structure is fully described by the matrix of pairwise correlations ρ. The (linear) correlation between two random variables X and Y with nonzero finite variances σ(X) and σ(Y ) is defined as
E[XY ] − E[X]E[Y ] ρ(X,Y ) = . (2.11) σ(X) σ(Y )
Linear correlation is measure of linear dependence, since Y = aX + b almost surely iff |ρ(X,Y )| = 1. For joint distributions other than the multivariate Gaussian, the correlation does not completely define the dependency structure 4. Other measures of dependency have been proposed; however, since the Gaussian copula is the industry standard in credit risk modeling, the correlation is commonly used as measure of dependency, despite its limited applicability. Due to its particular importance for the valuation of CDOs, correlation is also used as quotation device, see Chapter 3.
4This even applies to other distributions belonging to the class of elliptical distribution, like the Student-t distribution, see Appendix B.
10 2 Modeling of Multivariate Default Risk 2.3 Factor Models
2.3 Factor Models
Factor models are a widely used simplification in the context of credit risk modeling and provide an efficient way to deal with a large number of credits with default dependence [18]. We assume a low- (d-) dimensional random variable –the factor–
Z ∼ GZ conditional upon which default times become independent, i.e. the Z- conditional defaults are assumed independent events with probability
P(τ ≤ t|Z = z) ≡ pi(t|Z). (2.12) The independence of the conditional default probabilities allows for the application of recursion arguments that lead to large gains in the efficiency of the calculation of the (discrete) conditional loss probabilities P (w; t|Z) ≡ P(L(t) = wu|Z), see Sec. 2.3.3. The unconditional loss distribution then follows from integrating P (w; t|Z) over the distribution GZ of Z, Z P (w; t) = P (w; t|Z = z) dGZ (z). (2.13) Ω The integration domain Ω depends on the properties of the random variable(s) that are taken as factor(s), e.g. Ω = Rd for a d-dimensional Gaussian factor structure, see d Sec. 2.3.1, and Ω = R × R+ for a d-dimensional Student-t factor structure, see Sec. 2.3.2.
The form of the conditional default probability pi(t|Z): R+ ×Ω → [0, 1] emerges from the used factor model with the constraint that the (unconditional) default probabili- ties of the individual firms are preserved,
Z ! pi(t|Z = z) dGZ (z) = pi(t). (2.14) Ω Following the structural model approach and notation, the random variable X de- scribing the asset value in a factor model is given by 5 q 2 Xi = βiZ + 1 − βi i. (2.15) The systematic factor Z (which can be thought of as a common macroeconomic driver or, loosely speaking, ’the market’) is shared by all Xi, and i is the residual driver idiosyncratic to firm i and is independent of Z and j ∀j 6= i. The β’s are called factor loadings. In the following, we describe two common models: the Gaussian factor model and the Student-t factor model. 5 The products in general need to be taken as scalar products, βiZ ≡ βi · Z, βiβj ≡ βi · βj , 2 βi ≡ βi · βi
11 2.3 Factor Models 2 Modeling of Multivariate Default Risk
2.3.1 Gaussian Copula
The Gaussian copula model [28] is the industry standard model. For a Gaussian copula, Z and are independent N (0, 1) distributed. Being a weighted sum of two
standard Gaussian random variables, Xi is also standard Gaussian and the correlation ρ between two obligors directly evaluates to
β β , for i 6= j ρ(X ,X ) = i j (2.16) i j 1 , for i = j
i.e. X = (X1,...,XN ) is a N-dimensional Gaussian variable with correlation matrix ρ. With Φ the standard normal distribution, the conditional default probability reads ! Ci(t) − βiZ pi(t|Z) = (Xi < Ci(t)|Z) = Φ . (2.17) P p 2 1 − βi
Since Xi ∼ N (0, 1) and following Eq. (2.6) the calibration equation for Ci(t) reads
−1 Ci(t) = Φ (pi(t)). (2.18)
2.3.2 Student-t Copula
The Student-t distribution has fatter tails than the Gaussian distribution and is known to generate tail dependence in the joint distribution, two properties that are also observed in the market. We make use of the following model:
rν q rν rν q X = β · V + 1 − β2 · V = β V + 1 − β2 V , (2.19) i i g i g i g i i i
2 where V and Vi are independent N (0, 1), and g is an independent χ (ν) random vari- q ν able that is common to both the systematic process Z = g V and the idiosyncratic q ν process i = g Vi. In order to prevent infinite low-order moments of X, we impose the restriction that ν > 2. Note that the factor Z is now two-dimensional and has a bivariate distribution func- q ν tion. The asset variable Xi, though, reads Xi = g Yi with Yi ∼ N (0, 1) which amounts to a Student-t distribution with ν degrees of freedom, Xi ∼ tν, as desired. Also, recall that with Y being a N-dimensional standard Gaussian variable with 2 q ν correlation matrix ρ, and g a scalar χ (ν) random variable, X = g Y follows a
12 2 Modeling of Multivariate Default Risk 2.3 Factor Models
N-dimensional Student-t distribution6 with correlation matrix ρ 7 and ν degrees of freedom. Thus, we have in fact constructed a Student-t copula. For the conditional default probability we now obtain
p g ! ν Ci(t) − βiV pi(t|V, g) = Φ , (2.20) p 2 1 − βi
−1 where the calibration equation reads Ci(t) = tν (pi(t)). Here, the strength of the model chosen above becomes apparent: if we had chosen two independent χ2 dis- tributed variables for the systematic and the idiosyncratic factor, then Xi would have been a sum of two Student-t variables whose distribution cannot be expressed in closed form. The calibration of C would therefore have required a numerical root finding in contrast to the analytic expression above. On the downside, the model above requires a double integral in Eq. (2.13) over both factor components g and V .
2.3.3 The Portfolio Loss Distribution
From the conditional default probabilities, the conditional loss distribution P (w; t|Z) can be calculated.
Let us first consider the case of a homogeneous setup, i.e. li ≡ l, pi ≡ p, and βi ≡ β are identical for all obligors. It follows that the loss distribution (conditional and unconditional) is related to the distribution of the number K of defaulted names by the transformation P (w; t) = P(K = w/l; t), i.e. we only have to consider the density of the number of defaults. Given the conditional independence of defaults, the density is given by the binomial density function,
N (K = n; t|Z) = p(t|Z)n (1 − p(t|Z))N−n = b(n; N, p(t|Z)). (2.21) P n
In the more general inhomogeneous setup, the conditional loss distribution is given by the convolution product of all conditional single name loss distributions 8. This convolution can be computed by established methods for the calculation of continuous
6The generalization of the univariate Student-t distribution to a multivariate distribution is not unique. We use the definition used in Ref. [7]. 7Hence, the construction in Eq. 2.19 preserves the Gaussian correlation structure, i.e. as before we get ρ(Xi,Xj) = βiβj, for the derivation see Appendix B. This means that the factor loadings β have the same meaning as before and form a decomposition of the correlation matrix. 8The distribution of the sum X + Y of two independent random variables with respective dis- ∞ P R 0 tributions GX (x) and GY (x) is given by the convolution G (x) = x0=−∞ GX (x)GY (x − x). The distribution of X1 + ... + Xn, Xi independent, can be computed from this by induction.
13 2.3 Factor Models 2 Modeling of Multivariate Default Risk
convolutions of independent random variables, see Ref. [20] for a Fourier transform based method. A particularly simple and efficient method [7] is based on recursion: a loss unit u is
introduced, such that the individual losses li are discretized in integer units wi of u, 9 li = wi u . If Ln is the loss associated with a portfolio containing the first n names
(in arbitrary order), then Pn(w; t|Z) ≡ P(Ln(t) = wu|Z) is recursively given by
Pn+1(w; t|Z) = Pn(w − wn+1; t|Z) pn+1(t|Z) + Pn(w; t|Z)(1 − pn+1(t|Z)). (2.22)
From this, the loss distribution P (w; t|Z) = PN (w; t|Z) can be built, starting from
the boundary case of the empty portfolio, P0(w; t|Z) = δw,0. Note that Eq. (2.22) corresponds to a nested loop: new names are successively added to the portfolio and for each new name the probabilities for all possible losses
w = 0, . . . , wmax,n + wn+1 need to be computed. Therefore, the computational time of numerical implementations will typically depend roughly quadratically on the size of the portfolio (cp. Sec. 5.3).
2.3.4 Fitting the Factor Loadings
Suppose we have inferred a correlation matrix ρ for the reference portfolio from the market, e.g. from asset correlations (cp. Chapter 3). In general, the factor loadings cannot be chosen such that the correlation matrix is exactly reproduced, since ρ has (N − 1)(N − 2)/2 independent entries whereas there are only N factor loadings for each factor. Therefore, one aims to find factor loadings β that approximate ρ best with a given number of factors M. This optimization problem can be numerically solved with a number of methods, an iterative algorithm based on principal component analysis (PCA) is sketched in Ref. [7]. From the spectral decomposition of the correlation matrix, ρ = V ΛV T , only the √ √ eigenvectors of the M largest eigenvalues are used, β 1 = λ1v1, . . . ,ββM = λ1vM (where the eigenvalues are labeled in decreasing order of magnitude), yielding an PM T approximation ρ˜ to ρ, ρ˜ = i=1 β i β i .
In general, the factor loadings β i are not optimal yet, since the optimization procedure above minimizes tr ρ −ρ˜ ρ −ρ˜T , i.e. it uses a Frobenius norm on all elements of ρ − ρ˜. However, the diagonal elements of ρ don’t need to be approximated, since the
9See Ref. [7] for a discussion about the optimal size of the loss unit in order to keep rounding errors at the one hand and computational effort at the other hand at a tolerable level.
14 2 Modeling of Multivariate Default Risk 2.3 Factor Models
construction of the factor models (2.15) assures that ρ(Xi,Xi) = 1, and we only need to consider non-diagonal elements of ρ. Therefore, a modified correlation matrix (2) ρij , for i 6= j ρij = (2.23) ρ˜ij , for i = j
(2) is approximated in the next iteration, yielding β i , etc. In this thesis we will follow a different approach, though, as we will construct factor loadings that build a model correlation matrix with some predefined structure, see Chapter 4.
2.3.5 Large Homogeneous Portfolio Approximation
For an asymptotically large, homogeneous portfolio it is possible to derive a closed- form solution for the loss distribution. Specifically, we make the following assump- tions: • The portfolio consists of an infinite number of entities N.
• The notional amounts, single name default probabilities, recovery rates, and pair-wise correlations are identical for all entities.
• There is a single factor determining the conditional default probabilities. This is referred to as the Large Homogeneous Portfolio (LHP) approximation. We now K consider the average number of defaulted obligors N and use the fact that conditional on the factor Z defaults are independent and follow the law of large numbers, K = p(Z)|Z = 1, (2.24) P N i.e. the average number of defaulted obligors equals the conditional default probability almost surely. As above, we obtain the unconditional (now continuous) distribution by integrating out the factor, K Z ∞ K F (α) = P ≤ α = P ≤ α|Z = z dGZ (z) N −∞ N Z ∞ = P(p(Z) ≤ α|Z = z) dGZ (z). (2.25) −∞ p(z) is not a random variable but a deterministic function and we can restrict the integration domain to the region where p(z) ≤ α, i.e. where z ≥ p−1(α) (for p(z) is a decreasing function), Z ∞ −1 F (α) = dGZ (z) = 1 − GZ (p (α)). (2.26) p−1(α)
15 2.3 Factor Models 2 Modeling of Multivariate Default Risk
2.3.5.1 Gaussian Copula
For the Gaussian copula, Eq. (2.26) evaluates to ! ! C − p1 − β2 Φ−1(α) p1 − β2 Φ−1(α) − C F (α) = 1 − Φ = Φ , (2.27) β β where we have used the symmetry of the Gaussian density function in the last step. The density function reads d 1 f(α) = F (α) = −g (p−1(α)) dα Z p 0(Z(α)) ! p1 − β2 Φ−1(α) − C p1 − β2 = ϕ , (2.28) β βϕ Φ−1(α) where ϕ is the density of the standard normal distribution.
2.3.5.2 Student-t Copula
As pointed out above, the conditional default probability (2.20) in the Student-t copula model differs from the Gaussian one by being conditioned on two independent variables, thereby failing the conditions for the LHP approximation stated above. This issue is resolved in Ref. [35] by introducing a new factor variable η,
rg η := C − βV, (2.29) ν thus transforming Eq. (2.20) into ! η p(η) = Φ . (2.30) p1 − β2
Eq. (2.26) then turns into
Z p−1(α) −1 F (α) = dGη(η) = Gη(p (α)), (2.31) −∞ since p(η) is an increasing function.
It is shown in Ref. [35], that the density function gη can be written in a closed form as finite sum over incomplete gamma functions 10,
2 ν−1 − νt X ν − 1 g (t) = A e 2(C2+νβ2) (Bt)ν−1−kI (−Bt), (2.32) η k k k=0
10Assuming that ν ∈ N and p < 50% [35].
16 2 Modeling of Multivariate Default Risk 2.3 Factor Models
where for each k ∈ N0 the function Ik : R 7→ R is given by
( k−1 k+1 k+1 x2 2 2 Γ 2 Γ 2 , 2 , for x ≥ 0 Ik(x) = (2.33) k Ik(0) + (−1) (Ik(0) − Ik(−x)) , for x < 0
1 R ∞ −t k−1 with Γ(k, x) = Γ(k) x e t dt, and the constants A and B are defined as
ν ν−1 ν 2 β C2+νβ2 C A := ,B := . (2.34) √ ν−1 ν p 2 2 2 π 2 Γ 2 β C + νβ This can be used to express the density function of the fraction of defaulted obligors as 1 f(α) = g (p−1(α)) η p 0(η(α)) p p 1 − β2 = g 1 − β2 Φ−1(α) . (2.35) η ϕ Φ−1(α)
17 Chapter 3
Default Correlation
As pointed out in the previous chapters, the valuation of CDO tranches involves a number of parameters: the reference portfolio is described by single-name credit spreads, recovery rates, and losses, as well as the correlation between the obligors; for the CDO being a cash-flow instrument, its valuation naturally also requires risk-free interest rates and accrual factors. The correlation, though, stands out from this list as it is the only parameter specific to CDOs (or other basket credit derivatives), whereas the others can be observed from the liquid CDS and interest rate markets.
Empirical investigations suggest that defaults tend to cluster together and that there is positive correlation even between companies in different sectors; therefore, the as- sumption of uncorrelated defaults is not justified and correlation has to be considered. Moreover, the strength of the correlation crucially affects the prices and risk profiles of the CDO tranches. The determination of the correlation therefore receives particular interest and can be approached from two opposite directions.
One can consider default correlation as model input that needs to be estimated ex- ternally, to the best of one’s knowledge. This is complicated by the fact that there is little empirical data on the relationship between defaults, and the underlying process driving the defaults is not easily observable. In theory, information about default cor- relation should be extracted from historical defaults; this generally suffers from the low number of observed defaults, though. For the estimation of the default correlation one therefore typically resorts to the correlation of historical equity market returns which is used as proxy for the asset return correlation of latent variable models like the Merton model (2.3).
18 3 Default Correlation 3.1 Implied Correlation
3.1 Implied Correlation
Conversely, one can think of correlation as a quotation convention. With the advent of standardized STCDOs on liquid CDS indices and the resulting increased liquidity in the CDO market, prices can be expected to follow from supply and demand rather than from a theoretical model. The correlation implied from these prices can be used as a quotation device to facilitate a comparison of prices across tranches and products, the line of thought being similar to implied volatilities of equity markets 1. Given the developments in the standardization of the credit derivative markets, there is also a need for a commonly agreed method of quoting this implied correlation. As standard model, the one-factor Gaussian copula with a flat correlation, βi ≡ β, ρij ≡ ρ ∀i, j is widely used. However, like the Block-Scholes model in equity markets, the one-factor Gaussian copula is known to be too simplistic to reflect market reality. It can therefore not be expected that a single correlation number is sufficient to fit the prices of all tranches simultaneously, but rather that each tranche trades at its own implied correlation. In the following, we describe two popular methods that have been developed for quoting the implied correlation with respect to the tranche.
3.1.1 Compound Correlation
Compound correlation is the direct extension of the Black-Scholes implied volatility and was the quotation convention initially used in the market. By inverting the standard pricing model, it is possible to find the level of default correlation that equates the observed market spread and the theoretical spread for a tranche. The (compound) correlation affects the different tranches in different ways. In general, an increase in default correlation means that it becomes more likely to observe many or few defaults. An equity tranche, for instance, is not much affected by the occurrence of many defaults as it absorbs losses only up to its detachment point. However, the occurrence of only a few defaults reduces the expected loss in the tranche and therefore the fair spread. Thus, in terms of prices, an equity tranche investor is long correlation. Conversely, if defaults in the portfolio are completely uncorrelated, the probability of the occurrence of many losses is very low, i.e. the losses are very unlikely to hit the senior tranche. Therefore, senior tranche investors are short correlation. Mezzanine tranches, finally, are in general not monotonic in (compound) correlation: a long position in a mezzanine tranche with attachment point aL and detachment 1However, market prices may be observed that are not attainable by any choice of correlation.
19 3.1 Implied Correlation 3 Default Correlation
point aH can be seen as a long position in an equity tranche with detachment point aH plus a short position in an equity tranche with detachment point aL, since the expected loss can be decomposed into
E [LaL,aH ] = E [L0,aH ] − E [L0,aL ]. (3.1)
Since both ”virtual” tranches are long correlation but enter with opposite sign, the root finding algorithm for the determination of the compound correlation often doesn’t yield a unique compound correlation but two correlations that produce the same tranche PV, see Sec. 5.2 for an example. This ambiguity is one of the biggest drawbacks of compound correlation. Moreover, as we will discuss in Sec. 3.2, compound correlation is not easy to extend to the pricing of tranches with non-standard strikes. The upside of compound correlation is its transparency and intuitive meaning as it directly maps to the asset correlation of latent variable models.
3.1.2 Base Correlation
Base correlation has been developed by JP Morgan [32] and makes use of the mono- tonicity of equity tranches. The basic idea of base correlation is the decomposition of all tranches into combinations of (virtual, non-traded) base (=equity) tranches, as given in Eq. (3.1). In contrast to compound correlation, where the two virtual base tranches are priced at the same correlation, base correlation allows for pricing each base tranche at a different correlation. This is done via a bootstrapping mech- anism where the base correlation from the first tranche is used to solve for the base correlation of the second one, and so on:
1. Let Va−b(s, ρ) denote the PV (from the investor’s perspective) of the tranche with attachment point a and detachment point b, with spread s, and (flat)
correlation ρ. The base correlation of the equity tranche, ρ0−K1 , solves the
equation ! V0−K1 (s0−K1 , ρ0−K1 ) = 0, (3.2)
where s0−K1 is the market spread for the 0−K1 tranche (i.e. the equity tranche). This is equivalent to the determination of the compound correlation for the equity tranche; hence, base correlation and compound correlation are identical for the equity tranche.
20 3 Default Correlation 3.1 Implied Correlation
2. The next tranche K1 − K2 is equivalent to a portfolio of a long position in a
0−K2 base tranche and a short position in a 0−K1 base tranche, both using the
K1 − K2 market spread. We define the base correlation ρ0−K2 as the correlation that solves ! VK1−K2 (sK1−K2 , ρK1−K2 ) = V0−K2 (sK1−K2 , ρ0−K2 ) − V0−K1 (sK1−K2 , ρ0−K1 ) . | {z } | {z } =0 by definition of ρK1−K2 known, <0 since sK1−K2 3. The procedure continues in the same way for the higher tranches. Since the PV of the 0 − K2 base tranche, V0−K2 (sK1−K2 , ρ0−K2 ), is the only changing component in Eq. (3.3) (the base correlation of the lower strike, ρ0−K1 , is fixed during the root finding) and is an increasing function of the base correlation ρ0−K2 , the base correlation is guaranteed to be unique. Moreover, by means of Eq. (3.3), the existence of base correlations for all tranches guarantees that Eq. (3.1) holds, which is a basic arbitrage-free requirement. With respect to the whole tranche structure this means that the sum of the protection legs of all tranches equals the sum of the protection legs of all underlying CDS. This is also the reason that sometimes base correlation fails to find a solution for the super senior tranche: since the expected loss (and hence the spread) over the whole tranche structure doesn’t depend on the correlation (cp. Eq. (2.2)), a market spread that only slightly deviates from this spread cannot be reproduced anymore by any base correlation value. However, there are some pitfalls with base correlation, too, the most obvious being the difficulty to build intuition about it. Since every tranche is actually characterized by two correlations (one for the lower, one for the upper strike), it is not obvious how base correlation changes with default or asset correlation. In fact, an increased default correlation can result in the decrease of the base correlation for some tranches [37]. One also needs to keep in mind that the determination of the base correlation for a particular tranche requires the knowledge about the prices for all more junior tranches and that base correlation is only unique given the set of attachment points [37]. The base correlation of the a − b tranche thus depends not only on a and b but also on the position of all prior attachment points, making the comparison e.g. between the European and American markets difficult due to the different tranche structure. 21 3.2 Correlation Smile 3 Default Correlation 3.2 Correlation Smile As pointed out above, the implied correlation is typically not constant throughout the tranches but is known to exhibit a ”smile” (in terms of compound correlation) or a skew (in terms of base correlation), reflecting the difference between real world pricing and the assumptions of the (simplistic) Gaussian copula model, see Fig. 3.1. The main features of the smile/skew are relatively high spreads for equity and senior tranches (resulting in a low implied compound correlation for equity tranches and a high correlation for senior tranches) and low spreads for mezzanine tranches. Figure 3.1: Example of a correlation smile/skew observed in the market (DJ iTraxx, 13 April 2006). Figure was taken from Ref. [15]. The reasons for the shape of the smile/skew are manifold, including model effects and supply-demand conditions. On the model side, it is obvious that the single correlation number isn’t able to capture the dependency structure of a portfolio of 125 credits; also, the Gaussian copula model does not reflect the empirical joint distribution of default times as it underestimates the probability of a very low or very high number of defaults. On the supply-demand side it is worth noting that mezzanine tranches are very popular among investors, resulting in a tightening of spreads. On the other hand, the CDO market is segmented, i.e. different parts of the seniority structure are driven by different investors (hedge funds buying equity tranches, retail investors looking for mezzanine tranches, and pension funds focusing on senior tranches), and there are only few market participants that seek to exploit perceived mispricing across different tranches. A number of other effects have been suggested as explanation, see e.g. Refs. [4, 32]. 22 3 Default Correlation 3.2 Correlation Smile An important application of the correlation smile/skew is the interpolation in order to calculate spreads for customized tranches on standard indices with non-standard attachment points. In this context, the shape of the base correlation skew, which is typically close to linear, appears to be much better suited for an interpolation than the more complicated smile shape of compound correlation. However, as it is shown in Ref. [37], the spread errors from a linear interpolation of the base correlation skew can still be large. In rare cases, the interpolation can even yield negative expected losses for some tranches. 3.2.1 Modeling Approaches A starting point for modeling approaches is the loss distribution implied by the mar- ket. Since tranche prices are essentially integrals over the loss distribution, cp. Eqs. (1.2) and (1.3), the implied loss distribution can be backed out from the market by differentiating. Compared to the loss distribution generated from the Gaussian cop- ula, one observes a fat upper tail, indicating an increased probability of high loss scenarios (yielding a high spread for senior tranches), and a lower probability of zero (or few) defaults (equivalent to a high spread for the equity tranche), see Fig. 3.2. Whereas the fat upper tail can be modeled e.g. by switching from the Gaussian to a Student-t copula, the challenge is to do so without also raising the probability of small losses. 50% 100% un-skewed un-skewed skewed skewed Prob P(L 0% 0% 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 Normalised Loss Normalised Loss Figure 3.2: Distribution of losses, as produced by a Gaussian copula (”un-skewed”) and observed in the market (”skewed”). Figure was taken from Ref. [24]. A number of alternative copulas have been proposed to overcome this issue, including Double-t, Clayton, Archimedean, and Marshall-Olkin copulas, we refer to Ref. [11] for an overview. Also extensions to the Gaussian copula model have been investi- gated: stochastic correlation interprets the correlation as (binary) stochastic variable 23 3.2 Correlation Smile 3 Default Correlation [10], which –in a further extension termed ”random factor loadings” [6]– possibly de- pends on the factor Z (low correlation in bullish markets, high correlation in bearish markets). Random recovery [6] in turn treats recovery rates stochastically, following e.g. a Gaussian, Student-t, or beta distribution. 24 Chapter 4 Modeling the Correlation Matrix In the previous chapter, a number of modeling techniques for the correlation smile were mentioned. In this work, we want to pursue a different approach, namely by considering the full matrix of pairwise correlations between the obligors instead of a constant and equal correlation between all names. As an example, the left panel of Fig. 4.1 shows the implied compound and base correlation for a non-flat correlation matrix whose structure is indicated in the right panel. Apparently, the correlation seen by the individual tranches differs, giving rise to a non-flat implied correlation curve. 40 Compound correlation Base correlation ρ=0.55 ρ=0.05 25 30 ) % ( n 50 o i t r a l o e 20 g r i r l o b C O 75 d e i l p m I 10 100 0 125 0 5 10 15 20 25 25 50 75 100 125 Detachment point (%) Obligor Figure 4.1: Implied compound and base correlation curve (left panel) for the block structure of the correlation matrix indicated in the right panel. Used parameters: see Tab. C.2 in Appendix C. Within the factor model framework, this approach leaves us with N ·d free parameters (the factor loadings), with d being the number of factors. The fitting of the correlation 25 4.1 Empirical Correlations and RMT 4 Modeling the Correlation Matrix smile thus turns into a high-dimensional optimization problem which was tackled in Ref. [23] by using Evolutionary Algorithms. In contrast, the idea of this work is to impose a structure for the correlation matrix in order to drastically reduce the number of fitting parameters while still keeping enough degrees of freedom to allow for rich variations of the implied correlation curve. To this end, we will let ourselves guide by the structure of empirical correlation matrices obtained from the market. 4.1 Empirical Correlations and RMT As mentioned in Chapter 3, the correlation of equity returns is often used as a proxy for default correlation. Denoting the equity return (price change) of name i by δxi(t) and assuming that the δx’s have been rescaled to constant unit volatility, the corre- lation reads T 1 X 1 C = δx (t) δx (t) = MMMM T , (4.1) ij T i j T t=1 where M is a N × T matrix (N being the number of assets), representing the N time series of price changes of length T . The period T is in practice always limited, resulting in empirical correlations that are to a large degree noisy (i.e. random). Random matrix theory [13] (RMT) makes predictions about universal properties of (real and symmetric) random matrices, in particular spectral properties. Deviations from these universal predictions then iden- tify non-random, system specific properties. The density ρA(λ) of eigenvalues of a matrix A is defined as N 1 dn(λ) X ρ (λ) = = δ(λ − λ ), (4.2) A N dλ i i where n(λ) is the number of eigenvalues of A smaller than λ. Assuming that the matrix M from above is completely random, i.e. its elements are iid random with arbitrary distribution, ρC (λ) is universal and exactly known in the limit N,T → ∞, with a fixed ratio Q = T/N ≥ 1 and reads [31] Q p(λ − λ)(λ − λ ) ρ (λ) = max min with C 2πσ2 λ (4.3) max 2 p λmin = σ (1 + 1/Q ± 2 1/Q), 26 4 Modeling the Correlation Matrix 4.1 Empirical Correlations and RMT 2 with λ ∈ [λmin, λmax], and σ the variance of the elements of M . Eq. (4.3) is called Marˇcenko-Pastur law [31]; an important prediction besides the functional form of the 1 spectral density is that ρC vanishes outside the interval [λmin, λmax]. The investigation of empirical correlation matrices of stock price returns in various markets in Refs. [27, 34] now reveals that the spectrum obtained from these empirical matrices shows – within the Marˇcenko-Pastur bounds λmin, λmax– a good agreement with the ”pure noise” hypothesis as expressed in Eq. (4.3). However, ≈ 6% of the eigenvalues exceed the upper bound significantly, in particular the highest eigenvalue with a value of λ1 ≈ 25 · λmax. It is shown in Ref. [27] that the corresponding eigenvector has almost equal components on all names and thus can be identified with the market. Ref. [34] elaborates further on the structure of the eigenvectors corresponding to the other large eigenvalues and shows that subsets (groups) of names of different sizes contribute to them. To conclude the most important findings from the investigation of empirical correla- tion matrices of equity returns, we assert that: • Only a small number of (large) eigenvalues deviates from the Marˇcenko-Pastur law (4.3) and thus carries some information. 2 • The highest eigenvalue is much larger than all other eigenvalues and can be identified with the market, i.e. there is a common systematic risk factor to which all names are significantly exposed to. • The correlations that account for the other large eigenvalues are neither com- pletely localized (i.e. only few names contribute to them), nor very extended. Instead, they span groups of names of different sizes. In the following we discuss the task of constructing correlation matrices with these properties in a factor model framework. 1For finite N, the edges of the interval are smoothed, i.e. there is a small probability to find eigenvalues outside the interval. We will not account for finite N effects here. 2However, care must be taken here: the Marˇcenko-Pastur law applies to independent sample elements whereas the correlation matrix is only a measure of linear dependence. If sampled from heavily tailed distributions, e.g. a multivariate Student-t distribution, even uncorrelated random data may lead to an eigenspectrum that doesn’t correspond to the Marˇcenko-Pastur law. Hence, the number of eigenvalues exceeding the upper bound might not give a correct indication of the number of real driving factors [17]. 27 4.2 Group Models 4 Modeling the Correlation Matrix 4.2 Group Models It is already intuitively clear that there is a close connection between the eigenvectors of the correlation matrix and the factor loading vectors in the factor framework since both are forming a decomposition of the correlation matrix. The case of the com- pletely random correlation matrix reflected by the Marˇcenko-Pastur law obviously corresponds to the zero-factor model, i.e. there is no systematic factor in Eq. (2.15). The correlation matrix in Eq. (2.16) then equals the identity matrix. Note that this is the model correlation matrix of the zero-factor model, i.e. without noise dressing. By sampling finite time series’ from the zero-factor model and computing the corre- lation matrix from the sample, one would retrieve the noise dressed matrix with the Marˇcenko-Pastur spectrum. Going to the one-factor model, the model spectrum exhibits a large eigenvector as- sociated with the systematic variable (the market): for simplicity we assume equal factor loadings for all names, βi ≡ β, so that the correlation matrix can be written as a weighted sum of the identity matrix I N and the unit matrix (i.e. a matrix consisting of all ones) J N , 1 β2 ... 2 2 2 C = β 1 = (1 − β )I N + β J N , (4.4) . . . .. and the characteristic equation for C reads