Factor Modeling in Finance
I discuss multi-factor modeling, which generalizes many early financial models into a common prediction and risk framework.
In the 1950s, Harry Markowitz developed modern portfolio theory (Markowitz, 1952). In particular, modern portfolio theory introduced the idea of a risk–reward trade-off between a portfolio’s expected return and its volatility. In its framing, investors want portfolios with high expected returns and low volatility in actual outcomes. This framework is sometimes called mean–variance analysis. However, Markowitz did not define “risk” beyond the notion of portfolio volatility.
Subsequently, several researchers (Sharpe, 1964; Mossin, 1966; Lintner, 1965; Treynor, 1961) extended mean–variance analysis one step further to argue that there is a single explanatory variable, often called a “factor” in both statistics and finance, that explains expected returns: the market. This model became known as the capital asset pricing model (CAPM). The CAPM argues that the only real risk is the market, and so the only real factor is exposure to the market. Formally, let and be random variables denoting the return of asset and the market , and let be the non-random risk-free rate of return. Then the CAPM states,
In words, the expected excess return of asset is a linear function of the expected excess return of the market. The linear relationship in Equation can be derived directly from the mean–variance analysis framework, particularly from linear efficient frontier. See my post on the CAPM for this derivation. The CAPM is useful because it is both simple and intuitive. There is a single factor, the market, and a single exposure to that factor, , which captures an asset’s sensitivity to the market. Higher (lower) means that when the market moves up or down, the return for asset moves up or down faster (slower).
However, empirical evidence has not agreed with the CAPM (Fama & French, 2004). To quote (Chamberlain & Rothschild, 1982),
Few believe that asset returns are well described by their first two moments or that some well-defined set of marketable assets contains most of the investment opportunities available to individual investors. Casual observation is sufficient to refute one of the main implications of the CAPM—that everyone holds the market portfolio.
Subsequent economists have proposed many multi-factor models, meaning models with multiple explanatory variables. An early and famous multi-factor model, for example, is the Fama–French three-factor model (Fama & French, 1993). Formally, let “SMB” denote a “small minus big” factor where size is measured via market capitalization, and let “HML” denote a “high minus low” factor, meaning the book-to-market ratio (book or accounting value to market value). Then the Fama–French three-factor model is
Of course, one can extend this logic to any number of macroeconomic indicators, and so there are many other multi-factor models. As an aside, the factors are not indexed by . The interpretation of this is that the factors are macroeconomic in nature: exposure to the market, size relative to the market, relative value, interest rates, and so on. Each asset is exposed to or loads on these factors in some quantity, defined by the parameters , which are indexed by .
A natural extension to multi-factor models with factors specified by some modeler would be a model with latent or unobservable factors. In other words, rather than specifying the factors in advance, can we use multivariate statistics to simply infer the factors? This is precisely what arbitrage pricing theory (APT) does (Ross, 1976). Formally, let denote the number of latent factors, let denote the -th systematic factor that is common to all assets, and let denote the asset’s sensitivity or factor loading onto the -th factor. Finally, let denote a linear model’s intercept, and let denote white noise. Then APT models risky asset returns as
APT is quite a general, and at this point I think it would be quite clear to a statistician where all this is going: a general framework for multi-factor modeling that looks a lot like linear-Gaussian models such as factor analysis or principal components analysis (PCA).
Multi-factor models
In finance, factor modeling exploits standard methods from multivariate statistics to model returns, variances, and correlations (Rosenberg & McKibben, 1973). To reiterate, a factor is an explanatory variable, and these factors can be observable or unobservable.
Let be the number of assets, be the number of time periods, and be the number of factors. Ideally, . In words, there are far fewer macroeconomic factors that explain asset returns than their are unique assets. Let and denote the return and white noise of asset at time . Let be a -vector of factors and let be a -vector of “loadings” or exposures of asset onto this factor at time . Note that the factors are common across assets (no index ), while the loadings are common across time (no index ). Then the multi-factor model is
As with linear regression, we can push the intercept into the dot product by adding a dummy factor . In finance, can be interpreted as the idiosyncratic return, or the return of asset at time that is not shared across assets or time. Note that all other parts of the return in Equation are common to all assets via the factors.
We assume that the idiosyncratic return has zero mean, , and that these specific returns are uncorrelated with each other,
Furthermore, we assume that the factors are zero mean, , and are uncorrelated with the idiosyncratic return . This means that their covariance is zero:
Since , this implies that the cross term must be zero as well, that
Finally, let denote the covariance of . Since is zero mean, this implies
Ex ante, is a random variable. These assumptions induce some probability distribution onto , and we can derive the first and second moments. The first moment of is:
So returns are zero mean if the factors are zero mean. The covariance between returns is
We cannot simplify without some assumptions. Typically, we restrict , so that this term is the covariance of the factors at time . This is called a cross-sectional analysis, where the “cross-section” is the slice of all assets at time .
As we will see, Equation is a big leap forward, as it explicitly represents the riskiness of two assets through their idiosyncratic risk and their exposure to the factors’ risk. What this suggests is that not only are factors useful for predicting asset returns, they are useful for modeling portfolio volatility.
Inference
What parameters are we actually estimating in a multi-factor model? It depends on how we frame the problem. Let’s look at several scenarios.
Classic (implicit) factor model
In an implicit factor model, we assume the factor loadings are known and seek to estimate the unknown (implicit) factors. This approach uses a cross-sectional setup and induces the classic formulation of factor analysis. In a cross-sectional analysis, data are grouped by time period, meaning we care about differences between assets within each time period. Thus, we can rewrite Equation in vector form as
where
And we can represent the idiosyncratic return in vector-form as
While is indexed by , is not. This is a cross-sectional assumption: at each time period, the covariance of the noise does not change. Since is a diagonal matrix, the error terms are uncorrelated across assets.
Equations and above can be written in vector form as
Typically, is assumed to be a diagonal matrix. See A1 for derivations.
Finally, to estimate the factors, , we fit linear regressions using Equation .
Time-series-based (explicit) factor model
In an explicit factor model, we assume the factors are known a priori (explicit) and seek to estimate the unknown factor loadings. In some sense, this is the most intuitive setup and the one most analogous to the models in the introduction. For example, in the Fama–French three factor model, we know the three factors a priori, and estimating the loadings is equivalent to estimating the parameters in Equation but for all assets .
This requires a slightly different setup, a time-series regression. Instead of Equation , we represent the problem as
Each of the columns of is a -vector . This represents the -th factor varying across time. This is a time-series regression rather than a cross-sectional regression because now the independent variable is the returns for a single asset across time. Any distributional assumption on is an assumption about this time series. The error terms still have spherical errors but w.r.t. time:
Now we fit linear regressions using Equation to estimate for each asset. Here, the real goal is to assess the goodness-of-fit of the model, assuming the factors . If the model has a high coefficient of determination or if the estimated coefficients are statistically significant, then this suggests that the investor has selected useful factors.
Statistical analysis
In the final approach, we assume both the factors and the loadings are unknown. To estimate both quantities jointly, we use standard methods in multivariate statistics such as factor analysis or PCA. Formally, factor analysis is the cross-sectional model defined in Equation , while probabilistic PCA is identical to factor analysis except the idiosyncratic returns have a common variance, i.e.
Inference for factor analysis typically requires assuming that is multivariate normally distributed, which induces a multivariate normal assumption on the returns:
This assumption is not too unreasonable if we assume is a vector of log returns rather than raw returns. We can then write down the log likelihood,
and use maximum likelihood estimation or expectation–maximization (EM) to infer the parameters and . We can then estimate the factors from the inferred parameters.
Alternatively, we can use PCA. In the PCA-based approach, we first compute the sample covariance matrix of the returns, . Then we take as the factors the first eigenvectors corresponding to the largest eigenvalues of .
See my previous post on factor analysis for details on fitting factor analysis using EM. Alternatively, see (Tipping & Bishop, 1999) for a discussion of probabilistic PCA.
Risk-factor modeling
So far, we have viewed factors as useful macroeconomic indicators that are correlated with or predictive of asset returns. However, if a factor predicts a return, it is natural to think of it as a risk-factor as well. What do I mean? Recall that in the mean–variance analysis framework, the objective is to maximize our portfolio’s expected return while minimizing its variance. See my post on mean–variance analysis if this claim does not make sense. Formally, if is an -vector of portfolio weights, then the unconstrained objective is:
where now is an -vector of assets in a portfolio at time and where is the covariance of those assets at time . We might add constraints such as the weights summing to unity, but the essence of the problem is to maximize the returns of the positions we take and to minimize the risk of those positions, as captured by their variances and covariances.
However, is an matrix, which can be quite large and quite sparse. Think about how many stocks there are, for example, and how the number of available stocks changes across time. To compute the optimal , an optimizer may supply many values for and then compute many times. In risk-factor modeling, we replace with the decomposition in Equation :
In fact, we should do this if we believe Equation is true, since that equation states that is really a linear combination of factors. This allows us to rewrite the objective in Equation as
Notice that is a diagonal matrix of idiosyncratic variances, so the right-most term in Equation can be computed with a dot product and scalar-vector multiplication. And typically we assume that the factors are uncorrelated, meaning that is a diagonal matrix as well. Since , computing is much faster than computing . Furthermore, this decomposition elegantly handles the sparse matrix . Rather than estimating the covariances between all combinations of assets, we model asset correlations through their loadings onto common factors.
Conclusion
In finance, multi-factor modeling generalizes many early financial models into a common framework, where returns are linear functions of macroeconomic explanatory variables, called “factors”. In this framework, we can either assume we know the factors or the loadings but not both, in which case inferring the other amounts to a multivariate linear regression. Or we can use standard methods from multivariate statistics, such as factor analysis or PCA, to infer both the factors and loadings jointly. In a cross-sectional analysis, the natural extension of these ideas to model risk (portfolio variances and covariances) through the low-rank approximation induced by the factor model.
Appendix
A1. Unconditional moments
The unconditional mean is zero, since by assumption:
The unconditional variance is
The cross-terms are zero because and are mean-zero and uncorrelated, so:
- Markowitz, H. (1952). Portfolio selection. Journal of Finance.
- Mossin, J. (1966). Equilibrium in a capital asset market. Econometrica: Journal of the Econometric Society, 768–783.
- Lintner, J. (1965). The valuation of risk assets and the selection of risky investments in stock portfolios and capital budgets. The Review of Economics and Statistics, 222–224.
- Treynor, J. L. (1961). Market value, time, and risk. Time, and Risk (August 8, 1961).
- Fama, E. F., & French, K. R. (2004). The capital asset pricing model: Theory and evidence. Journal of Economic Perspectives, 18(3), 25–46.
- Chamberlain, G., & Rothschild, M. (1982). Arbitrage, factor structure, and mean-variance analysis on large asset markets. National Bureau of Economic Research Cambridge, Mass., USA.
- Fama, E. F., & French, K. R. (1993). Common risk factors in the returns on stocks and bonds. Journal of Financial Economics, 33(1), 3–56.
- Ross, S. A. (1976). The arbitrage theory of capital asset pricing. In Handbook of the fundamentals of financial decision making: Part I (pp. 11–30). World Scientific.
- Rosenberg, B., & McKibben, W. (1973). The prediction of systematic and specific risk in common stocks. Journal of Financial and Quantitative Analysis, 8(2), 317–333.
- Tipping, M. E., & Bishop, C. M. (1999). Probabilistic principal component analysis. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 61(3), 611–622.