There’s a cycle in the finance world whereby good ideas become products, which sell well and spur demand for more good ideas. Current case in point: the cycle involving investable factors.
“Factors” is the catch-all term for the mechanisms that drive asset prices, and factors are being discovered almost as quickly as they can be packaged and sold to the waiting public.
The investment firms AQR Capital Management and Dimensional Fund Advisors are prominent factor users. The Maryland State Retirement and Pension System, a $52 billion plan that covers more than 400,000 workers and retirees, uses factor products. Many long/short equity hedge funds have incorporated factor hedging into their strategies, and the proliferation of low-cost exchange-traded funds and index mutual funds has made factor-style investing accessible to the retail market.
Fidelity, Vanguard, and BlackRock all offer online explainers of factors, and BlackRock has a head of factor-based strategies who proclaims on the company’s website that “factor investing is the way of the future.” “Institutional investors and active managers have been using factors to manage portfolios for decades,” reads BlackRock’s pitch. “Today, data and technology have democratized factor investing to give all investors access to these historically persistent drivers of return.” BlackRock’s site features an infographic that highlights 12 factors of importance, divided into two categories: macroeconomic (capturing broad risks across asset classes) and style (explaining returns in just one asset class).
And many dozens more are circulating, presumably keys to greater investment returns. Duke’s Campbell R. Harvey, Texas A&M’s Yan Liu, and University of Oklahoma’s Heqing Zhu have identified more than 300 factors in academic literature. City University of Hong Kong’s Guanhao Feng (a recent graduate of Chicago Booth’s PhD Program), Yale’s Stefano Giglio, and Booth’s Dacheng Xiu have collected and investigated over 100 of them—ranging from employee growth to maximum daily return.
But has the hunt for investable factors gone too far? Feng, Giglio, and Xiu are suggesting that it has. On one hand, factors are helpful. If an investor wants to have a truly balanced portfolio, she should do more than make sure she owns both stocks and bonds, and in theory she can use factors to make sure her investments truly represent a diverse basket of assets whose returns are driven by different things.
But are there really 300 separate characteristics associated with higher asset returns, or only a handful of things really driving stock prices? The hundreds of factors that appear in academic research are based on many possible company characteristics, which could, in theory, capture many types of risk. However, Feng, Giglio, and Xiu are skeptical that all these factors are useful. Two portfolios that look different from each other—for example, one that overweighs small companies and another that is heavy on illiquid companies—could end up giving investors exposure to the same underlying risk. That is, many of these factors may be redundant. Is the factor industry an example of a good idea that has gone too far?
Before there were 300 factors, there was just one, kind of. The original factor—or the forerunner of factor analysis—is, in a sense, the capital asset pricing model. Largely developed by Stanford’s William F. Sharpe, who was awarded the Nobel Memorial Prize in Economic Sciences in 1990, the CAPM posits that investors are paid for both the time value of their money and the risk they assume in holding an asset. The model has helped investors determine how much compensation they should demand for holding a risky asset.
The formula is not reliably predictive. It generates an expected return on assets based in part on the beta of a security, or its covariance with the market. Some research on the CAPM has suggested that low-beta stocks tend to have higher returns than predicted, while higher-beta stocks do not live up to expectations. Still, some finance professionals refer to the CAPM as “the one-factor model,” where market beta is the sole explanatory factor of expected returns.
In 1992, expanding on the CAPM, Chicago Booth’s Eugene F. Fama and Dartmouth’s Kenneth R. French published a three-factor model, identifying two more things that generate returns: size and value. They observed that small-cap companies outperformed large-cap ones over time, and companies with low price-to-book ratios outperformed companies with higher price-to-book ratios.
Investors generally accepted the Fama-French three-factor model, yet it still seemed incomplete to many adherents of Fama’s efficient-markets hypothesis. If markets are efficient and assets priced accordingly, a return is the compensation an investor receives for taking on some risk by holding an asset. Factors capture this risk, so when investors take on exposures to certain factors in their portfolios, they earn a risk premium as compensation. But some mutual-fund managers consistently outperformed or underperformed market benchmarks. Were they taking on risks beyond just three? The natural inclination of an efficient-markets believer, when confronted with an outperforming manager, is to assume that the manager is taking on and being paid for unquantified risks that are driving the asset returns. In 1997, Mark Carhart, who founded Kepos Capital, published an article that suggested momentum as a fourth factor.
We’re not saying those 300 factors are fake. It may be true that some deliver significant risk premia for investors. But they could also be simply duplicating a few other important factors.
Between this point and 2016, researchers developed another 300 factors, some of which are based on the idea that investors’ biases and behaviors can affect asset prices. Take momentum, where the recent direction and slope of stock returns is used to predict a stock’s future returns. When behavioral economists contributed some factors, that caused some debate about whether some of those factors were still grounded in the efficient-markets hypothesis.
Even Fama and French contributed to the growing pile of factors. In 2013, Fama was awarded the Nobel Memorial Prize in Economic Sciences for his work on asset pricing. Less than two years later, he and French came out with research in which they suggest two more explanatory factors in stock returns: profitability (high-earnings-growth companies get higher returns) and investment (companies with high total asset growth typically generate lower-than-average returns).
Competition within academia is fueling a rush to discover or develop new factors for publication, says Fama. “You have thousands of academics in this area, all of whom need to come up with papers to publish, so the door is open wide.”
Close the barn door
But there’s now a notion that it’s time to partially close this door, or at least make sure researchers test new factors vigorously, using international stock market data and at least relating the intuition behind the factors to economic models, such as the dividend discount model.
Feng, Giglio, and Xiu suggest that many touted discoveries are mere subsets of the original three, or the more recent five, or perhaps a few other already-discovered factors. This redundancy can be tough for researchers to spot. Just as investors evaluate a mutual fund or hedge fund by comparing it to a benchmark index, researchers evaluate a new factor by comparing it to a benchmark factor model, to see if it produces an extra return. If researchers omit important factors from the model, new factors may appear to be useful, even though they may be simply replicating the ones that were omitted.
For example, size (as measured by a company’s market capitalization) is a long-accepted factor and component of stock returns. Small-capitalization stocks tend to outperform large-cap stocks over time. This suggests an investor who takes on extra risk by investing in smaller companies receives compensation for doing so.
But small-cap companies also tend to be less liquid than large-cap ones. So it’s reasonable to suggest that liquidity might also be a factor in returns, and our investor is also being compensated for holding less-liquid, and therefore riskier, assets. But in the case of stocks, these researchers hypothesize that liquidity may stand alone as a factor or it may, in fact, simply be the size factor with a different label.
“We’re not saying those 300 factors are fake,” says Xiu. “It may be true that some deliver significant risk premia for investors. But they could also be simply duplicating a few other important factors.” Xiu and his colleagues argue that many of these factors are either useless or account for return variations that have already been explained. “We think many of these newly proposed factors are redundant and are the result of somewhat arbitrary choices of existing factors as controls—one form of abusive data mining,” they write.
Torturing the data
Such mining may be prompted by the requirements of journals. Goodhart’s law, named after London School of Economics’ Charles Goodhart, says that “when a measure becomes a target, it is no longer a measure.” This could be what’s happening as factor hunters search too strenuously for factors that will adhere to the measures journals require for publication.
Specifically, journals require a new factor to have a probability value (P value) of less than 5 percent, meaning that a test shows that the factor will give off false positives fewer than 5 percent of the time. Since factor foragers can target the P value, they are susceptible, consciously or not, to data dredging and mining (presenting patterns within data as statistically significant without presenting a hypothesis to explain causality) or outright P-value hacking (setting up the sample to get the desired result).
“There is a saying that if you torture the data enough, eventually they will surrender,” says Xiu.
Most of the existing research papers that propose a new factor compare it to an existing benchmark, such as the Fama-French three-factor model, and find that their new factor is useful in explaining more variation in the expected return across stocks. But if researchers consistently compare their new factors to the same benchmarks, and not to other newly discovered factors, it’s possible they essentially keep “discovering” the same factor again and again but give it different names.
Alternatively, some researchers, instead of using the Fama-French factors as a benchmark, pick an arbitrary factor that makes their results look significant.
Xiu and his coresearchers instead propose using the entire set of existing factors (the factor zoo, as some finance academics have termed it) as a benchmark for newly proposed factors—a more robust way to test whether or not newly proposed factors really bring something new to the table. Thus, old factors could help determine if newly proposed factors are just old ones in fancy repackaging.
The researchers say the challenge of using the entire zoo is the curse of dimensionality, a well-known problem that could render standard methods used to evaluate new factors (such as a simple linear regression) infeasible to apply. To counteract this, they propose using a machine-learning method to select the best benchmark.
Using this method, they analyzed 99 factors, pitting factors discovered in 2010 and later against those published earlier to determine which newer factors are truly new. They find that a few of the newly proposed factors do seem to have some utility, notably the profitability and investment factors proposed by Fama and French, while many others fall short. Feng, Giglio, and Xiu find that 13 of the 99 factors hold some explanatory power, which is enough to keep the factor hunt going, but also a result that suggests using skepticism when assessing the hundreds of factors in the academic literature.
Developing more rigorous factors
Econometrics researchers have been testing and refining asset pricing models since the 1970s, but Feng, Giglio, and Xiu focus on newly proposed factors. They believe that the existing approach to finding factors will inevitably lead to mistakes or a proliferation of ideas that aren’t fully vetted.
“What we are trying to do is to propose a statistical methodology that can tame the factors or anomalies, and we suggest a more robust approach in trying to identify an important one,” says Xiu.
Is that new factor useful?
Click below for a visual translation of an equation that explains how to test a proposed factor’s value.
He says that the techniques they use to evaluate factor efficacy could be applied to mutual-fund and hedge-fund portfolios, so that investors would be able to quantify the sources of returns generated by their managers. For example, a portfolio manager investing in frontier markets might identify the risks associated with emerging economies as a factor that drives returns. However, if those frontier-market companies are also smaller than most, or undervalued, it might really be the size or value factors driving returns.
“There may be many different smart beta portfolios out there that seem distinct but are actually taking on the same risk,” says Xiu.
Another item of interest is whether or not a factor’s persistence is affected by its use. It could be that a factor’s wide adoption undermines its utility, as people crowd into previously unknown trades. Momentum, for example, is a well-established factor prone to such reversals. Price momentum can attract investment dollars that are abruptly withdrawn once the price of a stock has appreciated.
Feng, Giglio, and Xiu want to apply machine learning to asset pricing apart from taming factors or anomalies. After all, if the factor industry is one example of the finance profession cycling out of control, it probably isn’t the only one.