Careful computational analysis of 40 years of stock prices explains important anomaly
One of the key principles in asset pricing -- how we value everything from stocks and bonds to real estate -- is that investments with high risk should, on average, have high returns.
"If you take a lot of risk, you should expect to earn more for it," said Scott Murray, professor of finance at George State University. "To go deeper, the theory says that systematic risk, or risk that is common to all investments" -- also known as 'beta' -- "is the kind of risk that investors should care about."
This theory was first articulated in the 1960s by Sharpe (1964), Lintner (1965), and Mossin (1966). However, empirical work dating as far back as 1972 didn't support the theory. In fact, many researchers found that stocks with high risk often do not deliver higher returns, even in the long run.
"It's the foundational theory of asset pricing but has little empirical support in the data. So, in a sense, it's the big question," Murray said.
ISOLATING THE CAUSE
In a recent paper in the Journal of Financial and Quantitative Analysis, Murray and his co-authors Turan Bali (Georgetown University), Stephen Brown (Monash University) and Yi Tang (Fordham University), argue that the reason for this 'beta anomaly' lies in the fact that stocks with high betas also happen to have lottery-like properties - that is, they offer the possibility of becoming big winners. Investors who are attracted to the lottery characteristics of these stocks push their prices higher than theory would predict, thereby lowering their future returns.
To support this hypothesis, they analyzed stock prices from June 1963 to December 2012. For every month, they calculated the beta of each stock (up to 5,000 stocks per month) by running a regression-- a statistical way of estimating the relationships among variables -- of the stock's return on the return of the market portfolio. They then sorted the stocks into 10 groups based on their betas and examined the performance of stocks in the different groups.
"Theory predicts that stocks with high betas do better in the long run than stocks with low betas," Murray said. "Doing our analysis, we find that there really isn't a difference in the performance of stocks with different betas."
They next analyzed the data again and, for each stock month, calculated how lottery-like each stock was. Once again, they sorted the stocks into 10 groups based on their betas and then repeated the analysis. This time, however, they implemented a constraint that required each of the 10 groups to have stocks with similar lottery characteristics. By making sure the stocks in each group had the same lottery properties, they controlled for the possibility that their failure to detect a difference in performance between in their original tests was because the stocks in different beta groups have different lottery characteristics.
"We found that after controlling for lottery characteristics, the seminal theory is empirically supported," Murray said.
In other words: price pressure from investors who want lottery-like stocks is what causes the theory to fail. When this factor is removed, asset pricing works according to theory.
IDENTIFYING THE SOURCE
Other economists had pointed to a different factor -- leverage constraints -- as the main cause of this market anomaly. They believed that large investors like mutual funds and pensions that are not allowed to borrow money to buy large amounts of lower-risk stocks are forced to buy higher-risk ones to generate large profits, thus distorting the market.
However, an additional analysis of the data by Murray and his collaborators found that the lottery-like stocks were most often held by individual investors. If leverage constraints were the cause of the beta anomaly, mutual funds and pensions would be the main owners driving up demand.
The team's research won the prestigious Jack Treynor Prize, given each year by the Q Group, which recognizes superior academic working papers with potential applications in the fields of investment management and financial markets.
The work is in line with ideas like prospect theory, first articulated by Nobel-winning behavioral economist Daniel Kahneman, which contends that investors typically overestimate the probability of extreme events -- both losses and gains.
"The study helps investors understand how they can avoid the pitfalls if they want to generate returns by taking more risks," Murray said.
To run the systematic analyses of the large financial datasets, Murray used the Wrangler supercomputer at the Texas Advanced Computing Center (TACC). Supported by a grant from the National Science Foundation, Wrangler was built to enable data-driven research nationwide. Using Wrangler significantly reduced the time-to-solution for Murray.
"If there are 500 months in the sample, I can send one month to one core, another month to another core, and instead of computing 500 months separately, I can do them in parallel and have reduced the human time by many orders of magnitude," he said.
The size of the data for the lottery-effect research was not enormous and could have been computed on a desktop computer or small cluster (albeit taking more time). However, with other problems that Murray is working on - for instance research on options - the computational requirements are much higher and require super-sized computers like those at TACC.
"We're living in the big data world," he said. "People are trying to grapple with this in financial economics as they are in every other field and we're just scratching the surface. This is something that's going to grow more and more as the data becomes more refined and technologies such as text processing become more prevalent."
Though historically used for problems in physics, chemistry and engineering, advanced computing is starting to be widely used -- and to have a big impact -- in economics and the social sciences.
According to Chris Jordan, manager of the Data Management & Collections group at TACC, Murray's research is a great example of the kinds of challenges Wrangler was designed to address.
"It relies on database technology that isn't typically available in high-performance computing environments, and it requires extremely high-performance I/O capabilities. It is able to take advantage of both our specialized software environment and the half-petabyte flash storage tier to generate results that would be difficult or impossible on other systems," Jordan said. "Dr. Murray's work also relies on a corpus of data which acts as a long-term resource in and of itself -- a notion we have been trying to promote with Wrangler."
Beyond its importance to investors and financial theorists, the research has a broad societal impact, Murray contends.
"For our society to be as prosperous as possible, we need to allocate our resources efficiently. How much oil do we use? How many houses do we build? A large part of that is understanding how and why money gets invested in certain things," he explained. "The objective of this line of research is to understand the trade-offs that investors consider when making these sorts of decisions."
Aaron Dubrow | EurekAlert!
Microtechnology industry is hiring – positive developments of past years continue
09.04.2018 | IVAM Fachverband für Mikrotechnik
RWI/ISL-Container Throughput Index with minor decline on a high overall level
20.03.2018 | RWI – Leibniz-Institut für Wirtschaftsforschung
The quality of additively manufactured components depends not only on the manufacturing process, but also on the inline process control. The process control ensures a reliable coating process because it detects deviations from the target geometry immediately. At LASER World of PHOTONICS 2019, the Fraunhofer Institute for Laser Technology ILT will be demonstrating how well bi-directional sensor technology can already be used for Laser Material Deposition (LMD) in combination with commercial optics at booth A2.431.
Fraunhofer ILT has been developing optical sensor technology specifically for production measurement technology for around 10 years. In particular, its »bd-1«...
The well-known representation of chemical elements is just one example of how objects can be arranged and classified
The periodic table of elements that most chemistry books depict is only one special case. This tabular overview of the chemical elements, which goes back to...
Light can be used not only to measure materials’ properties, but also to change them. Especially interesting are those cases in which the function of a material can be modified, such as its ability to conduct electricity or to store information in its magnetic state. A team led by Andrea Cavalleri from the Max Planck Institute for the Structure and Dynamics of Matter in Hamburg used terahertz frequency light pulses to transform a non-ferroelectric material into a ferroelectric one.
Ferroelectricity is a state in which the constituent lattice “looks” in one specific direction, forming a macroscopic electrical polarisation. The ability to...
Researchers at TU Graz calculate the most accurate gravity field determination of the Earth using 1.16 billion satellite measurements. This yields valuable knowledge for climate research.
The Earth’s gravity fluctuates from place to place. Geodesists use this phenomenon to observe geodynamic and climatological processes. Using...
Discovery by Brazilian and US researchers could change the classification of two species, which appear more akin to jellyfish than was thought.
The tube anemone Isarachnanthus nocturnus is only 15 cm long but has the largest mitochondrial genome of any animal sequenced to date, with 80,923 base pairs....
29.04.2019 | Event News
17.04.2019 | Event News
15.04.2019 | Event News
19.06.2019 | Physics and Astronomy
19.06.2019 | Information Technology
19.06.2019 | Materials Sciences