with Emmanuel A. Abbe and Amir E. Khandani, American Economic Review: Papers & Proceedings 102 (2012), 65-70.
Unlike other industries in which intellectual property is patentable, the financial industry relies on trade secrecy to protect its business processes and methods, which can obscure critical financial risk exposures from regulators and the public. We develop methods for sharing and aggregating such risk exposures that protect the privacy of all parties involved and without the need for a trusted third party. Our approach employs secure multi-party computation techniques from cryptography in which multiple parties are able to compute joint functions without revealing their individual inputs. In our framework, individual financial institutions evaluate a protocol on their proprietary data which cannot be inverted, leading to secure computations of real-valued statistics such as concentration indexes, pairwise correlations, and other single- and multi-point statistics. The proposed protocols are computationally tractable on realistic sample sizes. Potential financial applications include: the construction of privacy-preserving real-time indexes of bank capital and leverage ratios; the monitoring of delegated portfolio investments; financial audits, and the publication of new indexes of proprietary trading strategies.
with Jasmina Hasanhodzic and Emanuele Viola, Quantitative Finance 7, 1043-1050
We propose to study market efficiency from a computational viewpoint. Borrowing from theoretical computer science, we define a market to be efficient with respect to resources S (e.g., time, memory) if no strategy using resources S can make a profit. As a first step, we consider memory-m strategies whose action at time t depends only on the m previous observations at times t - m,...,t - 1. We introduce and study a simple model of market evolution, where strategies impact the market by their decision to buy or sell. We show that the effect of optimal strategies using memory m can lead to "market conditions" that were not present initially, such as (1) market bubbles and (2) the possibility for a strategy using memory m' > m to make a bigger profit than was initially possible. We suggest ours as a framework to rationalize the technological arms race of quantitative trading firms.
with Ely Dahan, Adlar J. Kim, Tomaso Poggio, and Nicholas T. Chan, Journal of Marketing Research, 48 (2011), 497-517.
Identifying winning new product concepts can be a challenging process that requires insight into private consumer preferences. To measure consumer preferences for new product concepts, the authors apply a 'securities of trading of concepts,' or STOC, approach, in which new product concepts are traded as financial securities. The authors apply this method because market prices are known to efficiently collect and aggregate private information regarding the economic value of goods, sevices, and firms, particularly when trading financial securities. This research compares the STOC approach against stated-choice, conjoint, constant-sum, and longitudinal revealed-preference data. The authors also place STOC in the context of previous research on prediction markets and experimental economics. The authors conduct a series of experiments in multiple product categories to test whether STOC (1) is more cost efficient than other methods, (2) passes validity tests, (3) measures expectations of others, and (4) reveals individual preferences, not just those of the crowd. The results also show that traders exhibit bias on the basis of self-preferences when trading. Ultimately, STOC offers two key advantages over traditional market research methods: cost efficiency and scalability. For new product development teams deciding how to invest resources, this scalability may be especially important in the Web 2.0 world, in which customers are constantly interacting with firms and one another in suggesting numerous product design possibilities that need to be screened.
with Amir E. Khandani and Adlar J. Kim, Journal of Banking & Finance 34 (2010), 2767-2787.
We apply machine-learning techniques to construct nonlinear nonparametric forecasting models of consumer credit risk. By combining customer transactions and credit bureau data from January 2005 to April 2009 for a sample of a major commercial bank's customers, we are able to construct out-of-sample forecasts that significantly improve the classification rates of credit-card-holder delinquencies and defaults, with linear regression R-squared's of forecasted/realized delinquencies of 85%. Using conservative assumptions for the costs and benefits of cutting credit lines based on machine-learning forecasts, we estimate the cost savings to range from 6% to 25% of total losses. Moreover, the time-series patterns of estimated delinquency rates from this model over the course of the recent financial crisis suggests that aggregated consumer-credit risk analytics may have important applications in forecasting systemic risk.
Journal of Indexes Q3 (2001), 26–35.
Artificial intelligence has transformed financial technology in many ways and in this review article, three of the most promising applications are discussed: neural networks, data mining, and pattern recognition. Just as indexes are meant to facilitate the summary and extraction of information in an efficient manner, sophisticated automated algorithms can now perform similar functions but at higher and more powerful levels. In some cases, artificial intelligence can save us from natural stupidity.
with Martin Haugh, Computing in Science & Engineering 3 (2001), 54–59.
The financial industry is one of the fastest-growing areas of scientific computing. Two decades ago, terms such as financial engineering, computational finance, and financial mathematics did not exist in common usage. Today, these areas are distinct and enormously popular academic disciplines with their own journals, conferences, and professional societies. One explanation for this area’s remarkable growth and the impressive array of mathematicians, computer scientists, physicists, and economists that are drawn to it is the formidable intellectual challenges intrinsic to financial markets. Many of the most basic problems in financial analysis are unsolved and surprisingly resilient to the onslaught of researchers from diverse disciplines. In this article, we hope to give a sense of these challenges by describing a relatively simple problem that all investors face when managing a portfolio of financial securities over time. Such a problem becomes more complex once real-world considerations factor into its formulation. We present the basic dynamic portfolio optimization problem and then consider three aspects of it: taxes, investor preferences, and portfolio constraints. These three issues are by no means exhaustive—they merely illustrate examples of the kinds of challenges financial engineers face today. Examples of other computational issues in portfolio optimization appear elsewhere.
Journal of the American Statistical Association 95 (2000), 629-635.
Ever since the publication in 1565 of Girolamo Cardano's treatise on gambling, Liber de Ludo Aleae (The Book of Games of Chance), statistics and financial markets have become inextricably linked. Over the past few decades many of these links have become part of the canon of modern finance, and it is now impossible to fully appreciate the workings of financial markets without them. This selective survey covers three of the most important ideas of finance—efficient markets, the random walk hypothesis, and derivative pricing models—that illustrate the enormous research opportunities that lie at the intersection of finance and statistics.
Foundations of Technical Analysis: Computational Algorithms, Statistical Inference, and Empirical Implementation
with Harry Mamaysky and Jiang Wang, Journal of Finance 55 (2000), 1705–1765.
Technical analysis, also known as "charting,'' has been a part of financial practice for many decades, but this discipline has not received the same level of academic scrutiny and acceptance as more traditional approaches such as fundamental analysis. One of the main obstacles is the highly subjective nature of technical analysis—the presence of geometric shapes in historical price charts is often in the eyes of the beholder. In this paper, we propose a systematic and automatic approach to technical pattern recognition using nonparametric kernel regression, and apply this method to a large number of U.S. stocks from 1962 to 1996 to evaluate the effectiveness of technical analysis. By comparing the unconditional empirical distribution of daily stock returns to the conditional distribution—conditioned on specific technical indicators such as head-and-shoulders or double-bottoms—we find that over the 31-year sample period, several technical indicators do provide incremental information and may have some practical value.
Blending Quantitative and Traditional Equity Analysis, edited by H. Russell Fogler, 1994. Charlottesville, VA: Association for Investment Management and Research
Data-snooping--finding seemingly significant but in fact spurious patterns in the data--is a serious problem in financial analysis. Although it afflicts all non-experimental sciences, data-snooping is particularly problematic for financial analysis because of the large number of empirical studies performed on the same datasets. Given enough time, enough attempts, and enough imagination, almost any pattern can be teased out of any dataset. In some cases, these spurious patterns are statistically small, almost unnoticeable in isolation. But because small effects in financial calculations can often lead to very large differences in investment performance, data-snooping biases can be surprisingly substantial. In this review article, I provide several examples of data-snooping biases, explain why it is impossible to eliminate them completely, and propose several ways to guard against the most extreme forms of data-snooping in financial analysis.
with Craig MacKinlay, Review of Financial Studies 3 (1990), 431–468.
Tests of financial asset pricing models may yield misleading inferences when properties of the data are used to construct the test statistics. In particular, such tests are often based on returns to portfolios of common stock, where portfolios are constructed by sorting some empirically motivated characteristic of the securities such as market value of equity. Analytical calculations, Monte Carlo simulations, and two empirical examples show the effects of this type of data snooping can be substantial.