1. Introduction
In this paper, we propose a general framework, under which optimal portfolio construction or investment activities can be carried out, along with a number of trading strategies. This paper adds to the knowledge of the investment management literature by introducing the following ideas. First, we recognize the time varying property of asset return distributions1, develop and call for new methods, potentially based on machine learning and panel regression, to conduct the model inference and portfolio optimization. Second, in order to build the investment management framework, we propose to use machine learning method to handle big data input and the dynamic programming problems, which might arise when we try to model the market uncertainty and formulate a dynamic optimal investment problem. Third, most importantly, we propose a completely different way to look at the randomness of the financial market. Previous work focuses on forecasting-type investment management techniques and tries to build various types of models to predict the uncertainties. Technical analysis is inevitably used. The authors of this paper, however, do not rely on the serial correlations of asset returns or analysis based on technical indicators. Instead, we effectively use the information from the cross-sectional data of asset returns and try to build various statistics, i.e., the crystal balls, that enable us to observe the future realizations of the market uncertainty in an aggregate or portfolio level. Investment strategies that are accurate at certain confidence levels are proposed and tested via simulation and backtesting studies. The analysis in this paper, for the first time in the financial literature, utilizes the cross sectional data of financial assets to infer their aggregate time series behavior. The proposed theories prove to be effective in both simulations, in an artificial environment, and backtesting studies with real market data. Last, but not least, we propose a combination of brute-force model-free approach, such as machine learning (reinforcement learning or Q-learning) in financial analysis, which can be found in [1] - [6] , and purely theoretical approaches such as no arbitrage pricing, hedging and dynamic stochastic general equilibrium (DSGE) studies. We try to find a balance between those methodologies, in order to yield better results, i.e., investment frameworks, investment management strategies and portfolio construction methodologies with good empirical performance. For example, we suggest using reinforcement learning (RL) to solve, potentially, the dynamic portfolio optimization problems, with minimal assumption on the underlying asset return dynamics. Moreover, we argue that, the underlying asset dynamics, often modeled via an Itô process in the stochastic analysis literature, can be combined with artificial neural network to yield better fits to the market data, meantime without losing the theoretical ground2.
Throughout the history of financial analysis, researchers and practitioners strive to build investment or trading strategies with short, medium or long term to benefit from the economic and market movements. However, inevitably in the literature, all the work focus on predicting future market movement with the information available from the past. For example, popular methods include, but are not limited to, trend following, mean-reversion or long-short strategies. Taking the class of trend following strategies as an example, some source of literature review can be found in [7] , where the methods identifying the trend of market movements and benefiting from riding that trend are surveyed. A loss will occur when market reverts and oscillates. More descriptions can also be found in [8] , [9] and [10] , as an incomplete list. On the other hand, mean reversion strategies build on the belief that the market variables will revert to their long run equilibrium levels, with short term adjustments or fluctuations, where trading opportunities appear. Work in this category includes [11] , [12] and [13] , among others. However, this investment method will suffer greatly if the spread between the two sets of chosen assets widens. Therefore, both classes of investment strategies incur losses with advert market movements. The root cause is that, information from the past might not be a good indicator of future market movements or at least it can only predict or explain a limited fraction of future volatility.
Moreover, current practice estimates model parameters directly using historical time-series data, which, potentially, introduces some problems. First, the distributions of economic or financial time series might be time varying, therefore a brute-force inference using historical data might result in significant estimation bias. Second, imposing additional model structure on the historical data series introduces further model risks.
All those facts encourage the authors of this paper to search for a new framework, under which portfolio management and trading activities are conducted. Ideally, first, this framework should be model-free such that it is able to incorporate and accommodate all the approaches, either model dependent or model independent, static or dynamic, into it. Second, this framework should allow efficient and accurate parameter inference that captures time dependent feature of the coefficients involved.
In addition to the proposal of a general framework, we work with a large class of optimal investment strategies based on a rotation of the original asset space. For some of the strategies, we do not try to predict the market movements from the past data directly, which we only utilize to get model parameter estimates. Here all the mentioned parameters are theoretically current time measurable and do not involve prediction. Instead, we try to identify, orthogonalize and isolate the random sources in the market and diversify away the randomness (asymptotically or exactly). Moreover, we do not try to model the serial correlation structure of the asset returns and mainly focus on the cross-sectional properties of them. Simulation and backtesting studies show good applicability of the investment models that we have developed. To the best of our knowledge, this paper is the first to discuss such orthogonalization and diversification, in the literature of investment and portfolio management.
In the end, numerical experiments are carried out. We find consistently good performance under both simulation and backtesting studies, which coincides with the basic intuition that if we get good parameter estimates, performance will be guaranteed by mathematical and probability laws. In spite of the good performance, limitations of the testing approach and remediation are discussed.
The organization of this paper is as follows. Section 2 describes the main investment framework. Section 3 introduces concrete investment strategies under the proposed framework. Section 4 performs simulation studies to test the models proposed, Section 5 backtests the models using equity data in the US and China markets and Section 6 concludes. Readers who are only interested in the investment strategies can skip Section 2, which contains the rigorous mathematical derivations, and directly start from Section 3.
2. The Optimal Investment Framework
This section contains the mathematical description and construction of the optimal investment framework. We first introduce the necessary probability space and tools required for further analysis. Next, we propose a rotated asset space which we will be working with instead of the original asset space. Afterwards, we write down the general formulation of the optimization problem and propose an investment framework to jointly solve the optimization problem and meantime perform parameter inference via machine learning (ML). Readers can directly start from Section 3 for concrete trading models, with the understanding that skipping this section does not prevent them from understanding the investment strategies.
2.1. Mathematical Setup
Assume that the randomness in the financial economic system under consideration is modeled by a filtered probability space
, where
represents the sample space, modeling the entire collection of possible outcomes of the system,
represents all the information in the system and
is the information filtration with
, satisfying the usual conditions, and
is the historical probability measure on
. There are M financial assets in the economic system, whose rate of return processes are denoted by an M-vector
. Suppose that
, meaning that
has finite variance-covariance structure for all
. Obviously we have the following (trivial) decomposition
(1)
Here process
is an M-vector generating the randomness, all the quantities are defined with proper dimensions, and it is obvious that
, i.e.,
generates all the information in the financial economic system. The last line of the above equation tries to impose some model structures on
and
, which can be of any functional form.
can be modeled by, for example, a joint Lévy process, a system of stochastic differential equations, a linear or nonlinear time series or even a collection of artificial neural networks. A detailed explanation of Equation (1) is postponed to the next section.
2.2. Economic Settings and Financial Market
Following Section 2.1, we assume that the source of randomness in the economy and the financial market can be represented by an N-dimensional (
) jointly independent and identically distributed (i.i.d.)3 stochastic process
(conditional on
) with zero mean and
,
for any
and
, where
is the Kronecker Delta and h is the smallest time increment under our consideration. Information filtration
is generated by U. Recall that, there are M primary financial assets traded in the market, whose rate of returns are denoted by an M-vector
. Suppose that we have the conditional decomposition4 (conditional on
)
(2)
where
is an
vector and
is an
matrix, which can both be estimated at some precision and accuracy at time
. Then, we know that
5 and
6.
Note that, this setting is very general as
can be 1-second, 1-day, 1-week or 1-year and it encompasses all the possible time frequencies.
can be any stochastic process materialized at time
, for every
.
also helps to model the cross-sectional and time series correlation structure of R. When
, we can consider the limiting case of Equation (2) as a system of stochastic differential equations with jumps (SDEJ), when
and
, i.e., they are functions of random sources
and
at time
. Equation (2) defines a general semi-martingale R when proper technical conditions are satisfied.
Remark 1 (On M, N and Factor Structure). If
, i.e., the number of financial assets R is larger than the number of random sources U, we are in an effectively complete market. As some (
, to be accurate) assets are redundant, we can choose N linearly independent assets in this case. However, if
, the market is incomplete. For the sake of generality, we study the case where
. Consider an orthonormal basis
and decompose
as
(3)
where
is the canonical inner product in
, i.e., the Hilbert space of all the stochastic processes that have finite second order moments at time t, and m runs from 1 to M. Equation (3) defines an infinite series expansion of
and we truncate the first M elements and write
(4)
Here
is considered to be the residual term. Then, if we denote
,
, likewise for
and
, we will have
(5)
Clearly, our analysis is asymptotic in nature under the assumption that
where
is an M-vector with entries all equal to 1. The above description justifies the analysis in this paper, the asymptotic investment framework and related strategies to be proposed in Section 3.
Remark 2 (More on Factor Structure). Mapping Equation (2) to the popular factor representation of asset returns R, we have
(6)
where
can be 0 or h. Equation (2) can be viewed as an equivalent form of Equation (6) after a proper Gram-Schmidt orthogonalization process on
. The validity of a factor representation is justified in Remark 1. Moreover, the determination of the factor space F requires a thorough theoretical and empirical study. For example, one choice of the factors is VIX index, studied in [14] . A deep insight of linear factor models can be found in [15] . Classical results can also be found in [16] , [17] , and references therein.
2.3. The Optimal Investment Asset Space
2.3.1. The Optimal Rotation
We illustrate ideas under the discrete-time setting, with the understanding that to solve continuous time models, time discretization is inevitable, which essentially reduces a continuous time problem to a discrete time one. We are seeking a rotation matrix
and a portfolio weight vector
of appropriate dimensions, such that the w-weighted average of the rotated asset space
(7)
is what we want to work with.
An Example of the Rotation
Rewrite Equation (2) as
(8)
assuming that
is of full-rank. Define the rotated assets as
, with the Moore-Penrose inverse
defining the rotation7. Note that, conditional on the information filtration
, i.e., all the public or private information available at time
,
is mutually orthogonal. Our goal is to find optimal portfolio weights
, on the rotated asset space
, for all
and
. The optimally realized return at time t is therefore
(9)
(10)
and
(11)
With the rotated assets
, we can do the following optimization, i.e., to minimize the impact of the error term
on the investment strategies that we try to develop. In the sequel, we will always assume
, without loss of generality, unless we want to discuss the cases with incomplete market.
The Pricing Kernel
After the rotation stated in the previous section, we can define the stochastic discount factor in the rotated asset space, projected onto the space spanned by the assets
, as
, where
is the (locally) risk-free rate and
, meaning that
lies in the orthogonal space of
in
. A linear transformation can bring the pricing kernel back to the original asset space.
2.3.2. Parameter Estimation via Regression Techniques
The General No Arbitrage Asset Pricing Formula. The classic asset pricing relation reads, under no arbitrage condition
(12)
where h is the smallest time increment,
denotes the asset price at time t,
is the stochastic discount factor evaluated at time t and
is the information filtration. Simple algebra transforms the above equation to
(13)
where
is the return of the locally risk-free asset. Of course, stochastic discount factor m depends on the information filtration
, which is assumed to be generated by an r-dimensional process X. Further assume that
.
Suppose that the asset span is denoted by
and the corresponding return process is
, i.e.,
is square-integrable. Equation (13) is equivalent to
(14)
Here
denotes the t-price of a zero-coupon bond with maturity time
. Because we have
and
8, we can, without loss of generality, assume that
. Then, we have
(15)
where
is the conditional variance-covariance matrix and therefore
(16)
Once we obtain the weight process
, we can price any asset in the span
. Therefore, the problem is to compute
and
for any
pair. Alternative method to compute the weight process can be found in [6] . Suppose that the asset returns depend on a set of state variables, aforementioned and denoted by X. Here we separate X into
, where Y is not asset specific and serves as a general factor process, but Z is asset specific, i.e., sub-vector
corresponds to asset
and
.
Time-Series Regression. For most of the papers in the literature, time-series regression is utilized to find the relation
(17)
This, of course, can serve as an option in our analysis. Functional form f can be represented by a basis function expansion or a deep artificial neural network. To obtain the values of
, we can run a time-series regression of
on
. This means we can utilize regression technique to compute the volatility estimates.
Panel Regression. In this paper, we emphasize the methodology using the panel regression technique. Following the notation above, we have common risk factors Y, asset specific risk factors
and
. We can run a panel regression of the following form
(18)
8
is the information sigma-algebra generated by
and
is the linear space spanned by R.
Then, we have
(19)
The benefit of doing so is three-fold. First, it can make use of the entire cross-section of asset return data. Second, it can give estimations of future risk factor returns X. Third, it generates more observations and can reduce the reliance on the past historical data series, i.e., k can be a small integer. To obtain the values of
, we can run a panel regression of
on
, where
is the new risk factor process adjusted for the dimensions of the variance-covariance matrix. In order to achieve better precision in function approximation via machine learning, we can interpolate the cross-sectional asset returns and formulate a cross-sectional curve to get as many samples as possible at any granularity and then perform the regression functional approximation via machine learning. For a meaningful interpolation we can first sort the cross-sectional asset returns and then interpolate the resulted curve via any interpolation technique, for example, linear, polynomial interpolation or even an artificial neural network. The detailed interpolation should be done as following. First sort
and the related sorted return series is
. Also do the same for X and denote
as the sorted series. Suppose the regressor variable is
. We interpolate p points between
. Assume that
pair corresponds to
. We choose p
equal distance points from the interpolated sequence for
and the goal is achieved.
General Discussions. Time-series regression seeks the same functional relation between the dependent and independent variables across time, while the functional relations can be time-varying via panel regression. However, the dependency on risk factors of each asset in the universe can be different for time-series regression while for panel regression the functional dependency is the same across assets. Moreover, in time-series regression, we need more samples than the number of factors. However, in panel regression, we can incorporate asset specific factors, for example, earnings-per-share or book to market ratio, into the regression framework. If the asset span contains 7000 assets, we use 10 different asset specific risk factors and look back 2 periods in time, in panel regression there will be 7000 observations and 20 independent variables. As discussed previously, we can also interpolate the 7000 asset returns to get more samples, potentially, infinity, to run the regression. The reasons for the preference on panel regression over time-series regression are as follows. First, the functional dependency of asset returns on state variables, i.e., factors, might be time varying. Therefore, using long historical data series might be inappropriate. Second, according to derivatives pricing theory, the functional form of asset returns on the state variables is the same across different assets, which means a panel regression is suitable.
2.4. The General Optimization Problem
Consider the following dynamic portfolio optimization problem in a stochastically varying financial environment
(20)
where
is a stopping time and
is a subspace of all the stopping times between
.
is the discount factor under the physical measure and
is the cumulative payoff measure of the investment portfolio process
adjusted by the risk, which can be path-dependent, where
, a proper space of optimal portfolio weights.
is an appropriate measure of performance, which can be conditional expectation, conditional quantile or other metrics. Note that, Equation (20) relates the investment portfolio choice practice to a dynamic problem with optimal stopping. Therefore the optimal exercise boundary can be computed. Specifically, G can be Sharpe ratio, Sortino ratio, max drawdown, or any other risk adjusted performance measure. Function G can depend on realized values of state variables, or conditional expected values that require us to use ANNs to predict. The ultimate output of the above optimization problem is a set of portfolio weights at each time t. Note that
can be modeled also via ANNs, according to [6] .
2.5. A New Investment Framework Incorporating Reinforcement Learning
2.5.1. The Big Picture
In [4] , [18] - [23] , reinforcement learning (RL) is introduced to solve dynamic hedging, portfolio optimization and asset pricing problems. The ability to handle dynamic programming makes RL suitable for those three types of problems, which are essentially the optimization of a target function (or functional) under dynamic constraints on control variables, where the state of the nature might evolve in a stochastic manner. The technical advantages and the ability to process huge amounts of data and high dimensional computations enable researchers to consider more complicated problems under realistic assumptions, for example, the consideration of market frictions, information asymmetry, stochastic differential games (that arise in the pricing practice of some types of equity swaps), market making and therefore mean-field games. Contrary to the traditional approach to relate those problems to BSDEs (backward stochastic differential equations) and solve them using machine learning methods, which is documented in [24] and [25] as an incomplete list, the RL approach seeks to solve the aforementioned problems in a more direct and brute-force manner, under the theoretical economic framework, therefore resulting in a model-free solution.
Back to our framework of investment, we propose to work under the rotated asset space
and use Equation (20) as our main optimization problem. We call for a general methodology to combine the steps of model estimation and dynamic optimization. The reason is that, under the observation that the distributions of asset returns are time varying, it introduces bias to estimate the model parameters
at time t directly using the data prior to it. However, RL makes it possible to optimize the portfolio weights in a model-free manner. Equation (20) is more general than mean-variance optimization because G can be functional of past or future trajectories of
, which results in a dynamic stochastic optimal control problem. Dynamic portfolio optimization takes into account the market regime changes. As [26] illustrates, there are two additional hedging terms that account for stochastic varying interest rate and market price of risk processes in the portfolio decomposition formula.
To summarize, the authors of this paper propose a new way of practice in both pricing and trading, to eliminate model dependence9, while maintaining the general economic or finance framework. This calls for the development of more advanced Artificial Neural Network (ANN) and RL techniques10.
2.5.2. Factor Construction
The first step to create the general investment framework is data processing and factor construction. Basically, factors represent the risk decomposition of any asset return in the universe and by bearing the risks, investors get rewarded in the financial market. Factors, which can be both qualitative and quantitative, should fall into the following categories. First, political environment and policies made by government and other authorities should be included. Second, macroeconomic factors, such as economic cycle, GDP, inflation, monetary policies and fiscal policies, should be included. Third, micro-economic and financial factors, such as market returns, yield to maturity for bonds, are helpful. Fourth, fundamental factors, such as earnings per share, book to market ratios, are crucial to equities. Fifth, technical indicator factors, such as the output from other predictive or trading models, namely, the resulted portfolio weights or predictions, stock market technical indicators and indicators from behavioral finance theories, i.e., market sentiment, real time news from natural language processing, should also be included. Please note that, there might be factors that do not fall into the above categories, such as weather conditions, as long as they are helpful in identifying the risk characteristics of an asset, we should include them. To summarize, this module processes raw market data and formulates different factors, for a meaningful risk decomposition of asset returns. Moreover, different factors might have different observation frequencies, sometimes, interpolation, in time dimension, is needed for granularity considerations.
2.5.3. Variables Prediction
9Eliminating model dependence introduces data dependence.
10One reference on this topic is [34] .
After the factors are constructed, we can use the methodologies outlined in Section 2.3.2 to compute the conditional expected values for asset returns and factor returns. The predictions can be made at any time frequency, e.g., a second, an hour, a day, a week, or even a quarter. It would be helpful to mention that the frequency of factor input should match the frequency of prediction. Instead of using calendar as a measure of time flow, we can use trading volume or volatility as the metric. For example, we divide the data into small blocks by equal trading volumes or the accumulation of volatility. If reinforcement learning technique is used, this step will be merged into the next one: Portfolio Construction.
2.5.4. Portfolio Construction
After we obtain the predicted asset and factor returns, we can compute the portfolio weights based on them as inputs. The first candidate is mean-variance frontier or Black-Litterman model. As a second choice, we can utilize machine learning classification techniques, e.g., k-means method, to categorize the asset universe into small groups based on the predicted and realized values and make long-short decisions accordingly. Moreover, Bayesian constrained deep reinforcement learning can be used to formulate the dynamic optimal portfolios. We emphasize that, outputs from other investment models can be used as inputs to our framework either by formulating them as factors or via ensemble method, which is quite popular in machine learning literature and practice.
2.5.5. Risk Management and Attribution
There are many risk management techniques available in the literature. We mention two of them. First, dynamic portfolio insurance techniques can be used to reduce the max-drawdown of the constructed portfolio. Second, we can set stop-gain or stop-loss threshold to formulate a more prudent strategy. The last step is to perform a sensitivity and risk-attribution analysis to better understand the strategy performance.
3. Investment Strategies under the General Framework
3.1. A Brute Force Mean-Variance or Dynamic Portfolio Optimization
11According to [26] , as long as
, we will have
in probability. Therefore, in this situation, it is fine to assign equal weights to the original asset space.
12Alternatively, the joint distribution of U can be recovered directly from historical data.
13Otherwise empirical quantiles have to be used.
We refer the interested readers to [27] , [28] and [29] for an introduction to mean-variance optimization, a Bayesian approach that builds on [27] and an illustration of a typical dynamic portfolio optimization problem. Machine learning based methods to perform dynamic portfolio optimization are introduced in [19] and [23] . Applications of reinforcement learning on portfolio management can be found in [18] , [21] , [22] , [30] and [31] . Please note that, now we are working with the rotated asset returns
11. All of the above analysis can be applied to the rotated asset space under the proposed framework.
3.2. Quantile Investing
With Equation (9), we can build a trading model which is profitable at some confidence level. To do this, we need to assume a joint distribution for U12 in order to compute its quantile values13. Taking joint Gaussian distribution as an example, denote the top α-quantile of
as
and the bottom α-quantile of
as
. Then, if
, we long
and if
, we short
. This method applies at portfolio level or individual asset level.
3.3. Long-Short Portfolio
Long-short portfolio method means that we can score each asset, e.g., based on the past values and predictions of its returns or other variables, in the universe, and rank the cross section. Long the top (bottom) and short the bottom (top) quantiles of the asset span and formulate a trend following (contrarian) strategy. The classification of different groups of assets can be done via machine learning classification methods based on the realized and predicted values, i.e., the computed conditional expectations. Please note that, the financial market can be a perfect blend of momentum effect and mean-reversion effect. For example, the stocks that perform well might continue to perform well. However, the stocks with worst performance in the past might tend to perform better in the next period. This leads us to ask, whether, in the long run, mean-reversion effect is significant and, in short run, momentum effect is dominating or the opposite? Careful empirical investigations should be carried out to answer this question.
3.4. Eliminating the Randomness Asymptotically
3.4.1. The Case with Short-Selling
As a third attempt, we try to utilize strong law of large numbers (S-LLN hereafter) to eliminate the randomness represented by
at time
, if M is
large enough. Let
and we have
(21)
Here
is an M-vector. But note that, as U is jointly independent (or uncorrelated), when M (and therefore N) is large, we have
in probability or almost surely, according to the (strong) LLN. Therefore, we have, approximately
(22)
Equation (22) is extremely powerful. It equates the realization of a future random variable at time t to a variable which is known currently at
. If
, we can long the asset
and when
, we short
. Here
and
are two threshold values that trigger the algorithm, which can be time dependent or even stochastic.
3.4.2. The Case without Short-Selling
14T is a g-vector.
For countries or regions that do not permit short-selling or this activity is costly, we can use futures contract to continue our analysis. Consider the asset space
and g futures contracts
which start at time t with maturity time T14. Consider the following M-factor regression
(23)
where
is an i.i.d. sequence and
describes the covariance structure. Because, we have
(24)
then
(25)
(26)
Here
where GS denotes Gram-Schmidt orthogonalization under the canonical
-norm. Then, perform similar transformation
(27)
Same analysis follows from Equation (27). As we can see from the above algorithm, regression (23) has a large number of factors, which is difficult to implement in practice with limited computation power. Therefore, we do not test this algorithm in this paper.
3.4.3. The Impact of Estimation Error
In this section, we study the impact of estimation errors on the proposed optimal trading strategy in Section 3.4. First, assume that
is estimated with a zero-mean error
, where
is the error term with
. Then, we have
(28)
where
. It can be seen that when
is sufficiently large, we will have
(29)
(30)
Because
has zero mean, according to [32] , as
, we will have convergence in probability
, under mild technical conditions. Therefore, the error term will not impact the estimation of the true value of
if M is large, based on the assumption that
.
The next step is to estimate the impact of estimation error of volatility matrix on the algorithm given that the drift term
is estimated correctly. Suppose we have
(31)
(32)
Here
are estimation error terms which are uncorrelated with U, with
and
. A similar argument shows that the impact of the error terms will vanish if M is sufficiently large. The discussion of joint impact of estimation errors from both drift
and volatility term
is analogous with only more complicated formula.
3.5. Eliminating the Randomness Completely
Following Equations (6) and (7), we can write
(33)
(34)
Let us remind the readers that
,
,
and
, where J is the number of rotated assets chosen to formulate the optimal portfolio. We denote
and
. In this section, we assume that the market is incomplete with
and
. The null space of matrix
is denoted by
and its
dimension is
. Further denote
as a set of linearly independent vectors in
. Taking
as portfolio weights, we have
(35)
Therefore, we have J rotated assets
, which are conditional deterministic. We can perform mean-variance optimization or simply assign equal weights to
, based on the signs15 of
.
3.6. Constructing Factor Mimicking Portfolios
From the decomposition Formula (2), we can derive the equations to construct the factor mimicking portfolios. Suppose we have
(36)
15It is also interesting to find vectors
such that
achieves maximum.
where
is
,
is
,
is
,
is
and
is of
dimension. A simple multiplication of both sides of Equation (36) by the Moore-Penrose inverse of
yields
(37)
We can construct indexes of F by constructing portfolios of R according to Equation (37).
4. Simulation Study
4.1. Methodology
4.1.1. Direct Simulation
For direct simulation approach, We assume that
is estimated correctly and Equation (8) is already obtained. It is justifiable to use independent draws for daily realizations of
across time, as our methods do not try to forecast future values based on serial correlations, which is the main difference between the proposed trading models and almost all the other methods in the existing literature. Of course, simulating directly the market price of risk vector
avoids the estimations of model parameters based on historical data, which eliminates some estimation errors. Transaction cost is simplified to be 5.00% per year and subtracted from the realized strategy return at each month. The real backtesting study is postponed to Section 5. Note that, for risk management purposes, in real-world backtesting in the US equity market, we can mix our portfolio with some percentage of VIX index or add risk management to the current methodologies.
4.1.2. Calibration and Simulation 1
Next, we work with a more realistic setting in order to illustrate how calibration impacts the model performance. The data generating process (DGP) of asset returns R is assumed to be a Gaussian process
, where
is a combination of a trend term
plus a noise
which also follows a joint normal distribution
at each time t.
is again sampled from a joint Gaussian distribution
at each time t. The covariance structure of the DGP is modeled through
.
We first determine the number of assets M and the time scope T, which are set to be 200 and 240-months. Then, we simulate one trajectory of the M assets with T months into the future and perform the model calibration for
in Equation (2) and optimization analysis. For realistic values of
, we record the Sharpe and Sortino ratios of the strategies under equal weight assumption, scaled weight assumption and risk parity assumption, which are documented in Section 4.1.1.
4.1.3. Calibration and Simulation 2
The simulation study in this section compares the simple NAV curve produced by a naive trend following model and our equal-weight model. The trend following model proceeds as follows. Compute the T-period moving average of asset returns for each stock and thus formulate a vector of moving averages
at time t. Long the cross section with equal weights if
and short otherwise. The DGP of the simulation study in this section is an ARMA(1,1)-GARCH model with realistic coefficients, simulated using ugarchpath function offered by rugarch package in R programming language.
4.1.4. Calibration and Simulation 3
In this section, we try to test a long only strategy based on the computed conditional expectations via simulation study. The study proceeds as following. First, simulate asset-specific factor processes
and
as independent draws from random normal variables, where M is the number of assets in the cross-section and T is the number of periods for our consideration. Then transform the processes H and B by the following logic: replace the i-th column by the average of
-th and i-th column for each H and B. This step adds some serial correlations across time for
. Suppose that the asset return process follows
. We compute
via machine learning regression. Then, we sort the conditional expected returns computed, long the top 100 assets and record the performance. We set
and
days in this simulation study. Another two possible strategies that serve as the extensions of the one introduced in Section 3.4 are as follows. First, compute the mean value of the conditional expected returns and if it is positive, long the asset universe. We go short the asset universe otherwise. The asset span is rotated as proposed in Section 2.3.1. Second, find the maximum expected rotated asset return at each time t, and long this asset when the expected value is positive, hold to the next period and we go short otherwise.
4.2. Numerical Results
4.2.1. Simulation Study 1
In this simulation study, we test the methods outlined in Sections 2.4, 3.4 and 4.1.1. Specifically, we generate 36 months of Sharpe ratios
of the M assets, perform optimization at each month and record the net asset value curve (NAV curve). After obtaining the values of
, we generate U and therefore
. We generate
randomly at each month and add to the realized returns to account for the remaining randomness of the market. We assume that
is sampled from a Gaussian distribution with mean 0.95 and standard deviation 1.40, annualized. Those values are estimated from S&P500 historical data from 1990/01/01 to 2018/02/27. The frequency of data is set to be monthly. After
obtaining the time series of
, we estimate
. Table 1 shows the results in various cases. Figure 1 shows the net asset value (NAV) curves.
4.2.2. Simulation Study 2
In this simulation study, we test the methods outlined in Sections 2.4, 3.4 and 4.1.1. Table 2 shows the results in various cases. Figure 2 shows the net asset value (NAV) curves.
Table 1. Sharpe ratios of different strategies.
Figure 1. Net asset value curves of simulation study 1 for unit and equal weights with M = 300.
Figure 2. Net asset value curves of simulation study 1 for unit and equal weights and M = 300.
Table 2. Sharpe ratios of different strategies.
4.2.3. Simulation Study 3
Simulation Study 3 corresponds to the ideas presented in Sections 2.4, 3.4 and 4.1.2. The methodology is as follows. First, simulate the market rate of returns R using the DGP specified in Section 4.1.2. Second, with the realized time series of
, use a moving window of T periods to compute
(38)
(39)
Third, compute
. Fourth, U can be recovered by applying
. We report the Sharpe ratios in Table 3 and NAV curves are shown in Figure 3. We use a moving window of 12 months to estimate the joint variance-covariance matrix
via cov.shrink function in R language and follow the rule below to decompose
, where
,
is the orthogonal matrix of
Figure 3. Net asset value curves of simulation study 3 for scaled, equal and rotated weights.
Table 3. Sharpe ratios of different strategies in simulation study 3.
eigenvectors of
and
is a diagonal matrix collecting the square-root of eigenvalues of
.
4.2.4. Simulation Study 4
This experiment corresponds to Section 4.1.3. The NAV curves are displayed in Figure 4. We set
and
for this study.
4.2.5. Simulation Study 5
The simulation study in this section corresponds to the method documented in Section 4.1.4. NAV plot for long only strategy is on the left, and the NAV for law of large numbers based strategy is on the right (Figure 5).
5. Backtesting
5.1. US Equity Market
5.1.1. The Market Data
We use monthly US stock return data downloaded from the CRSP database provided by WRDS. The data span from January of 1999 to December of 2018. After excluding missing return data, there are 1,521,909 security-month observations. The data set is comprised of various publicly traded securities, of which common stock is the major type, accounting for 81.00% of the sample. Returns are winsorized at the 1st and 99th percentile within each share code-month group to mitigate the effect of outliers or simply truncated at
-level for some chosen
.
5.1.2. The Testing Methodologies
For US equities, we apply the methodology introduced in Section 2.3.2 to estimate the conditional expected asset returns and shrinkage for variance-covariance matrix using a T-day moving average window. Two scenarios are considered. For the first scenario, we aim at testing the claim that as long as the conditional expected asset returns can be correctly estimated, the model performance is guaranteed by S-LLN. Therefore, we assume an imperfect foresight for one period ahead: meaning that at each time t, we can use the returns materialized at time
to infer the conditional expectations of asset returns
Figure 4. Net asset value curves of simulation study 4 for naive trend following and equal weights strategy with (25, 50, 250) assets.
Figure 5. Net asset value curves of simulation study 5.
, where the estimation methodology and R are described in Section 2.3.2. The factor process is chosen also to be R, i.e., we let the cross-sectional returns to explain their own behavior. If the average of the cross-sectional of the expected returns is positive, we go long the entire universe with equal weights. We short the equally-weighted universe if the aforementioned quantity is negative. The second scenario tries to infer the conditional expected asset returns in the same way and obtain variance-covariance matrix through shrinkage method, using cov.shrink function in R. Under the rotated asset space
, we compute the conditional expected returns and rank them, long the top θ-percentage and short the bottom θ-percentage. The performance is evaluated quarterly. We do not account for transaction cost in this backtesting study. A preliminary test shows that a 30 bps cost per unit of portfolio weight change does not affect the testing results. Nor does the restriction of asset space to domestic US common shares affect the outcome. Our tests are robust in this sense.
5.1.3. Testing Results
For the first scenario, we get 100% positive quarterly returns and a Sharpe ratio of 1.87. Sortino ratio is therefore
. The result of the second scenario is summarized in Table 4, where
. Figure 6 shows the same methodology (Scenario 2) applied monthly. Annual Sortino ratio is again
.
5.2. China A-Shares
5.2.1. The Market Data
In addition to the US equity market, we test our algorithm with China A-shares. The monthly return data are downloaded from Wind terminal and span from
Table 4. US equity performance with 500 stocks randomly selected at the beginning of each quarter and performing the long-short strategy on the rotated asset space. Results are stable for different random draws. Sortino ratio is above 9.00.
Figure 6. Net asset value curve of scenario 2, applied monthly.
January of 2002 to August of 2018. We consider Shanghai and Shenzhen A-share markets. HS300 index data are also downloaded from Wind. Returns are winsorized as following. For positive returns, we apply a cap of 99.00% and for negative returns, a floor of −99.00% is imposed.
5.2.2. The Testing Methodology
We test the algorithm introduced in Section 3.4.1 with equally-weighted portfolios in the original asset space. The methodology is as follows. We compute the T-quarter moving average of past returns for each of the stocks in the A-share universe. Then, apply equal weights on the vector of the moving averages. If this number is positive, we long the equal-weight portfolio and if this number is negative, we short the equal-weight portfolio.
5.2.3. Testing Results
The net asset value curve (NAV curve) is shown in Figure 7. Annualized Sortino ratio is 4.67. Percentage of positive annual returns is 81.25%.
5.3. Limitations and Challenges
Theoretically sound, our general investment framework does rely heavily on the quality of parameter estimates for
and
, as illustrated by both simulation and backtesting studies. Better estimates result in better strategy performance. However, under the observations that the distributions of asset price returns are time varying, it becomes very hard to estimate the accurate values of the model parameters based merely on the time-series data as each point in the time series is sampled from a different distribution. Also, the time varying property might be different under various sample frequencies. Our general framework posts two challenges to the field of parameter estimation. First, find the right time frequency such that the model parameters can be estimated
Figure 7. Net asset value curve of backtesting using equal weights with moving window 4 quarters and 3101 stocks in the universe of China A-share market. The figure corresponds to long-short portfolios. Time ranges from 2003-8 to 2018-8.
accurately. Second, find the correct estimation method to minimize the numerical errors.
6. Conclusions and Future Research
In this paper, we outline a general framework of optimal investment and discuss several concrete investment strategies under this framework. The basic idea of the proposed investment methodologies is proper diversification and the elimination of the future market randomness. Simulation studies and backtesting show good performance of the proposed methods under this framework. Note that, the same ideas apply to all categories of investment strategies, i.e., trend following, mean-reversion, long-short, etc.. For example, the long-short strategy tries to score the assets and long or short certain classes of assets whose scores fall in predetermined sets. We can apply this type of analysis on the rotated asset space. The integration of the proposed investment framework with other classes of strategies is also interesting, which we leave to future research.
Moreover, the real market environment is, of course, much more complicated than the data generating processes we consider in the simulation study. The research for more advanced model parameter estimation techniques, when there is model uncertainty, time changing parameters and measurement errors, is highly important and necessary. Examples can be found in [33] and [1] . Moreover, similar to [22] , our framework can be combined with deep reinforcement learning techniques to simultaneously estimate the model parameters and meantime perform portfolio optimization. In [34] , a flexible stochastic volatility model framework is proposed based on neural network, which might also be a future direction to impose flexible model structures on
and perform parameter estimation. We leave the open questions in estimation to future research and meantime direct the interested readers to the relevant literature for more information.
Last, but not least, the simulation studies and backtesting in this paper focus on equity data. However, the methodologies can be applied to any asset class, for which the rate of return can be defined and computed.
NOTES
1This means that asset returns at different time might be sampled from different distributions.
2For example, let the diffusion coefficients be an ANN function-form.
3The requirement of being identically distributed for U can actually be relaxed under Lyapunov central limit theorem (L-CLT), which will cause a slight variation in our discussions later. However, this will not impact the algorithms to be introduced in Section 3.
4Actually, the conditionally joint independence of U can be replaced by being jointly uncorrelated under strong law of large numbers (S-LLN).
5Sometimes we write
.
6Equation (2) can be justified as an approximate relationship due to the infinite orthogonal basis expansion of R in the conditional Hilbert space of
random variables at each time t, see Remark 1.
can be any stochastic process. Of course, they are market, asset class and regime dependent, meaning that different situations might result in different modeling practice for them.
7Here we define rotation as a linear transformation.