Using simple returns or log-returns is an empirical issue, which depends in general on the problem at hand.
Dividend Payments
For assets with periodic dividend payments, asset returns must be redefined.
R{t+1} = \frac{P{t+1} + D{t+1}}{Pt} - 1
r{t+1} = \log(P{t+1} + D{t+1}) - \log(Pt)
Where D_{t+1} is the dividend payment of an asset between dates t and t+1
P{t+1} is the price of the asset at the end of t+1 (dividend not included in P{t+1}).
Most reference indices include dividend payments but not all.
Some reference indices (e.g., MSCI indices) are computed without (“price index”) or with dividend payments (“total return index”).
On Datastream, stock prices are available with both definitions.
Excess Returns
Excess return is simply the difference between the asset return and the return on the risk-free asset
Z{i,t+1} = R{i,t+1} - R_{0,t}
z{i,t+1} = r{i,t+1} - r_{0,t}
With R{0,t} and r{0,t} the simple return and log return of the risk-free asset.
The index t is used because the risk-free rate between t and t+1 is determined in t
The risk-free rate should satisfy two constraints:
Low risk of default (rating of the issuer)
Low risk of a loss of capital (choice of the maturity of the security)
In practice, the risk-free rate is often chosen as the return of a Treasury bill with the same maturity as the investment horizon
Definition of Volatility
The unconditional volatility is estimated as the sample standard deviation:
s = \sqrt{\frac{1}{T-1} \sum{t=1}^{T} (Rt - \hat{\mu})^2}
\sigma = \sqrt{\frac{1}{T} \sum{t=1}^{T} (Rt - \hat{\mu})^2}
Where R_t is the simple return on day t and \hat{\mu} is the sample mean over the T days.
s is the unbiased estimator
\sigma is the ML estimator, asymptotically unbiased
For the moment, we focus on a unique estimator. We will discuss the case of a dynamic volatility later on.
Characteristics of Financial Asset Returns (Stylized Facts)
Absence of Autocorrelations: Linear autocorrelations of asset returns are often insignificant, except for very small intraday time scales (~20 minutes, microstructure effects).
Heavy Tails: The unconditional distribution of returns displays a power-law or Pareto-like tail, with a finite tail index greater than two.
Gain/Loss Asymmetry: Large drawdowns are observed in stock prices, but equally large upward movements are not.
Aggregational Normality: As the time scale increases, the distribution of returns looks more like a normal distribution. The shape varies across time scales.
Intermittency: Returns display high variability, quantified by irregular bursts in time series of volatility estimators.
Volatility Clustering: Volatility measures display positive autocorrelation over several days, indicating that high-volatility events tend to cluster in time.
Conditional Heavy Tails: Even after correcting for volatility clustering (e.g., via GARCH models), residual time series still exhibit heavy tails, but less heavy than the unconditional distribution.
Slow Decay of Autocorrelation in Absolute Returns: Autocorrelation function of absolute returns decays slowly, roughly as a power law, suggesting long-range dependence.
Leverage Effect: Volatility measures of an asset are negatively correlated with returns of that asset.
Volume/Volatility Correlation: Trading volume is correlated with volatility.
Asymmetry in Time Scales: Coarse-grained measures of volatility predict fine-scale volatility better than the other way around.
Early Assumptions in Finance
Normality of Log-Returns: Convenient for applications like the Black-Scholes model. Consistent with the Central Limit Theorem if log-returns are i.i.d.
Time Independency (i.i.d. Process): Implied by the Efficient Market Hypothesis (EMH), which only requires unpredictability of returns.
Moments of a Random Variable
Definitions
Assume X (log-returns) has the following cumulative distribution function (cdf) FX(x) = Pr[X \leq x] = \int{-\infty}^{x} f_X(u) du
where f_X(x) is the probability distribution function (pdf)
The sample average is the simpler estimate of location. If the data sample is drawn from a normal distribution, the sample average is the optimal measure of location.
Variance:
It is the optimal measure for normal returns.
s^2 = \frac{1}{T-1} \sum{t=1}^{T} (rt - \hat{\mu})^2
\hat{\sigma}^2 = \frac{1}{T} \sum{t=1}^{T} (rt - \hat{\mu})^2
s^2 is the unbiased estimator, while \hat{\sigma}^2 is the ML estimator.
Standard Deviation:
The square root of the variance.
The mean and variance are not robust to outliers.
Higher Moments
The sample counterpart of the standardized skewness and kurtosis are
For a normal distribution, the skewness and the excess kurtosis are zero.
If \hat{S} < 0, the distribution has a fatter left tail.
If \hat{S} > 0, the distribution has a fatter right tail.
If \hat{K} < 3, the distribution has thinner tails than the normal.
If \hat{K} > 3, the distribution has fatter tails than the normal.
If we reject normality, we should probably use robust measures for higher moments.
Tests for Normality
Tests:
Jarque-Bera test (based on the moments of the distribution).
Kolmogorov-Smirnov test (based on the empirical distribution function).
Anderson-Darling test (based on the empirical distribution function).
Jarque-Bera Test
Based on the fact that under normality, skewness and excess kurtosis are jointly equal to zero.
Hypotheses
H_0: Skewness = 0 and Kurtosis = 3
The normal distribution is defined by the first two moments, so there is no constraint on the mean and variance under the null.
Under normality, the sample skewness and kurtosis are mutually independent.
Test Statistic:
JB = T \left[\frac{S^2}{6} + \frac{(K - 3)^2}{24}\right]
Under the null hypothesis, it is asymptotically distributed as a \chi^2(2).
If JB \geq \chi^2_{\alpha}(2), then we reject the null hypothesis at level \alpha.
Empirical Distribution Function Goodness-of-Fit Tests
Compare the empirical cdf and the assumed theoretical cdf F^*(x; \theta) (here, the normal distribution).
The time series (r1, r2, …, rT) is associated with some unknown cdf, Fr(x).
As the true distribution is unknown, it is approximated by the empirical cdf, G_T(x), defined as (step function with steps of height [1/T] at each observation)
GT(x) = \frac{1}{T} \sum{t=1}^{T} 1 {r_t \leq x }
The empirical cdf G_T(x) is compared with the assumed cdf F^*(x; \theta) to see if there is good fit.
Hypotheses
H0: Fr(x) = F^*(x; \theta) for all x.
HA: Fr(x) \neq F^*(x; \theta) for at least one value of x.
EDF tests are based on the result that if Fr(X) is the cdf of X, then the variable Fr(X) is uniformly distributed between 0 and 1.
Kolmogorov-Smirnov and Lilliefors Tests
A measure of the difference between two cdfs is the largest distance between the two functions G_T(x) and F^*(x; \theta).
KS Test Statistic:
KS = \max{1, …, T} |F^*(x; \theta) - GT(x)|
Recipe for the normal distribution N(\mu, \sigma^2):
If \mu and \sigma^2 are unknown, estimate \hat{\mu} and s^2 from the original data (Lilliefors test)
Sort the sample data by increasing order and denote the new sample \left{rt\right}{t=1}^{T}, with r1 \leq r2 \leq … \leq rT. By construction, we have GT(r_t) = \frac{t}{T}.
Evaluate the assumed theoretical cdf F^*(rt; \theta) for all values \left{rt\right}_{t=1}^{T}. Use (\mu, \sigma^2) if they are known or (\hat{\mu}, s^2) if they are not
Critical values: 0.805/\sqrt{T}, 0.886/\sqrt{T}, and 1.031/\sqrt{T} for a 10%, 5%, and 1% level test
Lecture 2: Characteristics of Financial Time Series – 2
Objectives
Addresses the time dependency of asset returns.
Explores the properties of correlation across asset returns.
Discusses issues with data.
Stationarity
Prices are generally non-stationary, while returns are typically stationary.
Prices are often non-stationary due to economic expansion, productivity increases, or financial crises.
Returns tend to fluctuate around a constant level, indicating a constant mean over time (mean-reverting process).
Most return sequences can be modeled as a stochastic process with time-invariant first two moments (weak stationarity).
Serial Correlation
No serial correlation: Autocorrelations of asset returns R_t are often insignificant, except for very small intraday time scales (≈ 20 minutes) where microstructure effects occur.
Absence of serial correlation does not mean the returns are independent.
Squared returns and absolute values of returns often display high serial correlation.
The fact that returns do not show any serial correlation does not mean that they are independent!
In fact, squared returns and the absolute value of returns display high serial correlation
Time Dependence of Moments
The time dependence of moments is crucial for testing market efficiency.
Assuming a weakly (or covariance) stationary time series (rt, …, rT), we have:
E[r_t] = \mu
\forall t V[r_t] = \sigma^2
\forall t Cov[rt, r{t-k}] = \gamma_k
\forall t, k
\gammak is the auto-covariance, measuring the dependency of rt with respect to its past observation r_{t-k}.
\gamma0 = V[rt] = E[(r_t - \mu)^2]
\gammak = \gamma{-k}
For a weakly stationary process, the autocovariance \gamma_k depends on the horizon k, but not on the date t.
Auto-Correlation Function (ACF)
The auto-correlation function (ACF) is preferred over auto-covariance as a measure of dependency because it is a bounded measure: