Square Root of Time Rule
A common heuristic for time-aggregating volatility is the square root of time rule. I discuss the big idea for this rule and then provide the mathematical assumptions underpinning it.
Big idea
In finance, the riskiness of an asset is often defined by its volatility, which is itself quantified as the standard deviation of the asset’s returns. However, the magnitude of volatility is a function of the frequency of the data. Lower-frequency data will have bigger returns, since the price has more time to move between samples. Therefore, the volatility of lower-frequency data will typically be higher than the volatility of higher-frequency data.
For example, consider Apple (AAPL) stock prices and returns, sampled daily and weekly (Figure ). Clearly, monthly prices will induce larger returns than daily prices. If we compute the standard deviation of the daily and weekly return series in Figure , we get very different estimates of volatility: and respectively. Thus, we would like a way to time-aggregate volatility, so that we can compare and reason about the riskiness of assets at different frequencies.
One way to handle this time aggregation is the square root of time rule, which is the heuristic that a higher-frequency volatility estimate can be scaled to a lower-frequency volatility estimate by multiplying by the square root of “time”. Here, “time” is the ratio between the number of high- to low-frequency samples.
For example, in Figure , the ratio between the number of daily-to-weekly samples is , since there are days in a week. Thus, to estimate weekly volatility from daily data, we would multiply the daily volatility by . To estimate daily volatility from weekly data, we would divide the weekly volatility by . Figure illustrates this computation by comparing daily returns to weekly returns that have been rescaled by . Now the rescaled weekly volatility is .
As a second example, consider the Black–Scholes equation, which scales volatility by the square root of the time to maturity for an options contract.
Deriving the rule
So why does this calculation work? Any random variable that is the sum of independent and identically distributed random variables will have a variance that scales linearly with the number of random variables in the sum. This basic idea applies to many random processes that make independence assumptions.
For example, imagine a Gaussian random walk with steps, where each step is i.i.d. The total distance after steps is the random variable ,
Then the variance of is
and the volatility is
As a second example, in a Poisson process, the arrival time of the -th element is the sum of exponentially distributed random variables,
Here, each is the waiting time between any two arrivals from a Poisson distribution. The variance of is
Again, this implies that the volatility of scales with the square root of the number of random variables, .
At this point, you might be asking: how does any of this apply to the returns in Figures and ? If is a random variable representing the log return (not the raw return) of an asset, then the cumulative return of buying an asset at time and holding it through time is simply the sum of these returns:
If you are unconvinced, see Equation in my previous post on log returns. Thus, when working with log returns, higher-frequency data can be time-aggregated into lower-frequency data by summing the log returns by into desired periods (e.g. weeks, months, years). In fact, this is precisely how I constructed Figures and : I computed daily log returns from daily price data and then computed weekly log returns by summing data within each week. As we saw, these time-aggregated or cumulative returns have a higher volatility. If we’d like, we can normalize the volatility using the square root of time rule.
As a final note, some resources incorrectly claim that the data must be independent and normally distributed, but normality is not a required assumption, as can be seen above.
Conclusion
The square root of time rule is a heuristic for rescaling the volatility estimate of a particular time series to a new data frequency. The rule assumes that our data are the sum of i.i.d. random variables. This applies to many random processes used in finance. Even when these basic assumptions do not hold, scaling volatility by the square root of time can still be a useful heuristic for estimating volatility to a first approximation.