How to Test the Assumption of Persistence

An assumption we often make in trading research is that the future will be at least a little like the past.

I see a lot of beginners making this assumption implicitly without recognising that they’re making it or thinking about whether it’s reasonable to do so.

That’s a mistake.

If you are making this assumption, you need to be aware of it and you need to have confidence that it’s a good assumption to be making.

There are a couple of ways we can explore whether it’s a good assumption:

  • Reason and eyeball
  • Data analysis

In this article, we’ll look into both of these approaches, but we’ll mostly focus on data analysis. I’ll show you some simple techniques that you can start using immediately.

Reason and eyeball

The first approach is simple – eyeballing charts and applying reason.

A simple example is the equity risk premium.

We observe total equity returns in the past being very positive:

equity returns chart
Source: Triumph of the Optimists

And we think we understand why – because equities are risky, they tend to trade lower than the discounted sum of their expected cash flows. Read more about the equity risk premium here.

So we make the assumption that they’ll continue to go up, at least over the long term, in the future.

Is that reasonable?

We can’t really know for sure, but some things that can help us weigh the evidence include:

  • Does it make economic sense? Yes, human risk preferences dictate that a risky investment trades lower than an equivalent risk-free investment.
  • Do we see evidence of it in the data? Yes, we see stocks going up in the long term – see the chart above.
  • Is the effect consistent over time? Yes, but of course there are periods where stock returns were negative – see the chart above.
  • Is the effect consistent elsewhere (other markets)? Yes, it shows up in nearly every developed market for which we have data:

chart of international stock returns
Source: Triumph of the Optimists

The answers to these questions can never provide us with absolute certainty, but if they’re all pointing in the right direction, then we can at least have some confidence that our assumption of persistence is reasonable.

In the case of the equity risk premium, I think we can have confidence in our assumption of persistence simply by applying some reason and eyeballing some data.

But that won’t always be the case. What if we need to dig into the data in more depth? How do we do that?

Using data analysis to explore the assumption of persistence

Next I want to show you some simple data analysis techniques you can use to explore the assumption of persistence.

We’ll do this in the context of portfolio construction and risk management.

When we create portfolios, we size positions based on trying to maximise some objective – usually risk-adjusted returns.

portfolio construction

We tend to forecast our risk metrics (variance, covariance) by estimating them and assuming they tend to stay roughly the same. Is that reasonable?

Let’s start with volatility persistence. We’ll go through each of our questions in turn.

Would it make economic sense for volatility to be persistent?

Yes, I think so, because a given asset has similar risk exposures over time.

Do we see evidence of persistence in the data?

To answer this using data analysis, we can look very directly by splitting our data up into periods that don’t overlap, and seeing if one period’s volatility is correlated with the next:

two-period example

The example in the image above is easy to understand, but it might not be obvious how to do do it with a continuous time-series. This is the process in a nutshell:

  • Start with daily OHLC observations because they’re easy to work with
  • Estimate volatility as the range in per cent between the high and the low (there are many ways to estimate volatility; this is just a simple one)
  • Copy the volatility time series and shift it by 1 day
  • Plot a scatter of the volatility estimate against the shifted volatility estimate and see if there’s some relationship

the simplest thing slide

Here’s how you would do this in R.

First set up our session, load some data for VTI, and plot it:

# setup
library(tidyverse)

# set chart options
options(repr.plot.width = 14, repr.plot.height = 7, warn = -1)
theme_set(theme_bw())
theme_update(text = element_text(size = 20))
# load some data - daily prices for macro etfs:
prices <- read_csv("https://raw.githubusercontent.com/Robot-Wealth/trader-tales/master/data/macro_etfs.csv")
prices %>% head
Rows: 34946 Columns: 8
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr  (1): ticker
dbl  (6): open, high, low, close, volume, closeadjusted
date (1): date

 Use `spec()` to retrieve the full column specification for this data.
 Specify the column types or set `show_col_types = FALSE` to quiet this message.
A tibble: 6 × 8
tickerdateopenhighlowclosevolumecloseadjusted
<chr><date><dbl><dbl><dbl><dbl><dbl><dbl>
DBC2010-01-04 25.17 25.240 25.072 25.24 2046900 23.239
EMB2010-01-04101.74101.930101.330101.93 120500 52.392
GLD2010-01-04109.82110.140109.310109.8016224100109.800
IEF2010-01-04 88.74 88.950 88.630 88.82 619100 66.539
TLT2010-01-04 89.84 90.100 89.580 89.81 2829100 60.998
TMF2010-01-04 77.65 78.475 77.275 78.10 9640 66.373
# get data for US stocks only
VTI <- prices %>% filter(ticker == "VTI")
VTI %>% head
A tibble: 6 × 8
tickerdateopenhighlowclosevolumecloseadjusted
<chr><date><dbl><dbl><dbl><dbl><dbl><dbl>
VTI2010-01-0456.8657.38056.8457.31225150044.181
VTI2010-01-0557.3457.54057.1157.53159770044.351
VTI2010-01-0657.5057.71557.4157.61212030044.412
VTI2010-01-0757.5557.88957.2957.85165670044.597
VTI2010-01-0857.7058.04657.5658.04165000044.744
VTI2010-01-1158.3058.35057.8558.12188310044.806
# plot
VTI %>%
  ggplot(aes(x=date, y=closeadjusted)) +
    geom_line()

Estimate daily volatility of VTI as the percentage range and shift:

VTI <- VTI %>%
  # calculate volatility as pct range
  mutate(range_pct = (high/low - 1) * 100) %>% 
  # ensure our observations are arranged by date
  arrange(date) %>%
  # shift our observations
  mutate(next_day_range_pct = lead(range_pct, 1)) %>%
  # remove NAs (we'll have an NA on our last observation as there's no tomorrow)
  na.omit()

VTI %>% tail
A tibble: 6 × 10
tickerdateopenhighlowclosevolumecloseadjustedrange_pctnext_day_range_pct
<chr><date><dbl><dbl><dbl><dbl><dbl><dbl><dbl><dbl>
VTI2024-03-25258.10258.49257.79257.922709549257.920.27153880.6024565
VTI2024-03-26258.80258.83257.28257.462589441257.460.60245650.7131507
VTI2024-03-27259.13259.85258.01259.734025990259.730.71315070.3156760
VTI2024-03-28259.80260.58259.76259.902663934259.900.31567600.6571826
VTI2024-04-01260.32260.38258.68259.703288458259.700.65718260.5113392
VTI2024-04-02257.50257.50256.19257.303206715257.300.51133920.6943092

Plot a scatter:

VTI %>% 
  ggplot(aes(x=range_pct, y=next_day_range_pct)) + 
  geom_point() +
  # add regression line
  geom_smooth(method = "lm", formula = 'y ~ x') +
  labs(
    title = "Daily range vs next day's range"
  )

Wee see a noisy positive relationship.

If VTI moved a lot yesterday, it tends to move a lot today as well.

So we feel confident that daily range is noisily persistent, based on what we observed in the data.

Overlapping data

An important consideration is overlapping data.

In our example, we had no overlapping data because each daily price bar gave us a single estimate of volatility for that day.

But say we estimated volatility using the average range over 30 days instead. Now we have a window of 30 days’ worth of price bars going into our estimate of volatility today.

You can see how this creates relationships between subsequent daily estimates of volatility: tomorrow’s estimate differs from today’s estimate by only a single data point out of 30.

If you have rolling windows of data in your estimate, then you need to remove the data points where they overlap, which can greatly reduce the amount of data you’re working with.

Is this effect consistent over time?

So far, we’ve looked at our entire data set. But somtimes that can hide important information.

One useful and easy thing to do is to split our data into separate time periods, repeat the analysis above, and see if the effect is consistent through time. If it is, then that would be further evidence that our effect is genuinely persistent.

For our daily VTI volatility estimates, we’ll split our data up by year and make a scatter plot for each year.

This is very simple in R. We just create a column for the year of each observation, then use facet_wrap to make a separate plot for each year:

# we need a bigger plot
options(repr.plot.width = 20, repr.plot.height = 14)

VTI %>%
  mutate(year = year(date)) %>%
  ggplot(aes(x=range_pct, y=next_day_range_pct)) + 
  geom_point() + 
  geom_smooth(method = 'lm', formula = 'y ~ x') +
  facet_wrap(~year, scales = "free") +
  labs(
    title = "Daily range vs next day's range by year"
  )

# reset plot options to default
options(repr.plot.width = 14, repr.plot.height = 7)

You can see that while some years are more volatile than others, the relationship between today’s and yesterday’s volatilities has been reasonably consistent over time. And while the slope of the regression line has changed over the years, it’s always up and to the right.

So this is another data point that we can use as evidence that US stock volatility is reasonably persistent.

Is the effect consistent elsewhere?

We can get more evidence by looking for the effect in other places we expect it to show up.

Let’s repeat our analysis, but this time we’ll do it for all the tickers in our data.

Again, this is quite simple in R. We make use of the group_by function to operate on each ticker separately:

prices %>% 
  # calculate the daily range
  mutate(range_pct = (high/low - 1) * 100) %>%
  # perform subsequent operations on each ticker separately
  group_by(ticker) %>% 
  arrange(date) %>%
  mutate(next_day_range_pct  = lead(range_pct, 1)) %>%
  na.omit() %>%
  # plot each ticker separately
  ggplot(aes(x=range_pct, y=next_day_range_pct)) + 
  geom_point() + 
  geom_smooth(method = 'lm', formula = 'y ~ x') +
  facet_wrap(~ticker, scales = "free") +
  labs(
    title = "Daily range vs next day's range by ticker"
  )

We see that the effect is consistent across assets as well.

So at this point, we have some fairly strong evidence for the persistence of volatility.

Summary

It’s OK to make the assumption of persistence. But you can get yourself into trouble if you don’t think about when you’re making it and whether it’s reasonable to do so.

The overarching point is to be aware of the assumptions you’re making, sometimes implicitly, and think about whether they reflect reality.

The simple data analysis techniques we used in this article can help you to determine when it’s reasonable to assume persistence.

You can apply this process to anything for which you’re making the assumption of persistence. A useful exercise is to repeat the analysis for correlation or covariance between pairs of assets. (Spoiler alert: you should find that correlations are also persistent, but that this persistence is much noisier than we saw for volatility.)

You might also like to repeat the analysis for asset returns – you’ll find almost no persistence at all, other than a very weak negative relationship at the daily timescale in some assets.

5 thoughts on “How to Test the Assumption of Persistence”

  1. As Kris is likely aware, the simple way to quantify the dependence of a day’s range on the previous day’s range is the autocorrelation coefficient of order 1. You can also compute ACF(2), ACF(3), etc to measure dependence on observations further back.

    Reply
    • Yep for sure, that’s a great way to quantify the dependence of a time-series with its previous values. While the scatter plot is useful to see the noisiness of the relationship, the ACF is an efficient way to summarise the average relationship. Two ways of looking at the same thing, and they both offer slightly different views.

      Reply
  2. Enjoyed the article, thanks. The regression line is there, but logging the axes (scale_y_log10(), scale_x_log10()) helps (me, at least) to visually accept the relationship better when the points are clumped like that.

    Reply

Leave a Comment