Prediction Intervals and Functional Forms of Simple Linear Regressions (CFA Level 1): Prediction Intervals vs Confidence Intervals for the Mean and Understanding the Difference. Key definitions, formulas, and exam tips.
Let’s talk about something we all love messing with—predictions. If you’re like most of us, you probably got excited the first time you built a regression model that seemed to forecast outcomes—even if you were just fooling around with a small data set in your earlier classes. Now, in this part of your CFA studies, you’ll see how that forecasting power is refined in practice. You’ll also learn one of the biggest watch-outs in regression analysis: distinguishing between predicting an individual new outcome versus estimating the mean response at specific values of your independent variable. They look similar at first glance, but the intervals can differ a lot.
Moreover, you’ll discover that not every scenario calls for a plain old linear relationship: it might be exponential, or perhaps your variable changes on a percentage basis. That’s where functional forms—including log-linear, linear-log, and log-log—come into play. It’s all about customizing your regression to fit the data and the real-world relationships you’re trying to represent.
One of the coolest aspects of linear regression, especially when we tie it back to investments and finance, is turning the model into a forward-looking tool. Suppose you’ve developed a simple regression of monthly returns of a particular stock (Y) against an economic indicator (X). You might want to know, “Given a particular value of my economic indicator, what might the stock’s return be?”
But there’s a big difference between:
Confidence intervals tell you about where the true mean of Y (for a given X) is likely to lie. Prediction intervals, on the other hand, aim to capture the likely range of a single new observation of Y. Because a single new observation can be affected by more variation (think day-to-day volatility in returns), prediction intervals are wider.
When we talk about “wider,” it’s basically because we add the extra uncertainty of the idiosyncratic noise around any single measurement. It’s like telling a friend: “Hey, if you do something a hundred times, the average will likely be in this narrower band—but that single trial outcome might wander off a bit more.”
Let’s say that you have your simple linear regression:
(1) Y = β₀ + β₁X + ε,
where ε is your random error term (often assumed to be normally distributed with mean zero and variance σ²). If you’ve estimated β₀ and β₁ from a sample, you have an estimated regression line:
(2) Ŷₓ = b₀ + b₁x,
where b₀ and b₁ are your sample estimates. Now imagine you want to predict a new observation Yₓ₍new₎ at some specific value x₍*₎. The formula for the prediction interval at confidence level (1 – α) typically looks like this:
(3) Ŷₓ₍₎ ± t(α/2, n – 2) × √[Var(Ŷₓ₍₎) + σ²],
where:
So that extra “+ σ²” in the square root is what accounts for the additional variability in your new individual observation.
In contrast, if you want just the average response—say you’re not predicting an individual monthly return but the expected monthly return for many months—your interval might look like this:
(4) Ŷₓ₍₎ ± t(α/2, n – 2) × √[Var(Ŷₓ₍₎)],
where we do not add that extra σ² inside the square root. This typically leads to a narrower interval since you’re focusing on the mean rather than individual outcomes.
Let’s run through a small hypothetical. Suppose you’re analyzing a portfolio’s monthly return Y against a single factor X (maybe it’s a lagged economic indicator, or the monthly return of a broad market index). After running the regression, you get:
Now let’s say you want to predict your portfolio’s return if X = 1.5%. So Ŷ₍1.5%₎ = 0.5% + 0.9(1.5%) = 0.5% + 1.35% = 1.85%.
For the mean response (confidence interval), you might end up with something like:
Ŷ₍1.5%₎ ± t(α/2, 28) × SE(Ŷ₍1.5%₎),
If the standard error of Ŷ₍1.5%₎ is, say, 0.3%, then a 95% CI might be 1.85% ± (2.05 × 0.3%) = 1.85% ± 0.615%, or approximately [1.235%, 2.465%].
For a new observation’s prediction interval, you have:
Ŷ₍1.5%₎ ± t(α/2, 28) × √[Var(Ŷ₍1.5%₎) + σ²].
Since σ² = 0.0004 (0.02²), you’re now adding that extra variance. The standard error might jump to 0.36% → 1.85% ± (2.05 × 0.36%) = 1.85% ± 0.738%, or approximately [1.112%, 2.588%]. That’s a wider interval, reflecting that a single month’s actual return can deviate more than the average predicted return.
One reason we keep harping on residual variance (σ²) is that it drives the spread in both types of intervals. In finance, returns often exhibit volatility clustering and potential heteroskedasticity (variance that changes over time). So you typically want to check the assumptions of your regression, making sure they’re reasonably valid for standard interval formulas to hold.
We’ve been working with Y = β₀ + β₁X for simplicity, but it’s not always the best shape for your data. Sometimes, log transformations or other modifications capture the relationship better. Let’s highlight some widely used functional forms:
(5) Y = β₀ + β₁X.
This is the standard approach. You assume that for every 1-unit change in X, Y changes by β₁ units—straight and direct. In finance, you might use a linear model to explain how the excess return on a commodity might change with an index drawdown, or how net income might change with sales.
(6) ln(Y) = β₀ + β₁X.
This implies that Y = e^(β₀ + β₁X). In many financial contexts (for instance, forecasting certain cost structures or certain risk premiums), an exponential-like growth pattern might ring truer. If β₁ is positive, Y grows exponentially as X increases linearly. A typical scenario is if Y is a price or a rating that grows in a multiplicative manner with X.
(7) Y = β₀ + β₁ ln(X).
Now X might be huge, or vary over multiple orders of magnitude—like AUM (Assets Under Management) at a fund. The relationship is: a small percentage change in X leads to a linear change in Y. This can be used if you suspect that Y changes systematically for each log-step in X. For instance, you might investigate how a bond’s price changes for each doubling of trade volume.
(8) ln(Y) = β₀ + β₁ ln(X).
This model is a favorite for measuring elasticity. If Y is an economic variable like consumer spending, and X is an income measure, then β₁ is directly interpreted as “if X changes by 1%, Y changes by approximately β₁%.” In asset pricing, you might similarly interpret the percentage change in price with respect to the percentage change in a macro factor.
This is where your expertise—and sometimes your intuition—comes into play. In my early days as a junior analyst, I remember simply plugging data into a plain vanilla linear regression, crossing my fingers, and hoping the line of best fit told the whole story. More often than not, it didn’t. Over time, I learned a few essentials:
Below is a quick Mermaid diagram to visualize how the dependent variable (Y) might respond to changes in the independent variable (X) for different functional forms. These relationships often help clarify when you might pick log-linear versus linear-log, etc.
flowchart LR
A["X <br/>(Independent)"] --> B["Linear: Y = β₀ + β₁X"]
A["X <br/>(Independent)"] --> C["Log-Linear: ln(Y) = β₀ + β₁X"]
A["X <br/>(Independent)"] --> D["Linear-Log: Y = β₀ + β₁ ln(X)"]
A["X <br/>(Independent)"] --> E["Log-Log: ln(Y) = β₀ + β₁ ln(X)"]
Each shape in this diagram emphasizes how the same X can be mapped to Y under different assumptions about the underlying relationship.
Forecasting Earnings per Share (EPS):
If a company’s earnings growth is roughly exponential (perhaps compounding at a certain rate), a log-linear model might better capture the relationship between time (X) and ln(EPS).
Portfolio Return Behavior:
Some practitioners use log-returns instead of simple returns because of the properties of compounding and the multiplicative nature of investment growth. A log-log model can help estimate elasticity of changes in a portfolio’s returns relative to changes in a benchmark or macroeconomic variable.
Comparative Statics in Bond Pricing:
If bond price changes are believed to be linear in the log of interest rates, then you’d rewrite your regression to reflect that. For instance, you might do a linear-log model if you suspect with each doubling of the yield, the bond price changes by a certain fixed number of basis points (it might be approximate, but can be tested).
Even though simple linear regression might appear straightforward compared to advanced topics you’ll tackle (like multi-factor models or simulation-based approaches), it remains foundational. On the CFA exam, especially for Level III (though building from these essential Level I details), you may see item-set questions requiring you to:
The exam often tests your ability to choose the correct approach for forecasting and to interpret the intervals properly—particularly under exam pressure, where you might forget that a single new observation’s forecast interval should be wider than the interval for the average predicted value.
Look for Key Phrases:
Watch the Functional Form:
Do a Sanity Check on Exponential Models:
Let’s suppose you have the following model to predict next month’s percentage return on an emerging markets fund (Y) from its expense ratio (X). You suspect that beyond a certain point, the fund’s expense ratio might erode returns exponentially. So you choose a log-linear model:
ln(Y) = β₀ + β₁X.
Your estimated coefficients from some historical data are:
Now if X = 1.2%, you get:
ln(Ŷ) = 3.40 – 0.35 × 1.2 = 3.40 – 0.42 = 2.98,
Ŷ = e^(2.98) ≈ 19.64%.
To form a confidence or prediction interval, you’d consider not only the standard error of the guess (Var(ln(Ŷ))) but decide: Are we talking about a single month’s return (new observation) or the expected average return at that expense ratio? For a single new month’s observation, you add the noise term inside the exponent. Realizing that ln(Y) is normally distributed with residual variance (0.12)², you might do a quick back-of-the-envelope calculation for a 95% interval by going:
2.98 ± (2.01 × 0.12) → [2.98 ± 0.2412],
Exponentiating the endpoints gives you a range for Y in percentage terms. In practice, you also have to consider whether the question wants a “mean response” in log-space or an “actual single observation,” which again leads to a slightly different formula (you’d add σ² to Var(Ŷₓ) if you’re dealing with a new observation).
Important Notice: FinancialAnalystGuide.com provides supplemental CFA study materials, including mock exams, sample exam questions, and other practice resources to aid your exam preparation. These resources are not affiliated with or endorsed by the CFA Institute. CFA® and Chartered Financial Analyst® are registered trademarks owned exclusively by CFA Institute. Our content is independent, and we do not guarantee exam success. CFA Institute does not endorse, promote, or warrant the accuracy or quality of our products.