Exam FRM Level 1 Ultimate Guide (Books 1 & 2)

Estimated reading time: 41 minutes



We present our extensive summary for books 1 and 2 in the Level 1 FRM course.

Note: You may download this entire content on our website’s shop page, free of charge.


FRM Level 1 Book 1 – Foundations of Risk Management

How we think about risk is the biggest determinant of whether we recognize or assess them properly

We isolate ten risk management building blocks

  1. The risk management process
  2. Identifying risk: knowns and unknowns
  3. Expected loss, unexpected loss, and tail loss
  4. Risk factor breakdown
  5. Structural change: from tail risk to systemic crisis
  6. Human agency and conflicts of interest
  7. Typology of risks and risk interactions 8. Risk aggregation
  8. Balancing risk and reward
  9. Enterprise risk management (ERM)

Each key risk type demands a specific set of skills and its own philosophical approach


Market Risk

Market prices and rates continually change, creating the potential for loss

Market risk is driven by:

General market risk:

This is the risk that an asset class will fall in value, leading to a fall in portfolio value

Specific market risk:

This is the risk that an individual asset will fall in value more than the general asset class


Credit Risk

Credit risk arises from the failure of one party to fulfill its financial obligations to another party

Examples of credit risk include:

  • A debtor fails to pay interest or principal on a loan (bankruptcy risk)
  • An obligor or counterparty is downgraded (downgrade risk)
  • A counterparty to a market trade fails to perform (counter-party risk)

Credit risk is driven by:

  • The probability of default of the obligor or counterparty
  • The exposure amount at the time of default
  • The amount that can be recovered in the event of a default


Liquidity Risk

Liquidity risk is used to describe two separate kinds of risk:

  • Funding liquidity risk and Market liquidity risk

Funding liquidity risk is the risk that covers the risk that a firm cannot access enough liquid cash and assets to meet its obligations

Market liquidity risk, sometimes known as trading liquidity risk, is the risk of a loss in asset value when markets temporarily seize up


Operational Risk

Operational risk can be defined as the “risk of loss resulting from inadequate or failed internal processes, people, and systems or from external events.”

It includes legal risk, but excludes business, strategic, and reputational risk

Business risks includes all the usual worries of firms, such as customer demand, pricing decisions, supplier negotiations, and managing product innovation


Strategic Risk

Strategic risk is distinct from business risk

Strategic risk involves making large, long-term decisions about the firm’s direction

Strategic risk is often accompanied by major investments of capital, human resources, and management reputation


Reputation Risk

Reputation risk is the danger that a firm will suffer a sudden fall in its market standing or brand with economic consequences


The Risk Management Process

  • Identify the risk
  • Analyze and Measure the risk
  • Assess the impact
  • Manage the risk

VaR was a popular risk aggregation measure in the years leading up to the global financial crisis

Financial risk managers have long recognized that they must overcome silo-based risk management process to build a broad picture of risk across risk types and business lines

We know this as Enterprise Risk Management (ERM)


Managing Financial Risk

The risk management process as a road-map can be seen:

  1. Identify risk appetite
  2. Map risks, make choices
  3. Operationalize risk appetite
  4. Implement
  5. Re-evaluate regularly to capture changes


Risk Hedging

Just because a risk can be hedged does not mean that it should be hedged

Hedging is simply a tool and, like any tool, it has limitations

Risk appetite describes the amount and types of risk a firm is willing to accept

This is in contrast to risk capacity, which describes the maximum amount of risk a firm can absorb


Governance of Risk

Corporate governance is the way in which companies are run

Corporate governance describes the roles and responsibilities of a firm’s shareholders, board of directors, and senior management

Basel III designed a macro-prudential overlay intended to reduce systemic risk and lessen pro-cyclicality

Ultimately, only four basic choices need to be made in the management of corporate risk:

  1. The choice to undertake or not to undertake certain activities
  2. The choice to transfer or not transfer either all or part of a given risk to a third party
  3. The choice to preemptively mitigate risk through early detection and prevention
  4. The choice to assume or not assume risk


The Risk Appetite Statement

Publishing a risk appetite statement (RAS) is an important component of corporate governance

The Financial Stability Board (FSB) describes an RAS as “a written articulation of the aggregate level and types of risk that a firm will accept or avoid in order to achieve its business objectives.”

The RAS includes both qualitative and quantitative statements


Internal Auditors

Internal auditors are responsible for:

  • Reviewing monitoring procedures
  • Tracking the progress of risk management system upgrades
  • Assessing the adequacy of application controls
  • Affirming the efficacy of vetting processes


Credit Risk Transfer Mechanisms

The core risk exposure for banks is credit risk

Banks have long had several ways to reduce their exposure to credit risk:

  • Purchasing insurance from a third-party guarantor/underwriter
  • Netting of exposures to counterparties
  • Marking-to-market/margining
  • Requiring collateral be posted
  • Termination/Put options
  • Reassignment of a credit exposure



Securitization involves the repackaging of loans and other assets into new securities that can then be sold in the securities markets

Securitization eliminates a substantial amount of risk (i.e., liquidity, interest rate, and credit risk) from the originating bank’s balance sheet when compared to the traditional buy-and-hold strategy


Modern Portfolio Theory and Capital Asset Pricing Model

A “rational investor” is an investor who is risk averse and seeks to maximize utility

Markowitz demonstrated that a rational investor should evaluate potential portfolio allocations based upon the associated means and variances of the expected rate of return distributions

The theory also assumes:

  • Capital markets are perfect
  • There are no taxes or transaction costs
  • All traders have costless access to all available information
  • Perfect competition exists among all market participants
  • Returns are normally distributed

According to Markowitz, the level of investment in a particular financial asset should be based upon that asset’s contribution to the distribution of the portfolio’s overall return (as measured by the mean and variance)

In other words, what matters is the covariability of the asset’s return with respect to the return of the overall portfolio

Along the efficient frontier, the only way to achieve a higher expected rate of return is by increasing the riskiness of the portfolio

The CAPM model shows that market equilibrium is achieved when all investors hold portfolios consisting of the riskless asset and the market portfolio


Arbitrage Pricing Theory and Multifactor Models

The capital asset pricing model (CAPM) is a single-factor model that describes an asset’s expected rate of return as a linear function of the market’s risk premium above a risk-free rate

Beta is the coefficient (i.e., the slope) of this relationship


Arbitrage Pricing Theory

Arbitrage Pricing Theory (APT) is based on the reasoning behind CAPM. However, it explains an asset’s expected rate of return as a linear function of several market factors

Arbitrage Pricing Theory assumes that there are no arbitrage opportunities

APT suggests that multiple factors can help explain the expected rate of return on a risky asset

These factors may include indices on stocks, bonds, and commodities, as well as macroeconomic factors


Effective Data Aggregation and Risk Reporting

Effective risk analysis requires sufficient and high-quality data

This makes data a major asset in today’s world

A bank with a limited ability to integrate data will have difficulties in satisfying the Basel principles

A key challenge is collecting data from the various internal and external sources and feeding it into risk analytics systems

Firms need to monitor their data on an ongoing basis to ensure accuracy and integrity

Risk data should be complete, reconciled with sources, and include all material risk disclosures at a granular level


The Key Governance Principles

Principle 1: Governance

Principle 2: Data architecture and IT infrastructure

Principle 3: Accuracy and Integrity

Principle 4: Completeness

Principle 5: Timeliness

Principle 6: Adaptability

Principle 7: Accuracy

Principle 8: Comprehensiveness

Principle 9: Clarity and usefulness

Principle 10: Frequency

Principle 11: Distribution

A study from PwC showed higher performance for compliance with Principles 7–11 (risk reporting) compared to Principles 3–6 (data aggregation)

Principles 1 (governance) and 2 (data architecture and infrastructure) have poor compliance rates


Enterprise Risk Management and Future Trends

At the enterprise level, risks may:

  • Negate each other (e.g., through netting and diversification) or
  • Exacerbate each other (e.g., through risk concentrations, contagion, and cross-over risks)

Enterprise risk management (ERM) applies the perspective and resources at the top of the enterprise to manage the entire portfolio of risks and account for them in strategic decisions

Another important feature of ERM is that it supports a consistent approach


ERM Benefits

Top ten benefits of ERM:

Helps firms define and adhere to risk appetites

Focuses oversight on most threatening risks

Identifies enterprise-scale risks generated at business line level

Manages risk concentrations across the enterprise

Manages emerging enterprise risks (e.g., cyber risk, AML (anti-money laundering) risk)

Supports regulatory compliance and stakeholder reassurance

Helps firms to understand risk-type correlations and cross-over risks

Optimizes risk transfer expenses in line with risk scale and total cost

Incorporates stress scenario capital costs into pricing and business decisions

Incorporates risk into business model selection and strategic decisions


Risk Culture

Risk culture can be thought of as the set of goals, values, beliefs, procedures, customs, and conventions that influence how staff create, identify, manage, and think about risk within an enterprise

The Financial Stability Board (FSB) has specified four key risk culture indicators:

  1. Accountability
  2. Effective communication and challenge
  3. Incentives
  4. Tone from the top


Learning from Financial Disasters

Over the last century, interest rate risk has caused the failure of individual firms as well as entire industries within the financial services sector

To mitigate interest rate risk, firms must manage their balance sheet structure such that the effect of any interest rate movement on assets remains highly correlated with the effect on liabilities

Funding liquidity risk can stem from external market conditions (e.g., during a financial crisis) or from structural problems within a bank’s balance sheet

Rogue trading can cause major financial institutions to collapse (as seen in the case of Barings Bank/Nick Leeson)

A main lesson from the Barings collapse is that reporting and monitoring of positions and risks (i.e., back-office operations) must be separated from trading

The case of Continental Illinois Bank is an example of how internal credit portfolio problems can precipitate a funding liquidity crisis

The 2007 failure of mortgage bank Northern Rock is a recent illustration of liquidity risk arising from structural weaknesses in a bank’s business model

In this case, a combination of an excessive use of short-term financing for long-term assets and a sudden loss of market confidence triggered a funding liquidity crisis that rapidly led to disaster

MGRM (Metallgesellschaft AG) was exposed to curve risk (i.e., the risk of shifts in the price curve between backwardation and contango)

Additionally, the firm was exposed to basis risk resulting from deviations between short-term prices and long-term prices

Long Term Capital Management (LTCM) failure reflected its inability to anticipate the dramatic increase in correlations and volatilities and the sharp drop in liquidity that can occur during an extreme crisis

LTCM also succumbed to an internal liquidity crunch brought on by large margin calls on its futures holdings

LTCM made heavy use of a Value-at-Risk (VaR) model as part of its risk control

Enron used “creative” (i.e., fraudulent) accounting practices to hide flaws in its actual financial performance

An example of Enron’s duplicity is a scheme by which the firm would build a physical asset and then immediately declare a projected mark-to-market profit on its books


Model Risk

Model risk can stem from using an incorrect model, incorrectly specifying a model, and/or using insufficient data and incorrect estimators

Banks may mitigate funding liquidity risk by reducing the maturity of their assets

VaR is a measure of the worst-case loss for a given normal market conditions

Cyber risk has become a critically important consideration in recent years


The Financial Crisis of 2007–2009

The cascade of events that came be known as the Great Financial Crisis of 2007–2009 (GFC) began with a downturn in the U.S. subprime mortgage market in the summer of 2007

The years preceding the crisis saw an exceptional boom in credit growth in the United States, a massive housing price bubble, and an excess of leverage in the financial system

February 2008 saw the nationalization of troubled U.K. mortgage lender Northern Rock, a victim of the first bank run that nation had experienced in 140 years

The following month, U.S. investment bank Bear Stearns was absorbed by J.P. Morgan Chase in a deal brokered by the U.S. Treasury Department and the Federal Reserve

The peak of the subprime crisis came in September 2008, which saw a cascade of events

  • Lehman Brothers declared bankruptcy
  • Morgan Stanley and Goldman Sachs, were converted to bank holding companies and became regulated by the Federal Reserve
  • Fannie Mae and Freddie Mac were nationalized
  • AIG was brought back from the brink of collapse via a USD 150 billion capital infusion by the U.S. Treasury and the Federal Reserve
  • In Europe, many countries had to step in to provide massive support to their banks
  • Dutch financial conglomerate Fortis was broken up and sold

Systemic risk is the risk that events at one firm, or in one market, can extend to other firms or markets

Systemic risk played a large role in exacerbating the impact of the crisis


GARP Code of Conduct

The GARP Code of Conduct sets forth principles of professional conduct for:

  • Global Association of Risk Professionals
  • Financial Risk Management and Energy Risk Professional certifications
  • Other GARP certification and diploma holders and candidates
  • GARP’s Board of Trustees, and Regional Directors
  • GARP Committee Members
  • GARP’s staff

These principles promote the highest levels of ethical conduct and disclosure and provide direction and support for both the individual practitioner and the risk management profession


Rules of Conduct

  • Professional Integrity and Ethical Conduct
  • Conflict of Interest
  • Confidentiality
  • Fundamental Responsibilities
  • General Accepted Practices


FRM Level 1 Book 2 – Quantitative Analysis

Fundamentals of Probability

A probability measures the likelihood that some event occurs

Probability is introduced through three fundamental principles:

  • The probability of any event is non-negative
  • The sum of the probabilities across all outcomes is one
  • The joint probability of two independent events is the product of the probability of each


Probability Ranges

Probabilities are always between 0 and 1 (inclusive)

An event with probability 0 never occurs

An event with a probability 1 always occurs


Conditional Probability

Conditional probability deals with computing a probability on that condition that another event occurs


Independent Events

Two events are independent if the probability that one event occurs does not depend on whether the other event occurs

Note that two types of independence—unconditional and conditional—do not imply each other

Events can be both unconditionally dependent (i.e., not independent) and conditionally independent

Similarly, events can be independent, yet conditional on another event they may be dependent


Random Variables

Probability can be used to describe any situation with an element of uncertainty

However, random variables restrict attention to uncertain phenomena that can be described with numeric values

This restriction allows standard mathematical tools to be applied to the analysis of random phenomena

Two functions are commonly used to describe the chance of observing various values from a random variable:

  1. The probability mass function (PMF) and
  2. The cumulative distribution function (CDF)

The PMF is particularly useful when defining the expected value of a random variable


Quantitative Moments

Moments are used to summarize the key features of random variables

A moment is the expected value of a carefully chosen function designed to measure a characteristic of a random variable

Four moments are commonly used in finance and risk management:

  1. The mean (which measures the average value of the random variable)
  2. The variance (which measures the spread/dispersion)
  3. The skewness (which measures asymmetry)
  4. The kurtosis (which measures the chance of observing a large deviation from the mean)


Quantile Function

The quantile function, which can be used to map a random variable’s probability to its realization, defines two moment-like measures:

  1. The median (which measures the central tendency of a random variable) and
  2. The interquartile range (which is an alternative measure of spread)


Random Variables

A discrete random variable assigns a probability to a set of distinct values

This set can be either finite or contain a countably infinite set of values

Random variables can be described precisely using mathematical functions

The set of values that the random variable may take is called the support of the function

In most applications in finance and risk management, the assumed distributions are continuous and without regions of zero probability


Common Univariate Random Variables

There are over two hundred named random variable distributions

Each of these distributions has been developed to explain key features of real-world phenomena

Risk managers model uncertainty in many forms, so this set includes both discrete and continuous random variables

There are three common discrete distributions:

  1. The Bernoulli
  2. The binomial
  3. The Poisson

The Bernoulli is a general-purpose distribution that is typically used to model binary events

The binomial distribution describes the sum of n independent Bernoulli random variables

The Poisson distribution is commonly used to model hazard rates, which count the number of events that occur in a fixed unit of time


Mixture Distributions

Mixture distributions are built using two or more distinct component distributions

A mixture is produced by randomly sampling from each component so that the mixture distribution inherits characteristics of each component

Mixtures can be used to build distributions that match important features of financial data


Normal Distributions

The normal distribution is the most commonly used distribution in risk management

The normal distribution is commonly referred to as a Gaussian distribution or a bell curve

A normal distribution has no skewness (because it is symmetrical) and a kurtosis of 3


Lognormal Distributions

The lognormal distribution is a simple transformation of a normal distribution and is the distribution underlying the famous Black-Scholes Merton model

A variable Y is said to be log-normally distributed if the natural logarithm of Y is normally distributed

In other words, if X = ln Y, then Y is log-normally distributed if and only if X is normally distributed


Chi-squared Distributions

The chi-squared distribution is frequently encountered when testing hypotheses about model parameters. It is also used when modeling variables that are always positive


Student’s t distribution

The Student’s t distribution is closely related to the normal, but it has heavier tails. The Student’s t distribution was originally developed for testing hypotheses using small samples

A Student’s t is a one-parameter distribution

This parameter, denoted by n, is also called the degrees of freedom parameter



The F is another distribution that is commonly encountered when testing hypotheses about model parameters

The F has two parameters known as the numerator and denominator degrees of freedom


Exponential Distributions

The exponential distribution uses a single parameter that determines both the mean and variance

The exponential distribution is closely related to the Poisson distribution


Exponential Variables

Exponential variables are “memoryless”, meaning that their distributions are independent of their histories


The Beta Distribution

The Beta distribution applies to continuous random variables with outcomes between 0 and 1

It is commonly used to model probabilities that naturally fall into this range


Multivariate Random Variables

Multivariate random variables are vectors of random variables

Multivariate random variables extend the concept of a single random variable to include measures of dependence between two or more random variables

Multivariate random variables are natural extensions of univariate random variables

These are defined using PMFs (for discrete variables) or PDFs (for continuous variables), which describe the joint probability of outcome combinations


Probability Mass Function

The probability mass function (PMF)/probability density function (PDF) for a bivariate random variable returns the probability that two random variables each take a certain value

The trinomial PMF has three parameters:

  1. n (i.e., the total number of experiments),
  2. p1, (i.e., the probability of observing outcome 1)
  3. p2 (i.e., the probability of observing outcome 2)



The expectation of a function of a bivariate random variable is defined analogously to that of a univariate random variable

Expectations are used to define moments of bivariate random variables in the same way that they are used to define moments for univariate random variables



The covariance is a measure of dispersion that captures how the variables move together

In a bivariate random variable, there are two variances and one covariance

Bivariate distributions can be transformed into either marginal or conditional distributions

A marginal distribution summarizes the information about a single variable and is simply a univariate distribution

A conditional distribution describes the probability of one random variable conditional on an outcome or a range of outcomes of another



Correlation measures the strength of the linear relationship between two variables and is always between -1 and 1

Correlation plays an important role in determining the benefits of portfolio diversification

When two random variables are independent, they must have zero correlation

However, if two random variables have zero correlation, they are not necessarily independent

Correlation is a measure of linear dependence

If two variables have a strong linear relationship (i.e., they produce values that lie close to a straight line), then they have a large correlation

If two random variables have no linear relationship, then their correlation is zero


Portfolio Return

The return on a portfolio depends on:

  • The distribution of the returns on the assets in the portfolio, and
  • The portfolio’s weights on these assets


Sample Moments

This segment describes how sample moments are used to estimate unknown population moments

When data are generated from independent identically distributed (iid) random variables, the mean estimator has several desirable properties:

  • It is (on average) equal to the population mean
  • As the number of observations grows, the sample mean becomes arbitrarily close to the population mean
  • The distribution of the sample mean can be approximated using a standard normal distribution

Data can also be used to estimate higher-order moments such as variance, skewness, and kurtosis

The first four (standardized) moments (mean, variance, skewness, and kurtosis) are widely used in finance and risk management to describe the key features of data sets



Quantiles provide an alternative method to describe the distribution of a data set

Quantile measures are particularly useful in applications to financial data because they are robust to extreme outliers



The mean estimator is a function that transforms data into an estimate of the population mean

An estimator is a mathematical procedure that calculates an estimate based on an observed data set

In contrast, an estimate is the value produced by an application of the estimator to data

The mean estimator is a function of random variables, and so it is also a random variable

The expected value of the mean estimator is the same as the population mean

The variance of the mean estimator depends on two values:

  1. The variance of the data
  2. The number of observations

The variance in the data is noise that obscures the mean

The more variable the data, the harder it is to estimate the mean of that data

The variance of the mean estimator decreases as the number of observations increases

So larger samples produce estimates of the mean that tend to be closer to the population mean


Means and Standard Deviations

Means and standard deviations are the most widely reported statistics. Their popularity is due to several factors:

  • The mean and standard deviation are often sufficient to describe the data
  • These two statistics provide guidance about the likely range of values that can be observed
  • The mean and standard deviation are in the same units as the data, and so can be easily compared

One challenge when using asset price data is the choice of sampling frequency


Mean Estimator Properties

The important properties of the mean estimator include the following:

  • The mean is unbiased
  • When the observed data are iid, the mean has a simple expression for its standard error
  • The mean estimator is BLUE
  • The mean estimator is consistent, and in large samples the estimated mean is close to the population mean
  • When the variance is finite, the distribution of the mean estimator can be approximated using the CLT


Hypothesis Testing

Hypothesis testing can be reduced to one universal question:

How likely is the observed data if the hypothesis is true?

Testing a hypothesis about a population parameter starts by specifying null hypothesis and an alternative hypothesis

The null hypothesis is an assumption about the population parameter

The alternative hypothesis specifies the population parameter values (i.e., the critical values) where the null hypothesis should be rejected

The critical values are determined by:

  • The distribution of the test statistic when the null hypothesis is true, and
  • The size of the test, which reflects our aversion to rejecting a null hypothesis that is in fact true

Observed data are used to construct a test statistic, and the value of the test statistic is compared to the critical values to determine whether the null hypothesis should be rejected


Components of Hypothesis Testing

A hypothesis test has six distinct components:

  1. The null hypothesis, which specifies a parameter value that is assumed to be true;
  2. The alternative hypothesis, which defines the range of values where the null should be rejected;
  3. The test statistic, which has a known distribution when the null is true;
  4. The size of the test, which captures the willingness to make a mistake and falsely reject a null hypothesis that is true;
  5. The critical value, which is a value that is compared to the test statistic to determine whether to reject the null hypothesis; and
  6. The decision rule, which combines the test statistic and critical value to determine whether to reject the null hypothesis

In some testing problems, the alternative hypothesis might not fully complement the null

The most common example of this is called a one-sided alternative, which is used when the outcome of interest is only above or only below the value assumed by the null

The test statistic is a summary of the observed data that has a known distribution when the null hypothesis is true

In an ideal world, a false (true) null would always (never) be rejected.

However, in practice there is a tradeoff between avoiding a rejection of a true null and avoiding a failure to reject a false null


Type I Errors

Rejecting a true null hypothesis is called a Type I error

The probability of committing a Type I error is known as the test size

The test size is chosen to reflect the willingness to mistakenly reject a true null hypothesis

  • The most common test size is 5%

The critical value depends on the distribution of the test statistic and defines a range of values where the null hypothesis should be rejected in favor of the alternative

  • This range is known as the rejection region

When the test statistic has a standard normal distribution, the critical value depends on both the size and the type of the alternative hypothesis (i.e., whether it is one-or two-sided)


Type II Errors

A Type II error occurs when the alternative is true, but the null is not rejected


Confidence Intervals

A confidence interval is a range of parameters that complements the rejection region

A 95% confidence interval contains the set of parameter values where the null hypothesis cannot be rejected when using a 5% test


P Values

A hypothesis test can also be summarized by its p-value

A p-value combines the test statistic, distribution of the test statistic, and the critical values into a single number that is always between 0 and 1

  • This value can always be used to determine whether a null hypothesis should be rejected
  • If the p-value is less than the size of the test, then the null is rejected

The p-value of a test statistic is equivalently defined as the smallest test size where the null is rejected

  • Any test size larger than the p-value leads to rejection, whereas using a test size smaller than the p-value fails to reject the null


Linear Regression

Linear regression is a widely applied statistical tool for modeling the relationship between random variables

Linear regression has many appealing features:

  • Closed-form estimators,
  • Interpretable parameters,
  • A flexible specification and
  • Can be adapted to a wide variety of problems

Regression analysis is the most widely used method to measure, model, and test relationships between random variables

  • It is widely used in finance to measure the sensitivity of a portfolio to common risk factors, estimate optimal hedge ratios for managing specific risks, and to measure fund manager performance

The bivariate linear regression model relates a dependent variable to a single explanatory variable

Regression is surprisingly flexible and can describe a wide variety of relationships

The Ordinary Least Squares (OLS) estimators, which have a simple moment-like structure and depend on the mean, variance, and covariance of the data


Dummy Random Variables

An important class of explanatory variable is known as a dummy

A dummy random variable is binary and only takes the value 0 or 1

Dummies are used to encode qualitative information (e.g., a bond’s country of origin)

A Dummy takes the value 1 when the observation has the quality and 0 if it does not

Dummies are also commonly constructed as binary transformations of other random variables

Example: a market direction dummy that encodes the return on the market as 1 if negative and 0 if positive


Regression with Multiple Explanatory Variables

Linear regression with a single explanatory variable provides key insights into OLS estimators and their properties

In practice, however, models typically use multiple variables where it is possible to isolate the unique contribution of each explanatory variable

A model built with multiple variables can also distinguish the effect of a novel predictor from the set of explanatory variables known to be related to the dependent variable


The k-Variate Regression Model

The k-variate regression model enables the coefficients to measure the distinct contribution of each explanatory variable to the variation in the dependent variable


The Fama-French Model

The Fama-French three-factor model is a leading example of a multi-factor approach

The Fama-French three-factor model expands upon CAPM by including two additional factors:

  • The size factor (which captures the propensity of small-cap firms to generate higher returns than large-cap firms) and
  • The value factor (which measures the additional return that value firms earn above growth firms)

Controls are explanatory variables that are known to have a clear relationship with the dependent variable


Model Fit

Model fit is assessed using R2, which measures the ratio of the variation explained by the model to the total variation in the data

While intuitive, this measure suffers from some important limitations:

  • It never decreases when an additional variable is added to a model and
  • It is not interpretable when the dependent variable changes.

The adjusted R2 partially addresses the first of these concerns


Regression Diagnostics

Ideally, a model should include all variables that explain the dependent variable and exclude all that do not

In practice, achieving this goal is challenging

Once a model has been selected, the specification should be checked for any obvious deficiencies

Omitting explanatory variables that affect the dependent variable creates biased coefficients

Including irrelevant variables does not bias coefficients

Determining whether a variable should be included in a model reflects a bias-variance tradeoff

Large models that include all conceivable explanatory variables are likely to have coefficients that are unbiased


Omitted Variables

An omitted variable is one that has a non-zero coefficient but is not included in a model

Omitting a variable has two effects:

First, the remaining variables absorb the effects of the omitted variable attributable to common variation

Second, the estimated residuals are larger in magnitude than the true shocks


Extraneous Variables

An extraneous variable is one that is included in the model but is not needed

This type of variable has a true coefficient of 0 and is consistently estimated to be 0 in large samples

The choice between omitting a relevant variable and including an irrelevant variable is ultimately a tradeoff between bias and variance

Larger models tend to have a lower bias but they also have less precise estimated parameters

Models with few explanatory variables have less estimation error but are more likely to produce biased parameter estimates

The bias-variance tradeoff is the fundamental challenge in variable selection



Homoscedasticity is one of the five assumptions used to determine the asymptotic distribution of an OLS estimator

It requires that the variance is constant and so does not systematically vary with any of the explanatory variables

When this is not the case, then the residuals are heteroskedastic


Heteroskedastic Residuals

When residuals are heteroskedastic, the standard errors can be estimated using White’s estimator (also called Eiker-White in some software packages)

Parameters can be tested using t-tests by using White’s standard error in the place of standard error used for homoscedastic data

On the other hand, F-tests, are not as easy to adjust for heteroskedasticity and so caution is required when testing multiple hypotheses if the shocks are heteroskedastic



Multicollinearity occurs when one or more explanatory variables can be substantially explained by the other(s)

Multicollinearity differs from perfect collinearity, where one of the variables is perfectly explained by the others

Multicollinearity is a common problem in finance and risk management because many regressors are simultaneously determined by and sensitive to the same news


Residual Plots

Residual plots are standard methods used to detect deficiencies in a model specification

An ideal model would have residuals that are not systematically related to any of the included explanatory variables


Stationary Time Series

Time-series analysis is a fundamental tool in finance and risk management

Many key time series (e.g., interest rates and spreads) have predictable components

Building accurate models allows past values to be used to forecast future changes in these series

A time series can be decomposed into three distinct components:

  • The trend, which captures the changes in the level of the time series over time
  • The seasonal component, which captures predictable changes in the time series according to the time of year
  • The cyclical component, which captures the cycles in the data

Whereas the first two components are deterministic, the third component is determined by both the shocks to the process and the memory (i.e., persistence) of the process

A time series is covariance-stationary if its first two moments do not change across time


Linear Processes

Any time series that is covariance-stationary can be described by a linear process

While linear processes are very general, they are also not directly applicable to modeling

Two classes of models are used to approximate general linear processes:

  • Autoregressions (AR)
  • Moving averages (MAs)

The ability of a model to forecast a time series depends crucially on whether the past resembles the future



Stationarity is a key concept that formalizes the structure of a time series and justifies the use of historical data to build models

Covariance stationarity depends on the first two moments of a time series:

  1. The mean
  2. The autocovariances

White noise is the fundamental building block of any time-series model

White noise processes have three properties:

  1. Mean zero
  2. Constant and finite variance
  3. No autocorrelation or autocovariance

The lack of correlation is the essential characteristic of a white noise process and plays a key role in the estimation of time-series model parameters

Dependent white noise relaxes the iid assumption while maintaining the three properties of white noise

Autoregressive models are the most widely applied time-series models in finance and economics


Non-Stationary Time Series

Covariance-stationary time series have means, variances, and autocovariances that do not depend on time

Any time series that is not covariance-stationary is non-stationary

This segment covers the three most pervasive sources of non-stationarity in financial and economic time series:

  • Time trends
  • Seasonalities
  • Unit roots (more commonly known as random walks)

Time trends are the simplest deviation from stationarity

Time trend models capture the propensity of many time series to grow over time



Seasonalities induce non-stationary behavior in time series by relating the mean of the process to the month or quarter of the year

Seasonalities can be modeled in one of two ways:

  1. Shifts in the mean that depend on the period of the year, or
  2. An annual cycle where the value in the current period depends on the shock in the same period in the previous year


Random Walks

Random walks (also called unit roots) are the most pervasive form of non-stationarity in financial and economic time series

All non-stationary time series contain trends that may be deterministic or stochastic

For deterministic trends (e.g., time trends and deterministic seasonalities), knowledge of the period is enough to measure, model, and forecast the trend

On the other hand, random walks are the most important example of a stochastic trend

A time series that follows a random walk depends equally on all past shocks

The correct approach to modeling time series with trends depends on the source of the non-stationarity

If a time series only contains deterministic trends, then directly capturing the deterministic effects is the best method to model the data

Unit roots generalize random walks by adding short-run stationary dynamics to the long-run random walk


Measuring Returns, Volatility, and Correlation

Financial asset return volatilities are not constant, and how they change can have important implications for risk management

Capturing the dependence among assets in a portfolio is also a crucial step in portfolio construction

In portfolios with many assets, the distribution of the portfolio return is predominantly determined by the dependence between the assets held

  • If the assets are weakly related, then the gains to diversification are large, and the chance of experiencing an exceptionally large loss should be small
  • If the assets are highly dependent, especially in their tails, then the probability of a large loss may be surprisingly high

The volatility of a financial asset is usually measured by the standard deviation of its returns

Implied volatility is an alternative measure that is constructed using option prices

Both put and call options have payouts that are nonlinear functions of the underlying price of an asset


The Black-Scholes-Merton Model

The most well-known expression for determining the price of an option is the Black-Scholes-Merton model

The Black-Scholes-Merton model relates the price of a call option to:

  • the interest rate of a riskless asset
  • the current asset price
  • the strike price
  • the time until maturity and
  • the annual variance of the return

All values in the Black-Scholes-Merton model, including the call price, are observable except the volatility

The Black-Scholes-Merton option pricing model uses several simplifying assumptions that are not consistent with how markets actually operate


Normal Distributions

A normal distribution is symmetric and thin-tailed, and so has no skewness or excess kurtosis

However, many return series are both skewed and fat-tailed

The Jarque-Bera (JB) test statistic is used to formally test whether sample skewness and kurtosis are compatible with an assumption that the returns are normally distributed

An alternative method to understand the non-normality of financial returns is to study the tails

Normal random variables have thin tails

The Student’s t is an example of a widely used distribution with a power law tail

Linear correlation is insufficient to capture dependence when assets have nonlinear dependence

Researchers often use two alternative dependence measures:

  1. Rank correlation (also known as Spearman’s correlation) and
  2. Kendal’s (tau)

These statistics are correlation-like: both are scale invariant, have values that always lie between -1 and 1

They are zero when the returns are independent

They are positive (negative) when there is in increasing (decreasing) relationship between the random variables


Rank Correlation

Rank correlation is the linear correlation estimator applied to the ranks of the observations

Rank correlation has two distinct advantages over linear correlation:

  1. It is robust to outliers because only the ranks, not the values of X and Y, are used
  2. It is invariant with respect to any monotonic increasing transformation of Xi and Yi

Linear correlation is only invariant with respect to increasing linear transformations


Simulation and Bootstrapping

Simulation is an important practical tool in modern risk management with a wide variety of applications

Examples of these applications include:

  • Computing the expected payoff of an option
  • Measuring the downside risk in a portfolio
  • Assessing estimator accuracy

Monte Carlo simulation is a simple approach to approximate the expected value of a random variable using numerical methods

Another important application of simulated methods is boot-strapping

Bootstrapping uses observed data to simulate from the unknown distribution generating the observed data

This is done by combining observed data with simulated values to create a new sample that is closely related to, but different from, the observed data

The key to understanding bootstrapping lies in one simple fact:

  • The unknown distribution being sampled from is the same one that produced the observed data

The bootstrap method avoids the specification of a model and instead makes the key assumption that the present resembles the past

Monte Carlo simulation and bootstrapping are closely related

  • Both methods use computer-generated values to numerically approximate an expected value


Simulation and Bootstrapping Differences

The fundamental difference between simulation and bootstrapping is the source of the simulated data:

When using simulation, the user specifies a complete data generating process (DGP) that is used to produce the simulated data

In bootstrapping, the observed data are used directly to generate the simulated data set without specifying a complete DGP

Monte Carlo experiments allow the finite sample distribution of an estimator to be tabulated and compared to its asymptotic distribution derived from the Central Limit Theorem (CLT)

Antithetic variates add a second set of random variables that are constructed to have a negative correlation with the iid variables used in the simulation

They are generated in pairs using a single uniform value

The biggest challenge when using simulation to approximate moments is the specification of the DGP

If the DGP does not adequately describe the observed data, then the approximation of the moment may be unreliable


Bootstrap Limitations

While bootstrapping is a useful statistical technique, it has its limitations. There are two specific issues that arise when using a bootstrap:

First, bootstrapping uses the entire data set to generate a simulated sample

The second limitation arises due to structural changes in markets so that the present is significantly different from the past

Both Monte Carlo simulation and bootstrapping suffer from the “Black Swan” problem—simulations generated using either method resemble the historical data

Bootstrapping is especially sensitive to this issue, and a bootstrap sample cannot generate data that did not occur in the sample

A good statistical model, on the other hand, should allow the possibility of future losses that are larger than those that have been realized in the past



Thanks so much for taking the time to read.

Whenever you are ready, try the following links for more information:


Success is near,

The QuestionBank Family