Depdence
Introduction to Dependence Metrics
Recommended Prerequesites
- Probability
- Probability II
- Introduction to Multivariate Distributions
- Distribution Distance
General Types
Autocorrelation
Autocorrelation measures the correlation of a time series with a lagged version of itself.
Given a time series where , the autocorrelation at lag k is defined as:
Autocorrelation is used in time series analysis to detect trends, seasonality, and repeating patterns. It plays a key role in models like AR (Auto-Regressive) models, where current values are modeled as linear functions of past values.
Pairwise Independence
Pairwise Independence is a weaker condition of independence between random variables. Two variables are pairwise independent if knowing the outcome of one provides no information about the outcome of the other, but pairwise independence does not necessarily imply full independence across all variables.
Random variables are pairwise independent if for all :
Conditional Dependence
Conditional Independence refers to a situation where two random variables are independent, given the value of a third variable. Conditional independence is a key concept in Bayesian networks, graphical models, and statistical modeling, where it allows for simplified representations of complex dependencies.
Let be random variables. X and Y are said to be conditionally independent given Z if:
Metrics
Spearman Correlation
Spearman correlation assesses how well the relationship between two variables can be described using a monotonic function.
Let be a set of paired observations. This could be thought of as having been at the same time point, for example.
Compare the pairs at two indices. Label the pair as:
- Concordant if
- Discordant if
In plain English, this means that if X is greater at time i than at time j, then Y is concordant if Y is greater at time i than it was at time j.
If the two variables have the opposite relationship, it is discordant.
Now, let us define the counts for each type of pair; i.e.:
Then we define Kendall's Tau as
Mutual Information
Unlike Kendall's Tau and Spearman's Rho, which assess specific types of dependence (such as monotonic or rank-based), MI captures all forms of dependence, including non-linear and complex relationships.
where is the joint probability density function of X and Y; and are the marginal densities of X and Y, respectively.
Tail Dependence
Tail dependence refers to the behavior of extreme values in multivariate distributions. It captures the likelihood that extreme outcomes in one variable will be accompanied by extreme outcomes in another.
For two random variables and , the upper tail dependence coefficient is defined as the limit of the conditional probability that Y exceeds a high threshold, given that X also exceeds that threshold:
where and are the marginal cumulative functions of X and Y respectively, and and are the quantiles corresponding to the probability level u.
Similarly, the lower tail dependence coefficient is defined as:
A value of 0 indicates no tail dependence. A value of 1 implies perfect tail dependence.