Mathematics Exam  >  Mathematics Notes  >  Mathematics for IIT JAM, GATE, CSIR NET, UGC NET  >  Partial correlation, CSIR-NET Mathematical Sciences

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET PDF Download

In probability theory and statistics, partial correlation measures the degree of association between two random variables, with the effect of a set of controlling random variables removed. If we are interested in finding whether or to what extent there is a numerical relationship between two variables of interest, using their correlation coefficient will give misleading results if there is another, confounding, variable that is numerically related to both variables of interest. This misleading information can be avoided by controlling for the confounding variable, which is done by computing the partial correlation coefficient. This is precisely the motivation for including other right-side variables in a multiple regression.

For example, if we have economic data on the consumption, income, and wealth of various individuals and we wish to see if there is a relationship between consumption and income, failing to control for wealth when computing a correlation coefficient between consumption and income would give a misleading result, since income might be numerically related to wealth which in turn might be numerically related to consumption; a measured correlation between consumption and income might actually be contaminated by these other correlations. The use of a partial correlation avoids this problem.

Like the correlation coefficient, the partial correlation coefficient takes on a value in the range from –1 to 1. The value –1 conveys a perfect negative correlation controlling for some variables (that is, an exact linear relationship in which higher values of one variable are associated with lower values of the other); the value 1 conveys a perfect positive linear relationship, and the value 0 conveys that there is no linear relationship.

The partial correlation coincides with the conditional correlation if the random variables are jointly distributed as the multivariate normal, other elliptical, multivariate hypergeometric, multivariate negative hypergeometric, multinomial or Dirichlet distribution, but not in general otherwise.[1]

 

 

Example 

Suppose we have the following data on three variables, X, Y, and Z:

X

Y

Z

2

1

0

4

2

0

15

3

1

20

4

1

These data have the feature that whenever Z = 0, X equals exactly twice Y, and whenever Z = 1, X is exactly 5 times Y. Thus, contingent on the value of Z, there is an exact relationship between X and Y; but the relationship cannot be said to be exact without reference to the value of Z.

In fact, if we compute the Pearson correlation coefficient between variables X and Y, the result is approximately 0.969, while if we compute the partial correlation between X and Y, using the formula given below, we find a partial correlation of 0.919. The computations were done using R with the following code.

> X = c(2,4,15,20)
> Y = c(1,2,3,4)
> Z = c(0,0,1,1)
> mm1 = lm(X~Z)
> res1 = mm1$residuals
> mm2 = lm(Y~Z)
> res2 = mm2$residuals
> cor(res1,res2)
[1] 0.919145
> cor(X,Y)
[1] 0.9695016

 

Formal definition 

Formally, the partial correlation between X and Y given a set of n controlling variables Z = {Z1, Z2, ..., Zn}, written ρXY·Z, is the correlation between the residuals eX and eY resulting from the linear regression of X with Z and of Y with Z, respectively. The first-order partial correlation (i.e. when n=1) is the difference between a correlation and the product of the removable correlations divided by the product of the coefficients of alienation of the removable correlations. The coefficient of alienation, and its relation with joint variance through correlation are available in Guilford (1973, pp. 344–345).[2]

 

Computation 

Using linear regression 

A simple way to compute the sample partial correlation for some data is to solve the two associated linear regression problems, get the residuals, and calculate the correlationbetween the residuals. Let X and Y be, as above, random variables taking real values, and let Z be the n-dimensional vector-valued random variable. We write xi, yi and zi to denote the ith of N i.i.d. observations from some joint probability distribution over real random variables X, Y and Z, with zi having been augmented with a 1 to allow for a constant term in the regression. Solving the linear regression problem amounts to finding (n+1)-dimensional regression coefficient vectors Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET and Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET such that

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

with N being the number of observations and (w, v) the scalar product between the vectors w and v.

The residuals are then

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

and the sample partial correlation is then given by the usual formula for sample correlation , but between these new derived values:

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

 

Using recursive formula 

It can be computationally expensive to solve the linear regression problems. Actually, the nth-order partial correlation (i.e., with |Z| = n) can be easily computed from three (n - 1)th-order partial correlations. The zeroth-order partial correlation ρXY·Ø is defined to be the regular correlation coefficient ρXY.

It holds, for any  Z0 ∈ Z, that

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

Naïvely implementing this computation as a recursive algorithm yields an exponential time complexity. However, this computation has the overlapping subproblems property, such that using dynamic programming or simply caching the results of the recursive calls yields a complexity of Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

Note in the case where Z is a single variable, this reduces to:

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

 

Using matrix inversion 

In Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET time, another approach allows all partial correlations to be computed between any two variables Xi and Xj of a set V of cardinality n, given all others, i.e., Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET, if the correlation matrix (or alternatively covariance matrix) Ω = (ωij), where ωij = ρXiXj, is positive definite and therefore invertible. If we define P = (pij)ijΩ−1, we have:

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

 

Interpretation 

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

Geometrical interpretation of partial correlation for the case of N=3 observations and thus a 2-dimensional hyperplane

 

Geometrical 

Let three variables X, Y, Z (where Z is the "control" or "extra variable") be chosen from a joint probability distribution over nvariables V. Further let vi, 1 ≤ i ≤ N, be N n-dimensional i.i.d. observations taken from the joint probability distribution over V. We then consider the N-dimensional vectors x (formed by the successive values of X over the observations), y (formed by the values of Y) and z (formed by the values of Z).

It can be shown that the residuals eX,i coming from the linear regression of X on Z, if also considered as an N-dimensional vector eX, have a zero scalar product with the vector z generated by Z. This means that the residuals vector lies on an (N–1)-dimensional hyperplane Sz that is perpendicular to z.

The same also applies to the residuals eY,i generating a vector eY. The desired partial correlation is then the cosine of the angle φ between the projections eX and eY of x and y, respectively, onto the hyperplane perpendicular to z.[3]:ch. 7

 

As conditional independence test 

With the assumption that all involved variables are multivariate Gaussian, the partial correlation ρXY·Z is zero if and only if Xis conditionally independent from Y given Z.[1] This property does not hold in the general case.

To test if a sample partial correlation Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NETimplies a true population partial correlation of 0, Fisher's z-transform of the partial correlation can be used:

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

The null hypothesis is H0 : ρXY.Z = 0, to be tested against the two-tail alternative Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET. We reject H0 with significance level α if:

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

where Φ(·) is the cumulative distribution function of a Gaussian distribution with zero mean and unit standard deviation, and N is the sample size. This z-transform is approximate and that the actual distribution of the sample (partial) correlation coefficient is not straightforward. However, an exact t-test based on a combination of the partial regression coefficient, the partial correlation coefficient and the partial variances is available.[4]

The distribution of the sample partial correlation was described by Fisher.[5]

 

Semipartial correlation (part correlation) 

The semipartial (or part) correlation statistic is similar to the partial correlation statistic. Both compare variations of two variables after certain factors are controlled for, but to calculate the semipartial correlation one holds the third variable constant for either X or Y but not both, whereas for the partial correlation one holds the third variable constant for both.[6] The semipartial correlation compares the unique variation of one variable (having removed variation associated with the Z variable(s)), with the unfiltered variation of the other, while the partial correlation compares the unique variation of one variable to the unique variation of the other.

The semipartial (or part) correlation can be viewed as more practically relevant "because it is scaled to (i.e., relative to) the total variability in the dependent (response) variable." [7]Conversely, it is less theoretically useful because it is less precise about the role of the unique contribution of the independent variable.

The absolute value of the semipartial correlation of X with Y is always less than or equal to that of the partial correlation of X with Y. The reason is this: Suppose the correlation of Xwith Z has been removed from X, giving the residual vector ex . In computing the semipartial correlation, Y still contains both unique variance and variance due to its association with Z. But ex , being uncorrelated with Z, can only explain some of the unique part of the variance of Y and not the part related to Z. In contrast, with the partial correlation, only ey (the part of the variance of Y that is unrelated to Z) is to be explained, so there is less variance of the type that ex cannot explain.

 

Use in time series analysis

In time series analysis, the partial autocorrelation function (sometimes "partial correlation function") of a time series is defined, for lag h, as

Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

This function is used to determine the appropriate lag length for an autoregression.

The document Partial correlation, CSIR-NET Mathematical Sciences | Mathematics for IIT JAM, GATE, CSIR NET, UGC NET is a part of the Mathematics Course Mathematics for IIT JAM, GATE, CSIR NET, UGC NET.
All you need of Mathematics at this link: Mathematics
556 videos|198 docs

FAQs on Partial correlation, CSIR-NET Mathematical Sciences - Mathematics for IIT JAM, GATE, CSIR NET, UGC NET

1. What is partial correlation?
Ans. Partial correlation is a statistical technique used to measure the relationship between two variables while controlling for the effects of one or more additional variables. It provides a way to assess the strength and direction of the relationship between two variables after removing the influence of other variables.
2. How is partial correlation different from ordinary correlation?
Ans. Ordinary correlation measures the relationship between two variables without considering the influence of any other variables. On the other hand, partial correlation measures the relationship between two variables while taking into account the effects of other variables. It allows us to determine the unique contribution of each variable in explaining the relationship between the two variables of interest.
3. How is partial correlation calculated?
Ans. Partial correlation is typically calculated using the formula: r12.3 = (r12 - r13 * r23) / sqrt((1 - r13^2) * (1 - r23^2)) where r12.3 represents the partial correlation between variables 1 and 2 while controlling for the influence of variable 3. r12, r13, and r23 represent the ordinary correlations between variables 1 and 2, variables 1 and 3, and variables 2 and 3, respectively.
4. What is the significance of partial correlation?
Ans. Partial correlation helps researchers to determine the relationship between two variables after removing the influence of other variables. It is particularly useful in situations where the relationship between two variables may be confounded by the effects of additional variables. By controlling for these confounding variables, researchers can gain a better understanding of the true relationship between the variables of interest.
5. In which fields is partial correlation commonly used?
Ans. Partial correlation is commonly used in various fields, including psychology, sociology, economics, and biology. It is particularly valuable in research studies where multiple variables need to be considered to understand the complex relationships between variables. Partial correlation allows researchers to isolate the unique contribution of each variable and determine its role in explaining the relationship between the variables of interest.
Explore Courses for Mathematics exam
Signup for Free!
Signup to see your scores go up within 7 days! Learn & Practice with 1000+ FREE Notes, Videos & Tests.
10M+ students study on EduRev
Related Searches

MCQs

,

shortcuts and tricks

,

UGC NET

,

CSIR-NET Mathematical Sciences | Mathematics for IIT JAM

,

CSIR-NET Mathematical Sciences | Mathematics for IIT JAM

,

UGC NET

,

Sample Paper

,

GATE

,

GATE

,

Previous Year Questions with Solutions

,

Partial correlation

,

Objective type Questions

,

CSIR NET

,

Exam

,

CSIR NET

,

Free

,

study material

,

video lectures

,

practice quizzes

,

Viva Questions

,

CSIR-NET Mathematical Sciences | Mathematics for IIT JAM

,

GATE

,

pdf

,

Extra Questions

,

Partial correlation

,

Partial correlation

,

past year papers

,

CSIR NET

,

Semester Notes

,

Important questions

,

ppt

,

UGC NET

,

mock tests for examination

,

Summary

;