SlideShare a Scribd company logo
CorrelationThe correlation is one of the most common and most useful statistics. A correlation is a single number that describes the degree of relationship between two variables. Let's work through an example to show you how this statistic is computed.Correlation ExampleLet's assume that we want to look at the relationship between two variables, height (in inches) and self esteem. Perhaps we have a hypothesis that how tall you are effects your self esteem (incidentally, I don't think we have to worry about the direction of causality here -- it's not likely that self esteem causes your height!). Let's say we collect some information on twenty individuals (all male -- we know that the average height differs for males and females so, to keep this example simple we'll just use males). Height is measured in inches. Self esteem is measured based on the average of 10 1-to-5 rating items (where higher scores mean higher self esteem). Here's the data for the 20 cases (don't take this too seriously -- I made this data up to illustrate what a correlation is):PersonHeightSelf Esteem1684.12714.63623.84754.45583.26603.17673.88684.19714.310693.711683.512673.213633.714623.315603.416634.017654.118673.819633.420613.6Now, let's take a quick look at the histogram for each variable:And, here are the descriptive statistics:VariableMeanStDevVarianceSumMinimumMaximumRangeHeight65.44.4057419.41051308587517Self Esteem3.7550.4260900.18155375.13.14.61.5Finally, we'll look at the simple bivariate (i.e., two-variable) plot:You should immediately see in the bivariate plot that the relationship between the variables is a positive one (if you can't see that, review the section on types of relationships) because if you were to fit a single straight line through the dots it would have a positive slope or move up from left to right. Since the correlation is nothing more than a quantitative estimate of the relationship, we would expect a positive correlation.What does a \"
positive relationship\"
 mean in this context? It means that, in general, higher scores on one variable tend to be paired with higher scores on the other and that lower scores on one variable tend to be paired with lower scores on the other. You should confirm visually that this is generally true in the plot above.Calculating the CorrelationNow we're ready to compute the correlation value. The formula for the correlation is:We use the symbol r to stand for the correlation. Through the magic of mathematics it turns out that r will always be between -1.0 and +1.0. if the correlation is negative, we have a negative relationship; if it's positive, the relationship is positive. You don't need to know how we came up with this formula unless you want to be a statistician. But you probably will need to know how the formula relates to real data -- how you can use the formula to compute the correlation. Let's look at the data we need for the formula. Here's the original data with the other necessary columns:PersonHeight (x)Self Esteem (y)x*yx*xy*y1684.1278.8462416.812714.6326.6504121.163623.8235.6384414.444754.4330562519.365583.2185.6336410.246603.118636009.617673.8254.6448914.448684.1278.8462416.819714.3305.3504118.4910693.7255.3476113.6911683.5238462412.2512673.2214.4448910.2413633.7233.1396913.6914623.3204.6384410.8915603.4204360011.561663425239691617654.1266.5422516.8118673.8254.6448914.4419633.4214.2396911.5620613.6219.6372112.96Sum =130875.14937.685912285.45The first three columns are the same as in the table above. The next three columns are simple computations based on the height and self esteem data. The bottom row consists of the sum of each column. This is all the information we need to compute the correlation. Here are the values from the bottom row of the table (where N is 20 people) as they are related to the symbols in the formula:Now, when we plug these values into the formula given above, we get the following (I show it here tediously, one step at a time):So, the correlation for our twenty cases is .73, which is a fairly strong positive relationship. I guess there is a relationship between height and self esteem, at least in this made up data!Testing the Significance of a CorrelationOnce you've computed a correlation, you can determine the probability that the observed correlation occurred by chance. That is, you can conduct a significance test. Most often you are interested in determining the probability that the correlation is a real one and not a chance occurrence. In this case, you are testing the mutually exclusive hypotheses:Null Hypothesis: r = 0Alternative Hypothesis: r  0The easiest way to test this hypothesis is to find a statistics book that has a table of critical values of r. Most introductory statistics texts would have a table like this. As in all hypothesis testing, you need to first determine the significance level. Here, I'll use the common significance level of alpha = .05. This means that I am conducting a test where the odds that the correlation is a chance occurrence is no more than 5 out of 100. Before I look up the critical value in a table I also have to compute the degrees of freedom or df. The df is simply equal to N-2 or, in this example, is 20-2 = 18. Finally, I have to decide whether I am doing a one-tailed or two-tailed test. In this example, since I have no strong prior theory to suggest whether the relationship between height and self esteem would be positive or negative, I'll opt for the two-tailed test. With these three pieces of information -- the significance level (alpha = .05)), degrees of freedom (df = 18), and type of test (two-tailed) -- I can now test the significance of the correlation I found. When I look up this value in the handy little table at the back of my statistics book I find that the critical value is .4438. This means that if my correlation is greater than .4438 or less than -.4438 (remember, this is a two-tailed test) I can conclude that the odds are less than 5 out of 100 that this is a chance occurrence. Since my correlation 0f .73 is actually quite a bit higher, I conclude that it is not a chance finding and that the correlation is \"
statistically significant\"
 (given the parameters of the test). I can reject the null hypothesis and accept the alternative.The Correlation MatrixAll I've shown you so far is how to compute a correlation between two variables. In most studies we have considerably more than two variables. Let's say we have a study with 10 interval-level variables and we want to estimate the relationships among all of them (i.e., between all possible pairs of variables). In this instance, we have 45 unique correlations to estimate (more later on how I knew that!). We could do the above computations 45 times to obtain the correlations. Or we could use just about any statistics program to automatically compute all 45 with a simple click of the mouse.I used a simple statistics program to generate random data for 10 variables with 20 cases (i.e., persons) for each variable. Then, I told the program to compute the correlations among these variables. Here's the result:          C1       C2       C3       C4       C5       C6       C7       C8       C9      C10C1     1.000C2     0.274    1.000C3    -0.134   -0.269    1.000C4     0.201   -0.153    0.075    1.000C5    -0.129   -0.166    0.278   -0.011    1.000C6    -0.095    0.280   -0.348   -0.378   -0.009    1.000C7     0.171   -0.122    0.288    0.086    0.193    0.002    1.000C8     0.219    0.242   -0.380   -0.227   -0.551    0.324   -0.082    1.000C9     0.518    0.238    0.002    0.082   -0.015    0.304    0.347   -0.013    1.000C10    0.299    0.568    0.165   -0.122   -0.106   -0.169    0.243    0.014    0.352    1.000This type of table is called a correlation matrix. It lists the variable names (C1-C10) down the first column and across the first row. The diagonal of a correlation matrix (i.e., the numbers that go from the upper left corner to the lower right) always consists of ones. That's because these are the correlations between each variable and itself (and a variable is always perfectly correlated with itself). This statistical program only shows the lower triangle of the correlation matrix. In every correlation matrix there are two triangles that are the values below and to the left of the diagonal (lower triangle) and above and to the right of the diagonal (upper triangle). There is no reason to print both triangles because the two triangles of a correlation matrix are always mirror images of each other (the correlation of variable x with variable y is always equal to the correlation of variable y with variable x). When a matrix has this mirror-image quality above and below the diagonal we refer to it as a symmetric matrix. A correlation matrix is always a symmetric matrix.To locate the correlation for any pair of variables, find the value in the table for the row and column intersection for those two variables. For instance, to find the correlation between variables C5 and C2, I look for where row C2 and column C5 is (in this case it's blank because it falls in the upper triangle area) and where row C5 and column C2 is and, in the second case, I find that the correlation is -.166.OK, so how did I know that there are 45 unique correlations when we have 10 variables? There's a handy simple little formula that tells how many pairs (e.g., correlations) there are for any number of variables:where N is the number of variables. In the example, I had 10 variables, so I know I have (10 * 9)/2 = 90/2 = 45 pairs.Other CorrelationsThe specific type of correlation I've illustrated here is known as the Pearson Product Moment Correlation. It is appropriate when both variables are measured at an interval level. However there are a wide variety of other types of correlations for other circumstances. for instance, if you have two ordinal variables, you could use the Spearman rank Order Correlation (rho) or the Kendall rank order Correlation (tau). When one measure is a continuous interval level one and the other is dichotomous (i.e., two-category) you can use the Point-Biserial Correlation. For other situations, consulting the web-based statistics selection program, Selecting Statistics at http://trochim.human.cornell.edu/selstat/ssstart.htm.
Correlation Example
Correlation Example
Correlation Example
Correlation Example
Correlation Example
Correlation Example
Correlation Example
Correlation Example

More Related Content

PDF
9. parametric regression
PPT
regression and correlation
PPTX
correlation and regression
PPTX
Chapter 16: Correlation (enhanced by VisualBee)
PPTX
Correlation and regression
PPTX
Correlation and regression
PPTX
Correlation & Regression
PDF
correlation_and_covariance
9. parametric regression
regression and correlation
correlation and regression
Chapter 16: Correlation (enhanced by VisualBee)
Correlation and regression
Correlation and regression
Correlation & Regression
correlation_and_covariance

What's hot (20)

PPTX
What is Simple Linear Regression and How Can an Enterprise Use this Technique...
PPTX
Correlation and regression
PPT
Correlation and regression
PPTX
Simple linear regression
PPTX
Correlation and regression
PPTX
Correlation Statistics
PPT
Correlation and Regression
PPTX
Regression
PDF
Statistics- chapter4.pdf
PDF
Simple correlation
PDF
Correlation in Statistics
PPTX
Partial correlation
PPT
Linear Correlation
PPTX
8 correlation regression
PDF
Kendall's ,partial correlation and scatter plot
PPTX
Correlation and Regression
PPT
Ch 7 correlation_and_linear_regression
ODP
Correlation
PPTX
Presentation on Regression Analysis
PPT
Sumit presentation
What is Simple Linear Regression and How Can an Enterprise Use this Technique...
Correlation and regression
Correlation and regression
Simple linear regression
Correlation and regression
Correlation Statistics
Correlation and Regression
Regression
Statistics- chapter4.pdf
Simple correlation
Correlation in Statistics
Partial correlation
Linear Correlation
8 correlation regression
Kendall's ,partial correlation and scatter plot
Correlation and Regression
Ch 7 correlation_and_linear_regression
Correlation
Presentation on Regression Analysis
Sumit presentation
Ad

Viewers also liked (10)

PPT
7.2 characteristics and evolution of stars
PDF
Inquiry based social studies isacs 2013
PPTX
PPTX
Types of evolution notes
PPTX
inquiry aproach in Social Studies
PPTX
Correlation analysis
DOCX
Questionnaire for the survey of electronics market(for school/college projects)
PPT
Social science and natural science
PPTX
Correlation ppt...
7.2 characteristics and evolution of stars
Inquiry based social studies isacs 2013
Types of evolution notes
inquiry aproach in Social Studies
Correlation analysis
Questionnaire for the survey of electronics market(for school/college projects)
Social science and natural science
Correlation ppt...
Ad

Similar to Correlation Example (20)

PPT
Frequency Tables - Statistics
PPT
2-20-04.ppt
PPT
2-20-04.ppthjjbnjjjhhhhhhhhhhhhhhhhhhhhhhhh
DOCX
BUS 308 – Week 4 Lecture 2 Interpreting Relationships .docx
DOCX
BUS 308 – Week 4 Lecture 2 Interpreting Relationships .docx
DOCX
MLR Project (Onion)
DOCX
For this assignment, use the aschooltest.sav dataset.The d
DOCX
8 Statistical SignificanceOK, measures of association are one .docx
DOCX
assignment 2
PPTX
4. correlations
DOCX
Week 4 Lecture 10 We have been examining the question of equal p.docx
PPTX
Spearman after priory man
DOCX
Two-Variable (Bivariate) RegressionIn the last unit, we covered
PPTX
PDF
Artificial Intelligence (Unit - 8).pdf
PDF
Section 5 - Improve Phase pdf Lean Six sigma
PDF
Introduction to correlation and regression analysis
PPTX
Machine learning session4(linear regression)
PPTX
Measure of Association
PPTX
data analysis
Frequency Tables - Statistics
2-20-04.ppt
2-20-04.ppthjjbnjjjhhhhhhhhhhhhhhhhhhhhhhhh
BUS 308 – Week 4 Lecture 2 Interpreting Relationships .docx
BUS 308 – Week 4 Lecture 2 Interpreting Relationships .docx
MLR Project (Onion)
For this assignment, use the aschooltest.sav dataset.The d
8 Statistical SignificanceOK, measures of association are one .docx
assignment 2
4. correlations
Week 4 Lecture 10 We have been examining the question of equal p.docx
Spearman after priory man
Two-Variable (Bivariate) RegressionIn the last unit, we covered
Artificial Intelligence (Unit - 8).pdf
Section 5 - Improve Phase pdf Lean Six sigma
Introduction to correlation and regression analysis
Machine learning session4(linear regression)
Measure of Association
data analysis

More from OUM SAOKOSAL (20)

PPTX
Class Diagram | OOP and Design Patterns by Oum Saokosal
PPTX
Android app development - Java Programming for Android
PDF
Java OOP Programming language (Part 8) - Java Database JDBC
PDF
Java OOP Programming language (Part 7) - Swing
PDF
Java OOP Programming language (Part 6) - Abstract Class & Interface
PDF
Java OOP Programming language (Part 5) - Inheritance
PDF
Java OOP Programming language (Part 4) - Collection
PDF
Java OOP Programming language (Part 3) - Class and Object
PDF
Java OOP Programming language (Part 1) - Introduction to Java
PDF
Javascript & DOM - Part 1- Javascript Tutorial for Beginners with Examples
PDF
Aggregate rank bringing order to web sites
DOC
How to succeed in graduate school
PDF
Google
PDF
E miner
PDF
Data preparation for mining world wide web browsing patterns (1999)
PDF
Consumer acceptance of online banking an extension of the technology accepta...
DOCX
When Do People Help
DOC
Mc Nemar
DOC
Sem Ski Amos
PPT
Sem+Essentials
Class Diagram | OOP and Design Patterns by Oum Saokosal
Android app development - Java Programming for Android
Java OOP Programming language (Part 8) - Java Database JDBC
Java OOP Programming language (Part 7) - Swing
Java OOP Programming language (Part 6) - Abstract Class & Interface
Java OOP Programming language (Part 5) - Inheritance
Java OOP Programming language (Part 4) - Collection
Java OOP Programming language (Part 3) - Class and Object
Java OOP Programming language (Part 1) - Introduction to Java
Javascript & DOM - Part 1- Javascript Tutorial for Beginners with Examples
Aggregate rank bringing order to web sites
How to succeed in graduate school
Google
E miner
Data preparation for mining world wide web browsing patterns (1999)
Consumer acceptance of online banking an extension of the technology accepta...
When Do People Help
Mc Nemar
Sem Ski Amos
Sem+Essentials

Recently uploaded (20)

PPTX
POLYCYSTIC OVARIAN SYNDROME.pptx by Dr( med) Charles Amoateng
PPT
Breast Cancer management for medicsl student.ppt
PDF
NEET PG 2025 | 200 High-Yield Recall Topics Across All Subjects
PPTX
History and examination of abdomen, & pelvis .pptx
PDF
Rheumatoid arthritis RA_and_the_liver Prof AbdelAzeim Elhefny Ain Shams Univ...
PPT
CHAPTER FIVE. '' Association in epidemiological studies and potential errors
PPT
ASRH Presentation for students and teachers 2770633.ppt
PPTX
Note on Abortion.pptx for the student note
PPTX
Neurotransmitter, Types of neurotransmitters,Neurotransmitter function, Neur...
PPTX
15.MENINGITIS AND ENCEPHALITIS-elias.pptx
PPTX
Chapter-1-The-Human-Body-Orientation-Edited-55-slides.pptx
PDF
Neuro ED Bet Sexologist in Patna Bihar India Dr. Sunil Dubey
PPTX
CME 2 Acute Chest Pain preentation for education
PPTX
Gastroschisis- Clinical Overview 18112311
PPTX
Electromyography (EMG) in Physiotherapy: Principles, Procedure & Clinical App...
PPTX
Neuropathic pain.ppt treatment managment
PDF
Khadir.pdf Acacia catechu drug Ayurvedic medicine
PDF
Handout_ NURS 220 Topic 10-Abnormal Pregnancy.pdf
PPTX
neonatal infection(7392992y282939y5.pptx
PDF
Therapeutic Potential of Citrus Flavonoids in Metabolic Inflammation and Ins...
POLYCYSTIC OVARIAN SYNDROME.pptx by Dr( med) Charles Amoateng
Breast Cancer management for medicsl student.ppt
NEET PG 2025 | 200 High-Yield Recall Topics Across All Subjects
History and examination of abdomen, & pelvis .pptx
Rheumatoid arthritis RA_and_the_liver Prof AbdelAzeim Elhefny Ain Shams Univ...
CHAPTER FIVE. '' Association in epidemiological studies and potential errors
ASRH Presentation for students and teachers 2770633.ppt
Note on Abortion.pptx for the student note
Neurotransmitter, Types of neurotransmitters,Neurotransmitter function, Neur...
15.MENINGITIS AND ENCEPHALITIS-elias.pptx
Chapter-1-The-Human-Body-Orientation-Edited-55-slides.pptx
Neuro ED Bet Sexologist in Patna Bihar India Dr. Sunil Dubey
CME 2 Acute Chest Pain preentation for education
Gastroschisis- Clinical Overview 18112311
Electromyography (EMG) in Physiotherapy: Principles, Procedure & Clinical App...
Neuropathic pain.ppt treatment managment
Khadir.pdf Acacia catechu drug Ayurvedic medicine
Handout_ NURS 220 Topic 10-Abnormal Pregnancy.pdf
neonatal infection(7392992y282939y5.pptx
Therapeutic Potential of Citrus Flavonoids in Metabolic Inflammation and Ins...

Correlation Example

  • 1. CorrelationThe correlation is one of the most common and most useful statistics. A correlation is a single number that describes the degree of relationship between two variables. Let's work through an example to show you how this statistic is computed.Correlation ExampleLet's assume that we want to look at the relationship between two variables, height (in inches) and self esteem. Perhaps we have a hypothesis that how tall you are effects your self esteem (incidentally, I don't think we have to worry about the direction of causality here -- it's not likely that self esteem causes your height!). Let's say we collect some information on twenty individuals (all male -- we know that the average height differs for males and females so, to keep this example simple we'll just use males). Height is measured in inches. Self esteem is measured based on the average of 10 1-to-5 rating items (where higher scores mean higher self esteem). Here's the data for the 20 cases (don't take this too seriously -- I made this data up to illustrate what a correlation is):PersonHeightSelf Esteem1684.12714.63623.84754.45583.26603.17673.88684.19714.310693.711683.512673.213633.714623.315603.416634.017654.118673.819633.420613.6Now, let's take a quick look at the histogram for each variable:And, here are the descriptive statistics:VariableMeanStDevVarianceSumMinimumMaximumRangeHeight65.44.4057419.41051308587517Self Esteem3.7550.4260900.18155375.13.14.61.5Finally, we'll look at the simple bivariate (i.e., two-variable) plot:You should immediately see in the bivariate plot that the relationship between the variables is a positive one (if you can't see that, review the section on types of relationships) because if you were to fit a single straight line through the dots it would have a positive slope or move up from left to right. Since the correlation is nothing more than a quantitative estimate of the relationship, we would expect a positive correlation.What does a \" positive relationship\" mean in this context? It means that, in general, higher scores on one variable tend to be paired with higher scores on the other and that lower scores on one variable tend to be paired with lower scores on the other. You should confirm visually that this is generally true in the plot above.Calculating the CorrelationNow we're ready to compute the correlation value. The formula for the correlation is:We use the symbol r to stand for the correlation. Through the magic of mathematics it turns out that r will always be between -1.0 and +1.0. if the correlation is negative, we have a negative relationship; if it's positive, the relationship is positive. You don't need to know how we came up with this formula unless you want to be a statistician. But you probably will need to know how the formula relates to real data -- how you can use the formula to compute the correlation. Let's look at the data we need for the formula. Here's the original data with the other necessary columns:PersonHeight (x)Self Esteem (y)x*yx*xy*y1684.1278.8462416.812714.6326.6504121.163623.8235.6384414.444754.4330562519.365583.2185.6336410.246603.118636009.617673.8254.6448914.448684.1278.8462416.819714.3305.3504118.4910693.7255.3476113.6911683.5238462412.2512673.2214.4448910.2413633.7233.1396913.6914623.3204.6384410.8915603.4204360011.561663425239691617654.1266.5422516.8118673.8254.6448914.4419633.4214.2396911.5620613.6219.6372112.96Sum =130875.14937.685912285.45The first three columns are the same as in the table above. The next three columns are simple computations based on the height and self esteem data. The bottom row consists of the sum of each column. This is all the information we need to compute the correlation. Here are the values from the bottom row of the table (where N is 20 people) as they are related to the symbols in the formula:Now, when we plug these values into the formula given above, we get the following (I show it here tediously, one step at a time):So, the correlation for our twenty cases is .73, which is a fairly strong positive relationship. I guess there is a relationship between height and self esteem, at least in this made up data!Testing the Significance of a CorrelationOnce you've computed a correlation, you can determine the probability that the observed correlation occurred by chance. That is, you can conduct a significance test. Most often you are interested in determining the probability that the correlation is a real one and not a chance occurrence. In this case, you are testing the mutually exclusive hypotheses:Null Hypothesis: r = 0Alternative Hypothesis: r 0The easiest way to test this hypothesis is to find a statistics book that has a table of critical values of r. Most introductory statistics texts would have a table like this. As in all hypothesis testing, you need to first determine the significance level. Here, I'll use the common significance level of alpha = .05. This means that I am conducting a test where the odds that the correlation is a chance occurrence is no more than 5 out of 100. Before I look up the critical value in a table I also have to compute the degrees of freedom or df. The df is simply equal to N-2 or, in this example, is 20-2 = 18. Finally, I have to decide whether I am doing a one-tailed or two-tailed test. In this example, since I have no strong prior theory to suggest whether the relationship between height and self esteem would be positive or negative, I'll opt for the two-tailed test. With these three pieces of information -- the significance level (alpha = .05)), degrees of freedom (df = 18), and type of test (two-tailed) -- I can now test the significance of the correlation I found. When I look up this value in the handy little table at the back of my statistics book I find that the critical value is .4438. This means that if my correlation is greater than .4438 or less than -.4438 (remember, this is a two-tailed test) I can conclude that the odds are less than 5 out of 100 that this is a chance occurrence. Since my correlation 0f .73 is actually quite a bit higher, I conclude that it is not a chance finding and that the correlation is \" statistically significant\" (given the parameters of the test). I can reject the null hypothesis and accept the alternative.The Correlation MatrixAll I've shown you so far is how to compute a correlation between two variables. In most studies we have considerably more than two variables. Let's say we have a study with 10 interval-level variables and we want to estimate the relationships among all of them (i.e., between all possible pairs of variables). In this instance, we have 45 unique correlations to estimate (more later on how I knew that!). We could do the above computations 45 times to obtain the correlations. Or we could use just about any statistics program to automatically compute all 45 with a simple click of the mouse.I used a simple statistics program to generate random data for 10 variables with 20 cases (i.e., persons) for each variable. Then, I told the program to compute the correlations among these variables. Here's the result: C1 C2 C3 C4 C5 C6 C7 C8 C9 C10C1 1.000C2 0.274 1.000C3 -0.134 -0.269 1.000C4 0.201 -0.153 0.075 1.000C5 -0.129 -0.166 0.278 -0.011 1.000C6 -0.095 0.280 -0.348 -0.378 -0.009 1.000C7 0.171 -0.122 0.288 0.086 0.193 0.002 1.000C8 0.219 0.242 -0.380 -0.227 -0.551 0.324 -0.082 1.000C9 0.518 0.238 0.002 0.082 -0.015 0.304 0.347 -0.013 1.000C10 0.299 0.568 0.165 -0.122 -0.106 -0.169 0.243 0.014 0.352 1.000This type of table is called a correlation matrix. It lists the variable names (C1-C10) down the first column and across the first row. The diagonal of a correlation matrix (i.e., the numbers that go from the upper left corner to the lower right) always consists of ones. That's because these are the correlations between each variable and itself (and a variable is always perfectly correlated with itself). This statistical program only shows the lower triangle of the correlation matrix. In every correlation matrix there are two triangles that are the values below and to the left of the diagonal (lower triangle) and above and to the right of the diagonal (upper triangle). There is no reason to print both triangles because the two triangles of a correlation matrix are always mirror images of each other (the correlation of variable x with variable y is always equal to the correlation of variable y with variable x). When a matrix has this mirror-image quality above and below the diagonal we refer to it as a symmetric matrix. A correlation matrix is always a symmetric matrix.To locate the correlation for any pair of variables, find the value in the table for the row and column intersection for those two variables. For instance, to find the correlation between variables C5 and C2, I look for where row C2 and column C5 is (in this case it's blank because it falls in the upper triangle area) and where row C5 and column C2 is and, in the second case, I find that the correlation is -.166.OK, so how did I know that there are 45 unique correlations when we have 10 variables? There's a handy simple little formula that tells how many pairs (e.g., correlations) there are for any number of variables:where N is the number of variables. In the example, I had 10 variables, so I know I have (10 * 9)/2 = 90/2 = 45 pairs.Other CorrelationsThe specific type of correlation I've illustrated here is known as the Pearson Product Moment Correlation. It is appropriate when both variables are measured at an interval level. However there are a wide variety of other types of correlations for other circumstances. for instance, if you have two ordinal variables, you could use the Spearman rank Order Correlation (rho) or the Kendall rank order Correlation (tau). When one measure is a continuous interval level one and the other is dichotomous (i.e., two-category) you can use the Point-Biserial Correlation. For other situations, consulting the web-based statistics selection program, Selecting Statistics at http://trochim.human.cornell.edu/selstat/ssstart.htm.