Karl Pearson's Coefficient of Correlation | Assumptions, Merits and Demerits
Last Updated :
23 Jul, 2025
What is Karl Pearson's Coefficient of Correlation?
The first person to give a mathematical formula for the measurement of the degree of relationship between two variables in 1890 was Karl Pearson. Karl Pearson's Coefficient of Correlation is also known as Product Moment Correlation or Simple Correlation Coefficient. This method of measuring the coefficient of correlation is the most popular and is widely used. It is denoted by 'r', where r is a pure number which means that r has no unit.
According to Karl Pearson, "Coefficient of Correlation is calculated by dividing the sum of products of deviations from their respective means by their number of pairs and their standard deviations."
Karl~Pearson's~Coefficient~of~Correlation(r)=\frac{Sum~of~Products~of~Deviations~from~their~respective~means}{Number~of~Pairs\times{Standard~Deviations~of~both~Series}}
Or
r=\frac{\sum{xy}}{N\times{\sigma_x}\times{\sigma_y}}
Where,
N = Number of Pair of Observations
x = Deviation of X series from Mean (X-\bar{X})
y = Deviation of Y series from Mean (Y-\bar{Y})
\sigma_x
= Standard Deviation of X series (\sqrt{\frac{\sum{x^2}}{N}})
\sigma_y
= Standard Deviation of Y series (\sqrt{\frac{\sum{y^2}}{N}})
r = Coefficient of Correlation
Karl Pearson's Coefficient of Correlation method can be used only when there is availability of quantitative measurements of different items of a series. However, there are various cases in which the direct measurement of the phenomenon under study is not possible. For example, different qualitative measures such as ability, kindness, honesty, beauty, etc., cannot be measured in quantitative terms. To study the correlation between two qualitative measures, one should use Spearman's Rank Correlation.
Karl Pearson's Coefficient of Correlation and Covariance
Karl Pearson's method of determining coefficient of correlation is based on the covariance of the given variables. Covariance is a statistical representation of the degree to which the two given variables vary together. Basically, Covariance is a number reflecting the degree to which the two variables vary together. The symbol of Covariance of two variables (say X and Y) is denoted by COV(X, Y).
COV(X,~Y)=\frac{\sum{(X-\bar{X})(Y-\bar{Y})}}{N}=\frac{\sum{xy}}{N}
The formula for calculating Karl Pearson's Coefficient of Correlation can be transformed into another easy formula as:
r=\frac{\sum{xy}}{N\times{\sigma_x}\times{\sigma_y}}
Or, r=\frac{\sum{xy}}{N}\times{\frac{1}{\sigma_x}}\times{\frac{1}{\sigma_y}}
Or, r=\frac{\sum{xy}}{N\times{\sqrt{\frac{\sum{x^2}}{N}}}\times{\sqrt{\frac{\sum{y^2}}{N}}}}
Or, r=\frac{\sum{xy}}{\sqrt{\sum{x^2}\times{\sum{y^2}}}}
Note: This method of determining Coefficient of Correlation should be applied only when the deviations of items are taken from actual means and not from assumed means.
Also Read:
Karl Pearson’s Coefficient of Correlation | Methods and Examples
Examples of Karl Pearson's Coefficient of Correlation
Example 1:
Determine the Coefficient of Correlation between X and Y.

The summation of the product of deviations of Series X and Y from their respective means is 200.
Solution:
The figures given are:
N = 30, σx = 4, σy = 3, and ∑xy = 200
r=\frac{\sum{xy}}{N\times{\sigma_x}\times{\sigma_y}}
=\frac{200}{30\times4\times3}=\frac{50}{90}=0.5
Coefficient of Correlation = 0.5
It means that there is a positive correlation between X and Y.
Example 2:
If the Covariance between two variables X and Y is 9.4 and the variance of Series X and Y are 10.6 and 12.5, respectively, then calculate the coefficient of correlation.
Solution:
Covariance between X and Y = \frac{\sum{xy}}{N}=9.4
Variance of X = σx2 = 10.6
\sigma_x=\sqrt{10.6}=3.25
Variance of Y = σy2 = 12.5
\sigma_y=\sqrt{12.5}=3.53
r=\frac{\sum{xy}}{N\times{\sigma_x}\times{\sigma_y}}
r=\frac{\sum{xy}}{N}\times{\frac{1}{\sigma_x}}\times{\frac{1}{\sigma_y}}
=9.4\times{\frac{1}{3.25}}\times{\frac{1}{3.53}}
r = 9.4 x 0.307 x 0.282 = 0.816
Coefficient of Correlation = 0.816
It means that there is quite a high degree of positive correlation between X and Y.
Features of Karl Pearson's Coefficient of Correlation
The main features of Karl Pearson's Coefficient of Correlation are as follows:
1. Knowledge of Direction of Correlation: This method of measuring coefficient of correlation gives us knowledge about the direction of the relationship between two variables. In other words, it tells us whether the relationship between two variables is positive or negative.
2. Size of Correlation: Karl Pearson's Coefficient of Correlation indicates the size of the relationship between two variables. Besides, Correlation Coefficient ranges between -1 and +1.
3. Indicates Magnitude and Direction: This method not only specifies the magnitude of the correlation between two variables but also specifies its direction. It means that, if two variables are directly related, then the correlation coefficient between the variables will be a positive value. However, if two variables are inversely related, then the correlation coefficient between the variables will be a negative value.
4. Ideal Measure: As this method is based on the most essential statistical measure, such as standard deviation and mean, it is an ideal/appropriate measure.
Note: The value of the Correlation Coefficient should always lie between -1 and +1.
- When r = +1, it means that there is perfect positive correlation.
- When r = -1, it means that there is perfect negative correlation.
- When r = 0, it means that there is no or zero correlation.
Assumptions of Coefficient of Correlation
The assumptions on which Karl Pearson's Coefficient of Correlation is based are as follows:
1. Linear Relationship: The first assumption in this method is that there is a linear relationship between the given two variables. It means that if the paired observations of the variables (say X and Y) are plotted on a scatter diagram, then it will always form a straight line.
2. Casual Relationship: Another assumption is that there is no cause-and-effect relationship between the given two variables. However, the cause-and-effect relationship exists between the forces affecting these variables. Besides, if there is no such type of relationship between the variables, then the correlation is meaningless.
3. Normal Distribution: It is also assumed that, if there are a large number of independent causes of some nature producing normal distribution, then the two given variables are affected by them. For example, Variables like demand, supply, height, weight, etc., are affected by multiple forces.
4. Error of Measurement: If the error of measurement is reduced to the minimum, then the coefficient of correlation is more reliable.
Properties of Coefficient of Correlation
1. Coefficient of Correlation is Independent of change of origin and scale of measurements: Coefficient of Correlation is not affected by the change of origin and scale of measurement.
2. Coefficient of Correlation lies between -1 and +1: The property of r also serves as a useful check on the correctness of the calculations. If the value of r lies outside the range, then it would mean that there is some error in the calculations.
3. Zero Correlation: If two variables (say X and Y) are independent of each other, in that case, the coefficient of correlation between them will be zero.
4. Measure of Linear Relationship: The coefficient of correlation is a measure that helps in determining the linear relationship between two variables. If both the variables (say X and Y) increase or decrease together, then r will be positive. However, if one variable increases when the other variable decreases or vice-versa, then r will be negative.
Merits of Karl Pearson's Coefficient of Correlation
Various advantages of Karl Pearson's Coefficient of Correlation are as follows:
1. Popular Method: Karl Pearson's Coefficient of Correlation is the most popular and widely used mathematical method to study the correlation between two variables.
2. Degree and Direction of Correlation: The value of correlation coefficient not only summarises the degree of correlation but also its direction.
Demerits of Karl Pearson's Coefficient of Correlation
Various disadvantages of Karl Pearson's Coefficient of Correlation are as follows:
1. Affected by Extreme Values: If the values of the two variables are extreme, then it would have a great impact on the value of correlation coefficient.
2. Assumption of Linear Relationship: While determining correlation coefficient, it is always assumed that there is a linear relationship without thinking whether the assumption is correct or not.
3. Time-Consuming Method: In comparison to other methods of determining correlation coefficients, this method takes more time.
4. Possibility of Wrong Interpretation: While interpreting the value of coefficient of correlation using this method, one has to be very careful. It is because the chances of misinterpreting the coefficient are more.
Also Read:
Correlation: Meaning, Significance, Types and Degree of Correlation
Methods of measurements of Correlation
Calculation of Correlation with Scattered Diagram
Spearman’s Rank Correlation Coefficient in Statistics
Similar Reads
Data Science Tutorial Data Science is a field that combines statistics, machine learning and data visualization to extract meaningful insights from vast amounts of raw data and make informed decisions, helping businesses and industries to optimize their operations and predict future trends.This Data Science tutorial offe
3 min read
Introduction to Machine Learning
What is Data Science?Data science is the study of data that helps us derive useful insight for business decision making. Data Science is all about using tools, techniques, and creativity to uncover insights hidden within data. It combines math, computer science, and domain expertise to tackle real-world challenges in a
8 min read
Top 25 Python Libraries for Data Science in 2025Data Science continues to evolve with new challenges and innovations. In 2025, the role of Python has only grown stronger as it powers data science workflows. It will remain the dominant programming language in the field of data science. Its extensive ecosystem of libraries makes data manipulation,
10 min read
Difference between Structured, Semi-structured and Unstructured dataBig Data includes huge volume, high velocity, and extensible variety of data. There are 3 types: Structured data, Semi-structured data, and Unstructured data. Structured data - Structured data is data whose elements are addressable for effective analysis. It has been organized into a formatted repos
2 min read
Types of Machine LearningMachine learning is the branch of Artificial Intelligence that focuses on developing models and algorithms that let computers learn from data and improve from previous experience without being explicitly programmed for every task.In simple words, ML teaches the systems to think and understand like h
13 min read
What's Data Science Pipeline?Data Science is a field that focuses on extracting knowledge from data sets that are huge in amount. It includes preparing data, doing analysis and presenting findings to make informed decisions in an organization. A pipeline in data science is a set of actions which changes the raw data from variou
3 min read
Applications of Data ScienceData Science is the deep study of a large quantity of data, which involves extracting some meaning from the raw, structured, and unstructured data. Extracting meaningful data from large amounts usesalgorithms processing of data and this processing can be done using statistical techniques and algorit
6 min read
Python for Machine Learning
Learn Data Science Tutorial With PythonData Science has become one of the fastest-growing fields in recent years, helping organizations to make informed decisions, solve problems and understand human behavior. As the volume of data grows so does the demand for skilled data scientists. The most common languages used for data science are P
3 min read
Pandas TutorialPandas (stands for Python Data Analysis) is an open-source software library designed for data manipulation and analysis. Revolves around two primary Data structures: Series (1D) and DataFrame (2D)Built on top of NumPy, efficiently manages large datasets, offering tools for data cleaning, transformat
6 min read
NumPy Tutorial - Python LibraryNumPy is a core Python library for numerical computing, built for handling large arrays and matrices efficiently.ndarray object â Stores homogeneous data in n-dimensional arrays for fast processing.Vectorized operations â Perform element-wise calculations without explicit loops.Broadcasting â Apply
3 min read
Scikit Learn TutorialScikit-learn (also known as sklearn) is a widely-used open-source Python library for machine learning. It builds on other scientific libraries like NumPy, SciPy and Matplotlib to provide efficient tools for predictive data analysis and data mining.It offers a consistent and simple interface for a ra
3 min read
ML | Data Preprocessing in PythonData preprocessing is a important step in the data science transforming raw data into a clean structured format for analysis. It involves tasks like handling missing values, normalizing data and encoding variables. Mastering preprocessing in Python ensures reliable insights for accurate predictions
6 min read
EDA - Exploratory Data Analysis in PythonExploratory Data Analysis (EDA) is a important step in data analysis which focuses on understanding patterns, trends and relationships through statistical tools and visualizations. Python offers various libraries like pandas, numPy, matplotlib, seaborn and plotly which enables effective exploration
6 min read
Introduction to Statistics
Statistics For Data ScienceStatistics is like a toolkit we use to understand and make sense of information. It helps us collect, organize, analyze and interpret data to find patterns, trends and relationships in the world around us.From analyzing scientific experiments to making informed business decisions, statistics plays a
12 min read
Descriptive StatisticStatistics is the foundation of data science. Descriptive statistics are simple tools that help us understand and summarize data. They show the basic features of a dataset, like the average, highest and lowest values and how spread out the numbers are. It's the first step in making sense of informat
5 min read
What is Inferential Statistics?Inferential statistics is an important tool that allows us to make predictions and conclusions about a population based on sample data. Unlike descriptive statistics, which only summarize data, inferential statistics let us test hypotheses, make estimates, and measure the uncertainty about our predi
7 min read
Bayes' TheoremBayes' Theorem is a mathematical formula used to determine the conditional probability of an event based on prior knowledge and new evidence. It adjusts probabilities when new information comes in and helps make better decisions in uncertain situations.Bayes' Theorem helps us update probabilities ba
13 min read
Probability Data Distributions in Data ScienceUnderstanding how data behaves is one of the first steps in data science. Before we dive into building models or running analysis, we need to understand how the values in our dataset are spread out and thatâs where probability distributions come in.Let us start with a simple example: If you roll a f
8 min read
Parametric Methods in StatisticsParametric statistical methods are those that make assumptions regarding the distribution of the population. These methods presume that the data have a known distribution (e.g., normal, binomial, Poisson) and rely on parameters (e.g., mean and variance) to define the data.Key AssumptionsParametric t
6 min read
Non-Parametric TestsNon-parametric tests are applied in hypothesis testing when the data does not satisfy the assumptions necessary for parametric tests, such as normality or equal variances. These tests are especially helpful for analyzing ordinal data, small sample sizes, or data with outliers.Common Non-Parametric T
5 min read
Hypothesis TestingHypothesis testing compares two opposite ideas about a group of people or things and uses data from a small part of that group (a sample) to decide which idea is more likely true. We collect and study the sample data to check if the claim is correct.Hypothesis TestingFor example, if a company says i
9 min read
ANOVA for Data Science and Data AnalyticsANOVA is useful when we need to compare more than two groups and determine whether their means are significantly different. Suppose you're trying to understand which ingredients in a recipe affect its taste. Some ingredients, like spices might have a strong influence while others like a pinch of sal
9 min read
Bayesian Statistics & ProbabilityBayesian statistics sees unknown values as things that can change and updates what we believe about them whenever we get new information. It uses Bayesâ Theorem to combine what we already know with new data to get better estimates. In simple words, it means changing our initial guesses based on the
6 min read
Feature Engineering
Model Evaluation and Tuning
Data Science Practice