Abstract
This chapter gives some concepts of correlation and regression analysis. Correlation comes prior to regression analysis. It starts with the concept of simple correlation coefficient; which gives the degree of linear relationship between two variables. One should draw scatter diagram in order to judge whether there exists any linear relation between the two variables. The correlation coefficient is not only invariant under changes of unit of measurements but also unaffected by changes of origin for both variables. The value of the correlation coefficient always lies in-between –1 and +1. As the scatter points move closer to the straight line, it moves to –1 or +1 depending on whether the relation is negative or positive. The straight-line relation between the two variables can be found by Least Squares (LS) method. The goodness of fit of the linear regression can be measured by the square of the simple correlation coefficient. Multiple Linear Regression Model is an extension of Simple Linear Regression Model. In a multiple linear regression model, we have more than two independent variables. The goodness of fit in this case is measured by coefficient of determination which is the square of the multiple correlation coefficient.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsNotes
- 1.
Center of Gravity is the \(\left( {\bar{x},\bar{y}} \right)\) point, where \(\bar{x}\) and \(\bar{y}\) are the arithmetic means of x values and y values, respectively.
- 2.
If y causes x, then we just interchange the two symbols (i.e., x is denoted as y and y is denoted as x) and y = a + bx and \(x = a^{{\prime }} + b^{{\prime }} y\) are equivalent.
- 3.
For example, proportion of expenditure on food and proportion of expenditure on non-food always have perfect linear relation, sum of these two variables being 1, leading to the value of the correlation coefficient to be perfectly as 1.
- 4.
If we are not specifically interested in the ith observation, we may simply write \(y = a + bx + e\).
- 5.
Strictly speaking, whether this transformation is valid depends on the nature of the error term. We have assumed that the errors are in a multiplicative form so that after transformation it is in the additive form.
- 6.
The assumption of normal distribution is not needed for unbiased estimation of the regression coefficients. All we need is that ei has mean 0 for all i.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Pal, M., Bharati, P. (2019). Introduction to Correlation and Linear Regression Analysis. In: Applications of Regression Techniques. Springer, Singapore. https://doi.org/10.1007/978-981-13-9314-3_1
Download citation
DOI: https://doi.org/10.1007/978-981-13-9314-3_1
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-9313-6
Online ISBN: 978-981-13-9314-3
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)