Abstract
Among the most important techniques in statistics is hypothesis testing. A hypothesis is a supposition about a certain state of affairs. It does not spring from a sudden epiphany or a long-standing conviction; rather, it offers a testable explanation of a specific phenomenon. A hypothesis is something that we can accept (verify) or reject (falsify) based on empirical data.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
In Excel 2010, this can be reached by clicking File → Options → Add-ins → Go.
- 2.
For a very good explanation of how to perform this test in SPSS, see https://www.youtube.com/watch?v=MJGk2sg4EZU on the how2stats YouTube channel.
- 3.
For a very good explanation of how to perform this test in Stata, see https://www.youtube.com/watch?v=ajzMeANAMzI on the Stata Learner YouTube channel.
- 4.
For a very good explanation of how to perform this test in Excel, see https://www.youtube.com/watch?v=wy8GVt7Ityk on the YouTube channel of https://alphabench.com
- 5.
Most statistical software packages perform this step automatically.
- 6.
For a very good explanation of how to do this in SPSS, see https://www.youtube.com/watch?v=dkobjvhxTro on the YouTube channel of Dr. Todd Grande.
- 7.
For a very good explanation of how to do this in Stata, see https://www.youtube.com/watch?v=2oJxerMCwIE and https://www.youtube.com/watch?v=NIwtaZqNFs8 on the Stata Learner YouTube channel.
- 8.
For a very good explanation of how to do this in Excel, see https://www.youtube.com/watch?v=xlgeta9FivI on the YouTube channel of Matthias Kullowatz and https://www.youtube.com/watch?v=mJtbhGETU88 on the YouTube channel of Dr. Todd Grande.
- 9.
Based on the results of the Kolmogorov–Smirnov test and the Shapiro–Wilk test (see Sect. 9.6.2), we must reject the hypothesis of a normal distribution.
- 10.
Based on the results of the Kolmogorov–Smirnov test and the Shapiro–Wilk test (see Sect. 9.6.2), we must reject the hypothesis of a normal distribution.
- 11.
In Excel 2010 this can be reached by clicking File → Options → Add-ins → Go.
- 12.
For a very good explanation of how to perform these steps using Excel, see https://www.youtube.com/watch?v=BlS11D2VL_U on the YouTube channel of Jim Grange or https://www.youtube.com/watch?v=X14z9r8FUKY on the YouTube channel of Dr. James Clark from the Kings College London (Essential Life Science Tutorials).
- 13.
- 14.
Based on the results of the Kolmogorov–Smirnov test and the Shapiro–Wilk test, we must reject the hypothesis of a normal distribution.
- 15.
See the file Chocopraline_colour_name_price.sav for SPSS and Chocopraline_colour_name_price.dta for Stata.
- 16.
The result of a single-factor univariate ANOVA for a factor with two traits is the same as that of a t-test for independent samples.
- 17.
Traditionally, the F-test (or its application to groups, Bartlett’s test) is used to measure equal variance. But these tests react very sensitively to deviations from the normal distribution, which is why the more robust Levene’s test is preferred. SPSS automatically performs Levene’s test for equal variance when performing ANOVA (choose Options → Homogeneity tests). Stata uses the one-way ANOVA to determine Bartlett’s test for equal variances. The Levene’s test calculation (w_0) is located under the heading Hypothesis Tests.
- 18.
Strictly speaking, all the measures of regression diagnostics (heteroskedasticity, autocorrelation, multicollinearity, etc.) should be performed when carrying out an ANCOVA (see Sect. 10.10).
- 19.
When using Excel 2010, select File → Options → Add-ins → GO instead.
- 20.
For a very good explanation of ANOVA in Excel, see https://www.youtube.com/watch?v=tPGPV_XPw-o on the YouTube channel of Dr. James Clark from the Kings College London (Essential Life Science Tutorials) and https://www.youtube.com/watch?v=JfUf5DR2Azs on the YouTube channel of StatisticsHowTo.com
- 21.
ti represents the respective number of rank scores for the value i. In our example, we have 4 rank scores of 2.5 for value 1, 3 rank scores of 6 for value 2, 8 rank scores of 11.5 for value 3, and 9 rank scores of 20 for value 4.
- 22.
The data in titanic.sav (SPSS), titanic.dta (Stata), and titanic.xls (Excel) contain figures on the number of persons on board and the number of victims. The data is taken from the British Board of Trade Inquiry Report (1990), Report on the Loss of the Titanic′ (S.S.), Gloucester (reprint).
- 23.
For a very good explanation of how to calculate the chi-square test of Independence using SPSS, see https://www.youtube.com/watch?v=wfIfEWMJY3s on the YouTube channel of ASK Brunel.
- 24.
Syntax command: tabulate class survived, cchi2 cell chi2 clrchi2 column expected row V.
- 25.
For a very good explanation of how to calculate the chi-square test of independence using Stata, see: https://www.youtube.com/watch?v=GZIi9zAlzIA on the StataCorp LLC YouTube channel.
- 26.
For a very good explanation of how to calculate the chi-square test of independence using Excel, see https://www.youtube.com/watch?v=ODxEoDyF6RI on the YouTube channel of Ken Blake.
- 27.
For a very good explanation of how to test for normal distribution using SPSS, see https://www.youtube.com/watch?v=dK-JNR3g_LU on the Dragonfly Statistics YouTube channel and https://www.youtube.com/watch?v=sQkB-AlJgPI on the HowToStats.com YouTube channel.
References
Backhaus, K., Erichson, B., Plinke, W., Weiber, R. (2016). Multivariate Analysemethoden. Eine anwendungsorientierte Einführung, 14th Edition. Berlin, Heidelberg: SpringerGabler.
Bortz, J. (1999). Statistik für Sozialwissenschaftler, 5th Edition. Berlin, Heidelberg: Springer.
Bortz, J., Lienert, G. A., Boehnke, K. (2000). Verteilungsfreie Methoden der Biostatistik, 2nd Edition. Berlin, Heidelberg: Springer.
Brown, M. B., Forsythe, A. B. (1974). Robust tests for the equality of variances. Journal of the American Statistical Association, 69, 364–367.
Conover, W.J. (1980). Practical nonparametric statistics. New-York: Wiley.
Dixon, W.J. (1954). Power under normality of several nonparametric tests. The Annals of Mathematical Statistics, 25: 610-614.
Field, A. (2005). Discovering Statistics Using SPSS. London: Sage.
Fisher, R.A., Yates, F. (1963). Statistical tables for biological, agricultural, and medical research. London: Oliver and Boyed.
Hair, J.F., Anderson, R.E., Tatham, R.L., Black, W.C. (1998). Multivariate Data Analysis, 5th Edition. London: Prentice Hall.
Kolmogorov, A. N. (1933). Sulla determinazione empirica di una legge di distribuzione. Giornale dell’ Istituto Italiano degli Attuari, 4, 83–91.
Kruskal, W. H., Wallis, W. A. (1952). Use of ranks in one-criterion variance analysis. Journal of the American Statistical Association, 47, 583–621.
Kruskal, W. H., Wallis, W. A. (1953). Errata: Use of ranks in one-criterion variance analysis. Journal of the American Statistical Association, 48, 907–911.
Mann, H.B., Whitney, D.R. (1947). On a test whether one of two random variables is stochastically larger than the other. The Annals of Mathematical Statistics, 18, 65-78.
Neyman, J. (1937). Outline of a theory of statistical estimation based on the classical theory of probability. Philosophical transactions of the Royal Society, Series A, 236.
Neyman, J. & Pearson, E. S. (1928a). On the use and interpretation of certain test criteria for purposes of statistical inference, part i. Biometrika, 20A, 175–240.
Neyman, J. & Pearson, E. S. (1928b). On the use and interpretation of certain test criteria for purposes of statistical inference, part ii. Biometrika, 20A, 263–294.
Popper, K. (1934). Logik der Forschung. Tübingen: Mohr.
Satterthwaite, F. E. (1946). An approximate distribution of estimates of variance components. Biometrics Bulletin 2: 110-114.
Scheffé, H. (1953). A method for judging all contrasts in the analysis of variance. Biometrika, 40, 87-104.
Shapiro, S. S., and R. S. Francia (1972). An approximate analysis of variance test for normality. Journal of the American Statistical Association, 67, 215–216.
Shapiro, S. S., and M. B. Wilk (1965). An analysis of variance test for normality (complete samples). Biometrika, 52, 591–611.
Smirnov, N. V. (1933). Estimate of deviation between empirical distribution functions in two independent samples. Bulletin Moscow University, 2, 3–16.
Spearman, C.E. (1904). The proof and measurement of association between two things. American Journal of Psychology, 15(1), 72–101.
Stevens, J. P. (1972). Four Methods of Analyzing between Variations for the k-Group MANOVA Problem, Multivariate Behaviorial Research, 7, 442-454.
Welch, B. L. (1947). The generalization of Student’s problem when several different population variances are involved. Biometrika, 34, 28–35.
Wilcoxon, F. (1945). Individual comparisons by ranking methods. Biometrics, 1, 80-83.
Wilcoxon, F. (1947). Probability tables for individual comparisons by ranking methods. Biometrics, 3, 119-122.
Witting, H. (1960). A generalized Pitman efficiency for nonparametric tests. The Annals of Mathematical Statistics, 31, 405-414.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Cleff, T. (2019). Hypothesis Testing. In: Applied Statistics and Multivariate Data Analysis for Business and Economics. Springer, Cham. https://doi.org/10.1007/978-3-030-17767-6_9
Download citation
DOI: https://doi.org/10.1007/978-3-030-17767-6_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-17766-9
Online ISBN: 978-3-030-17767-6
eBook Packages: Economics and FinanceEconomics and Finance (R0)