The self-normalized statistics in Chaps. 15 and 16 are Studentized statistics of the form \((\hat \theta - \theta )/\hat {se}\) which are generalizations of the t-statistic for \({\sqrt n} {(\bar X_n - \mu )}{/s_n}\) testing the null hypothesis that the mean of a normal distribution is μ, when the variance ō2 is unknown and estimated by the sample variance s2n. In Sect. 17.1 we consider another class of self-normalized statistics, called generalized likelihood ratio (GLR) statistics, which are extensions of likelihood ratio (LR) statistics (for testing simple hypotheses) to composite hypotheses in parametric models. Whereas LR statistics are martingales under the null hypothesis, GLR statistics are no longer martingales but can be analyzed by using LR martingales and the pseudo-maximization technique of Chap. 11. The probabilistic technique of pseudo-maximization via the method of mixtures has a fundamental statistical counterpart that links likelihood to Bayesian inference; this is treated in Sect. 17.2.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
(2009). Pseudo-Maximization in Likelihood and Bayesian Inference. In: Self-Normalized Processes. Probability and its Applications. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-85636-8_17
Download citation
DOI: https://doi.org/10.1007/978-3-540-85636-8_17
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-85635-1
Online ISBN: 978-3-540-85636-8
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)