ebook img

Large Sample Techniques for Statistics (Springer Texts in Statistics) PDF

628 Pages·2012·3.37 MB·English
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview Large Sample Techniques for Statistics (Springer Texts in Statistics)

Springer Texts in Statistics Series Editors: G. Casella S. Fienberg I. Olkin For other titles published in this series, go to http://www.springer.com/series/417 Jiming Jiang Large Sample Techniques for Statistics 1 C Jiming Jiang University of California Department of Statistics 1 Shields Avenue Davis, California 95616 USA [email protected] STS Editorial Board George Casella Stephen Fienberg Department of Statistics Department of Statistics University of Florida Carnegie Mellon University Gainesville, FL 32611-8545 Pittsburg, PA 15213-3890 USA USA Ingram Olkin Department of Statistics Stanford University Stanford, CA 94305 USA ISSN 1431-875X ISBN 978-1-4419-6826-5 e-ISBN 978-1-4419-6827-2 DOI 10.1007/978-1-4419-6827-2 Springer New York Dordrecht Heidelberg London Library of Congress Control Number: 2010930134 © Springer Science+Business Media, LLC 2010 All rights reserved. This work may not be translated or copied in whole or in part without the written permission of the publisher (Springer Science+Business Media, LLC, 233 Spring Street, New York, NY 10013, USA), except for brief excerpts in connection with reviews or scholarly analysis. Use in connec- tion with any form of information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed is forbidden. The use in this publication of trade names, trademarks, service marks, and similar terms, even if they are not identified as such, is not to be taken as an expression of opinion as to whether or not they are subject to proprietary rights. Printed on acid-free paper Springer is part of Springer Science+Business Media (www.springer.com) For my parents, Huifen and Haoliang, and my sisters, Qiuming and Dongming, with love Preface In a way, the world is made up of approximations, and surely there is no exception in the world of statistics. In fact, approximations, especially large sample approximations,are very important parts of both theoretical and ap- pliedstatistics.TheGaussiandistribution,alsoknownasthenormaldistribu- tion,ismerelyonesuchexample,duetothewell-knowncentrallimittheorem. Large-sample techniques provide solutions to many practical problems; they simplify our solutions to difficult, sometimes intractable problems; they jus- tify our solutions; and they guide us to directions of improvements. On the other hand, just because large-sample approximations are used everywhere, and every day, it does not guarantee that they are used properly, and, when the techniques are misused, there may be serious consequences. Example 1 (Asymptotic χ2 distribution). Likelihood ratio test (LRT) is one of the fundamental techniques in statistics. It is well known that, in the “standard” situation, the asymptotic null distribution ofthe LRT isχ2, with thedegreesoffreedomequaltothedifferencebetweenthedimensions,defined as the numbers of free parameters, of the two nested models being compared (e.g., Rice 1995, pp. 310). This might lead to a wrong impression that the asymptotic (null) distribution of the LRT is always χ2. A similar mistake mighttakeplacewhendealing withPearson’sχ2-test—the asymptoticdistri- bution of Pearson’sχ2-test is not always χ2 (e.g., Moore 1978). Example2(Approximationtoamean).Itmightbethoughtthat,inalarge sample,onecould alwaysapproximatethe meanofa randomquantitybythe quantity itself. In some cases this technique works. For example, suppose X1,...,Xn are observationsthat are independent and identica(cid:2)lly distributed (i.i.d.)suchthatμ=E(X1)(cid:2)=0.Thenonecanapprox(cid:2)imateE( ni=1Xi)=nμ bysimplyremo(cid:2)vingtheexpectat(cid:2)ionsign,thatis,by ni=1Xi.Thi√sisbecause the difference ni=1Xi −nμ = ni=1(Xi −(cid:2)μ) is of the order O( n), which is lower than the order of the (cid:2)mean of ni=1Xi. However, this technique completelyfailsifoneconsiders( ni=1Xi)(cid:2)2 instead.Toseethis,letusassume for sim(cid:2)plicity that Xi ∼N((cid:2)0,1). Then E( ni=1√Xi)2(cid:2)=n. On the other hand, since n X ∼ N(0,n), ( n X )2 = n{(1/ n) n X }2 ∼ nχ2, where i=1 i i=1 i i=1 i 1 VIII Preface χ21 is a rand(cid:2)om variable wi(cid:2)th a χ2 distribution with one degree of freedom. There(cid:2)fore,( ni=1Xi)2−(cid:2)E( ni=1Xi)2 =n(χ21−1),whichisofthesameorder of E( n X )2. Thus, ( n X )2 is not a good approximation to its mean. i=1 i i=1 i Example 3 (Maximum likelihood estimation). Here is another example of theso-calledlarge-sampleparadox.Becauseofthepopularityofthemaximum likelihood and its well-known large-sample theory in the classical situation, onemightexpectthatthemaximumlikelihoodestimatorisalwaysconsistent. However, this is not true in some fairly simple, and practical, situations. For example,NeymanandScott(1948)gavethefollowingexample.Supposethat two measurements are taken from each of the n patients. Let y denote the ij jthmeasurementfromtheithpatient,i=1,...,n,j =1,2.Supposethatthe measurementsareindependent andy isnormallydistributed with unknown ij meanμ andvarianceσ2.Then,asn→∞,themaximumlikelihoodestimator i of σ2 is inconsistent. The above are only a few examples out of many, but the message is just as clear: It is time to unravel such confusions. This book deals with large-sample techniques in statistics. More impor- tantly, we show how to argue with large-sample techniques and how to use thesetechniquestherightway.Itshouldbepointedoutthatthereisanexten- sive literature on large-sampletheory, including books and published papers, some of which are highly mathematical. Traditionally, there have been sev- eralapproachestointroducingthesematerials.Thefirstisthetheorem/proof approach, which provides rigorous proofs for all or most of the theoretical results (e.g., Petrov 1975). The second is the method/application approachi, which focuses on using the results without paying attention to any of the proofs(e.g., Barndorff-Nielsenand Cox1989).Our approachis somewherein between. Instead of giving a formal, technical proof for every result, we focus on the ideas of asymptotic argumentsand howto use the methods developed by these arguments in various less-than-textbook situations. Webegin byreviewingsomeof theverysimple andfundamentalconcepts that most of us have learned, say, from a calculus book. More specifically, Chapters 1–5 are devoted to a comprehensive review of the basic tools for large-sample approximations, such as the (cid:4)-δ arguments, Taylor expansion, different types of convergence, and inequalities. Chapters 6–10 discuss limit theoremsinspecificsituationsofobservationaldata.Theseincludetheclassi- cal case of i.i.d. observations, independent but not identically distributed ob- servationssuchasthoseencounteredin linear regression,empiricalprocesses, martingales, time series, stochastic processes, and random fields. Each of the first 10 chapters contains at least one section of case study as applications of the methods or techniques covered in the chapter. Some more extensive applications of the large-sample techniques are discussed in Chapters 11–15. Theareasofapplicationsincludenonparametricstatistics,linearandgeneral- izedlinearmixedmodels,small-areaestimation,jackknifeandbootstrap,and Markov-chainMonte Carlo methods. Preface IX Asmentioned,therehavebeenseveralmajortextsonsimilartopics.These include, in the order of year published: [1] Hall & Heyde (1980), Martingale Limit Theory and Its Application, Academic Press; [2] Barndorff-Nielsen & Cox (1989), Asymptotic Techniques for Use in Statistics, Chapman & Hall; [3] Ferguson(1996),A Course in Large Sample Theory, Chapman & Hall; [4] Lehmann (1999), Elements of Large-Sample Theory, Springer; and [5] Das- Gupta (2008), Asymptotic Theory of Statistics and Probability, Springer. A comparison with these existing texts would help to highlight some of the features of the current book. Text [2] deals with the case of independent ob- servations.Inpractice,however,theobservationsareoftencorrelated.Amain purposeof the currentbookis to introduce large-sampletheory andmethods for correlated observations, such as those in time series, mixed models, and spatial statistics. Furthermore, the approach of [2] is more like “use this for- mula,” rather than “why?” and “what’s the trick?.” In contrast, the current text focuses more on the way of thinking. For example, the current text cov- ers basic elements in asymptotic theory, such as (cid:4)-δ, O , and o , in addition P P to the asymptotic results, such as a formula of asymptotic expansion. This reflects the current author’s belief that methodology is more important and applicable to a broader range of problems than formulas. Text [3] provides an account of large-sample theory for independent ran- dom variables (mostly in the i.i.d. case) with applications to efficient esti- mation and testing problems. Several classical cases of dependent random variablesarealsoconsidered,suchasm-dependentsequences,rank,andorder statistics, but the basic method was to convert these to the case of indepen- dent observations plus some extra terms that are asymptotically negligible. The chapters are written in a theorem–proof style which is what the author intended to do. Like [2] and [3], text [4] deals with independent observations, mostly the i.i.d.case.However,theapproachof[4]hasmotivatedthecurrentauthor.For example,[4]beginswithverysimpleandfundamentalconceptsandeventually getstoamuchadvancedlevel.Itmightbeworthmentioningthatthecurrent author assisted Professor E. L. Lehmann in the mid-1990sduring his writing of book [4]. Text [5] provides a very comprehensive account of asymptotic theory in statistics and probability. However, similar to books [2]–[4], the focus of [5] is mainly on independent observations. Also, since a large number of topics need to be covered, it is unavoidablethat the coverageis a little sketchy. Unlike books [2]–[5], text [1] deals with one special case of dependent observations—the martingales. Whereas the martingale limit theory applies to a broad ranges of problems, such as linear mixed models and some cases of time series, it does not cover many other cases encountered in practice. Furthermore, the book starts at a relatively high level, assuming that the readerhastakenanadvancedcourseinprobabilitytheory.Asmentioned,the current book begins with very basic concepts in asymptotic arguments, such

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.