Empirical Likelihood Inference for Two-sample Problems

Empirical Likelihood Inference for Two-sample Problems
Author: Ying Yan
Publisher:
Total Pages: 40
Release: 2010
Genre:
ISBN:

Download Empirical Likelihood Inference for Two-sample Problems Book in PDF, Epub and Kindle

In this thesis, we are interested in empirical likelihood (EL) methods for two-sample problems, with focus on the difference of the two population means. A weighted empirical likelihood method (WEL) for two-sample problems is developed. We also consider a scenario where sample data on auxiliary variables are fully observed for both samples but values of the response variable are subject to missingness. We develop an adjusted empirical likelihood method for inference of the difference of the two population means for this scenario where missing values are handled by a regression imputation method. Bootstrap calibration for WEL is also developed. Simulation studies are conducted to evaluate the performance of naive EL, WEL and WEL with bootstrap calibration (BWEL) with comparison to the usual two-sample t-test in terms of power of the tests and coverage accuracies. Simulation for the adjusted EL for the linear regression model with missing data is also conducted.

Empirical Likelihood

Empirical Likelihood
Author: Art B. Owen
Publisher: CRC Press
Total Pages: 322
Release: 2001-05-18
Genre: Mathematics
ISBN: 1420036157

Download Empirical Likelihood Book in PDF, Epub and Kindle

Empirical likelihood provides inferences whose validity does not depend on specifying a parametric model for the data. Because it uses a likelihood, the method has certain inherent advantages over resampling methods: it uses the data to determine the shape of the confidence regions, and it makes it easy to combined data from multiple sources. It al

Empirical Likelihood and Bootstrap Inference with Constraints

Empirical Likelihood and Bootstrap Inference with Constraints
Author: Chunlin Wang
Publisher:
Total Pages: 172
Release: 2017
Genre:
ISBN:

Download Empirical Likelihood and Bootstrap Inference with Constraints Book in PDF, Epub and Kindle

Empirical likelihood and the bootstrap play influential roles in contemporary statistics. This thesis studies two distinct statistical inference problems, referred to as Part I and Part II, related to the empirical likelihood and bootstrap, respectively. Part I of this thesis concerns making statistical inferences on multiple groups of samples that contain excess zero observations. A unique feature of the target populations is that the distribution of each group is characterized by a non-standard mixture of a singular distribution at zero and a skewed nonnegative component. In Part I of this thesis, we propose modelling the nonnegative components using a semiparametric, multiple-sample, density ratio model (DRM). Under this semiparametric setup, we can efficiently utilize information from the combined samples even with unspecified underlying distributions. We first study the question of testing homogeneity of multiple nonnegative distributions when there is an excess of zeros in the data, under the proposed semiparametric setup. We develop a new empirical likelihood ratio (ELR) test for homogeneity and show that this ELR has a $\chi^2$-type limiting distribution under the homogeneous null hypothesis. A nonparametric bootstrap procedure is proposed to calibrate the finite-sample distribution of the ELR. The consistency of this bootstrap procedure is established under both the null and alternative hypotheses. Simulation studies show that the bootstrap ELR test has an accurate nominal type I error, is robust to changes of underlying distributions, is competitive to, and sometimes more powerful than, several popular one- and two-part tests. A real data example is used to illustrate the advantages of the proposed test. We next investigate the problem of comparing the means of multiple nonnegative distributions, with excess zero observations, under the proposed semiparametric setup. We develop a unified inference framework based on our new ELR statistic, and show that this ELR has a $\chi^2$-type limiting distribution under a general null hypothesis. This allows us to construct a new test for mean equality. Simulation results show favourable performance of the proposed ELR test compared with other existing tests for mean equality, especially when the correctly specified basis function in the DRM is the logarithm function. A real data set is analyzed to illustrate the advantages of the proposed method. In Part II of this thesis, we investigate the asymptotic behaviour of, the commonly used, bootstrap percentile confidence intervals when the parameters are subject to inequality constraints. We concentrate on the important one- and two-sample problems with data generated from distributions in the natural exponential family. Our attention is focused on quantifying asymptotic coverage probabilities of the percentile confidence intervals based on bootstrapping maximum likelihood estimators. We propose a novel local framework to study the subtle asymptotic behaviour of bootstrap percentile confidence intervals when the true parameter values are close to the boundary. Under this framework, we discover that when the true parameter is on, or close to, the restriction boundary, the local asymptotic coverage probabilities can always exceed the nominal level in the one-sample case; however, they can be, surprisingly, both under and over the nominal level in the two-sample case. The results provide theoretical justification and guidance on applying the bootstrap percentile method to constrained inference problems. The two individual parts of this thesis are connected by being referred to as {\em constrained statistical inference}. Specifically, in Part I, the semiparametric density ratio model uses an exponential tilting constraint, which is a type of equality constraint, on the parameter space. In Part II, we deal with inequality constraints, such as a boundary or ordering constraints, on the parameter space. For both parts, an important regularity condition in traditional likelihood inference, that parameters should be interior points of the parameter space, is violated. Therefore, the respective inference procedures involve non-standard asymptotics that create new technical challenges.

Sample Surveys: Inference and Analysis

Sample Surveys: Inference and Analysis
Author:
Publisher: Morgan Kaufmann
Total Pages: 667
Release: 2009-09-02
Genre: Mathematics
ISBN: 0080963544

Download Sample Surveys: Inference and Analysis Book in PDF, Epub and Kindle

Handbook of Statistics_29B contains the most comprehensive account of sample surveys theory and practice to date. It is a second volume on sample surveys, with the goal of updating and extending the sampling volume published as volume 6 of the Handbook of Statistics in 1988. The present handbook is divided into two volumes (29A and 29B), with a total of 41 chapters, covering current developments in almost every aspect of sample surveys, with references to important contributions and available software. It can serve as a self contained guide to researchers and practitioners, with appropriate balance between theory and real life applications. Each of the two volumes is divided into three parts, with each part preceded by an introduction, summarizing the main developments in the areas covered in that part. Volume 1 deals with methods of sample selection and data processing, with the later including editing and imputation, handling of outliers and measurement errors, and methods of disclosure control. The volume contains also a large variety of applications in specialized areas such as household and business surveys, marketing research, opinion polls and censuses. Volume 2 is concerned with inference, distinguishing between design-based and model-based methods and focusing on specific problems such as small area estimation, analysis of longitudinal data, categorical data analysis and inference on distribution functions. The volume contains also chapters dealing with case-control studies, asymptotic properties of estimators and decision theoretic aspects. Comprehensive account of recent developments in sample survey theory and practice Covers a wide variety of diverse applications Comprehensive bibliography

EMPIRICAL LIKELIHOOD TESTS FOR CONSTANT VARIANCE IN THE TWO-SAMPLE PROBLEM

EMPIRICAL LIKELIHOOD TESTS FOR CONSTANT VARIANCE IN THE TWO-SAMPLE PROBLEM
Author: Paul Shen
Publisher:
Total Pages: 19
Release: 2019
Genre: Mathematical statistics
ISBN:

Download EMPIRICAL LIKELIHOOD TESTS FOR CONSTANT VARIANCE IN THE TWO-SAMPLE PROBLEM Book in PDF, Epub and Kindle

In this thesis, we investigate the problem of testing constant variance. It is an important problem in the field of statistical influence where many methods require the assumption of constant variance. The question of constant variance has to be settled in order to perform a significance test through a Student t-Test or an F-test. Two of most popular tests of constant variance in applications are the classic F-test and the Modified Levene's Test. The former is a ratio of two sample variances. Its performance is found to be very sensitive with the normality assumption. The latter Modified Levene's Test can be viewed as a result of the estimation method through the absolute deviation from the median. Its performance is also dependent upon the distribution shapes to some extent, though not as much as the F-test. We propose an innovative test constructed by the empirical likelihood method through the moment estimation equations appearing in the Modified Levene's Test. The new empirical likelihood ratio test is a nonparametric test and retains the principle of maximum likelihood. As a result, it can be an appropriate alternative to the two traditional tests in applications when underlying populations are skewed. To be specific, the empirical likelihood ratio test of constant variance uses the optimal weights in summing the absolute deviations of observations from the median values, while the Modified Levene's test uses the simple averages. It is thus desired that the empirical likelihood ratio test is more powerful than the Modified Levene's test. Meanwhile, the empirical likelihood ratio test is expected to be as robust as the Modified Levene's test, as the empirical likelihood ratio test is also constructed via the same distance as the Modified Levene's test. A real-life data set is used to illustrate implementation of the empirical likelihood ratio test with comparisons to the classic F-test and the Modified Levene's Test. It is confirmed that the empirical likelihood ratio test performs the best.

Empirical Likelihood Methods in Biomedicine and Health

Empirical Likelihood Methods in Biomedicine and Health
Author: Albert Vexler
Publisher: CRC Press
Total Pages: 300
Release: 2018-09-03
Genre: Mathematics
ISBN: 1351001515

Download Empirical Likelihood Methods in Biomedicine and Health Book in PDF, Epub and Kindle

Empirical Likelihood Methods in Biomedicine and Health provides a compendium of nonparametric likelihood statistical techniques in the perspective of health research applications. It includes detailed descriptions of the theoretical underpinnings of recently developed empirical likelihood-based methods. The emphasis throughout is on the application of the methods to the health sciences, with worked examples using real data. Provides a systematic overview of novel empirical likelihood techniques. Presents a good balance of theory, methods, and applications. Features detailed worked examples to illustrate the application of the methods. Includes R code for implementation. The book material is attractive and easily understandable to scientists who are new to the research area and may attract statisticians interested in learning more about advanced nonparametric topics including various modern empirical likelihood methods. The book can be used by graduate students majoring in biostatistics, or in a related field, particularly for those who are interested in nonparametric methods with direct applications in Biomedicine.

In All Likelihood

In All Likelihood
Author: Yudi Pawitan
Publisher: OUP Oxford
Total Pages: 543
Release: 2013-01-17
Genre: Mathematics
ISBN: 0191650579

Download In All Likelihood Book in PDF, Epub and Kindle

Based on a course in the theory of statistics this text concentrates on what can be achieved using the likelihood/Fisherian method of taking account of uncertainty when studying a statistical problem. It takes the concept ot the likelihood as providing the best methods for unifying the demands of statistical modelling and the theory of inference. Every likelihood concept is illustrated by realistic examples, which are not compromised by computational problems. Examples range from a simile comparison of two accident rates, to complex studies that require generalised linear or semiparametric modelling. The emphasis is that the likelihood is not simply a device to produce an estimate, but an important tool for modelling. The book generally takes an informal approach, where most important results are established using heuristic arguments and motivated with realistic examples. With the currently available computing power, examples are not contrived to allow a closed analytical solution, and the book can concentrate on the statistical aspects of the data modelling. In addition to classical likelihood theory, the book covers many modern topics such as generalized linear models and mixed models, non parametric smoothing, robustness, the EM algorithm and empirical likelihood.

Sampling Theory and Practice

Sampling Theory and Practice
Author: Changbao Wu
Publisher: Springer Nature
Total Pages: 371
Release: 2020-05-15
Genre: Social Science
ISBN: 3030442462

Download Sampling Theory and Practice Book in PDF, Epub and Kindle

The three parts of this book on survey methodology combine an introduction to basic sampling theory, engaging presentation of topics that reflect current research trends, and informed discussion of the problems commonly encountered in survey practice. These related aspects of survey methodology rarely appear together under a single connected roof, making this book a unique combination of materials for teaching, research and practice in survey sampling. Basic knowledge of probability theory and statistical inference is assumed, but no prior exposure to survey sampling is required. The first part focuses on the design-based approach to finite population sampling. It contains a rigorous coverage of basic sampling designs, related estimation theory, model-based prediction approach, and model-assisted estimation methods. The second part stems from original research conducted by the authors as well as important methodological advances in the field during the past three decades. Topics include calibration weighting methods, regression analysis and survey weighted estimating equation (EE) theory, longitudinal surveys and generalized estimating equations (GEE) analysis, variance estimation and resampling techniques, empirical likelihood methods for complex surveys, handling missing data and non-response, and Bayesian inference for survey data. The third part provides guidance and tools on practical aspects of large-scale surveys, such as training and quality control, frame construction, choices of survey designs, strategies for reducing non-response, and weight calculation. These procedures are illustrated through real-world surveys. Several specialized topics are also discussed in detail, including household surveys, telephone and web surveys, natural resource inventory surveys, adaptive and network surveys, dual-frame and multiple frame surveys, and analysis of non-probability survey samples. This book is a self-contained introduction to survey sampling that provides a strong theoretical base with coverage of current research trends and pragmatic guidance and tools for conducting surveys.

Statistics for High-Dimensional Data

Statistics for High-Dimensional Data
Author: Peter Bühlmann
Publisher: Springer Science & Business Media
Total Pages: 568
Release: 2011-06-08
Genre: Mathematics
ISBN: 364220192X

Download Statistics for High-Dimensional Data Book in PDF, Epub and Kindle

Modern statistics deals with large and complex data sets, and consequently with models containing a large number of parameters. This book presents a detailed account of recently developed approaches, including the Lasso and versions of it for various models, boosting methods, undirected graphical modeling, and procedures controlling false positive selections. A special characteristic of the book is that it contains comprehensive mathematical theory on high-dimensional statistics combined with methodology, algorithms and illustrations with real data examples. This in-depth approach highlights the methods’ great potential and practical applicability in a variety of settings. As such, it is a valuable resource for researchers, graduate students and experts in statistics, applied mathematics and computer science.

Empirical Likelihood Methods for Pretest-Posttest Studies

Empirical Likelihood Methods for Pretest-Posttest Studies
Author: Min Chen
Publisher:
Total Pages: 130
Release: 2015
Genre:
ISBN:

Download Empirical Likelihood Methods for Pretest-Posttest Studies Book in PDF, Epub and Kindle

Pretest-posttest trials are an important and popular method to assess treatment effects in many scientific fields. In a pretest-posttest study, subjects are randomized into two groups: treatment and control. Before the randomization, the pretest responses and other baseline covariates are recorded. After the randomization and a period of study time, the posttest responses are recorded. Existing methods for analyzing the treatment effect in pretest-posttest designs include the two-sample t-test using only the posttest responses, the paired t-test using the difference of the posttest and the pretest responses, and the analysis of covariance method which assumes a linear model between the posttest and the pretest responses. These methods are summarized and compared by Yang and Tsiatis (2001) under a general semiparametric model which only assumes that the first and second moments of the baseline and the follow-up response variable exist and are finite. Leon et al. (2003) considered a semiparametric model based on counterfactuals, and applied the theory of missing data and causal inference to develop a class of consistent estimator on the treatment effect and identified the most efficient one in the class. Huang et al. (2008) proposed a semiparametric estimation procedure based on empirical likelihood (EL) which incorporates the pretest responses as well as baseline covariates to improve the efficiency. The EL approach proposed by Huang et al. (2008) (the HQF method), however, dealt with the mean responses of the control group and the treatment group separately, and the confidence intervals were constructed through a bootstrap procedure on the conventional normalized Z-statistic. In this thesis, we first explore alternative EL formulations that directly involve the parameter of interest, i.e., the difference of the mean responses between the treatment group and the control group, using an approach similar to Wu and Yan (2012). Pretest responses and other baseline covariates are incorporated to impute the potential posttest responses. We consider the regression imputation as well as the non-parametric kernel imputation. We develop asymptotic distributions of the empirical likelihood ratio statistic that are shown to be scaled chi-squares. The results are used to construct confidence intervals and to conduct statistical hypothesis tests. We also derive the explicit asymptotic variance formula of the HQF estimator, and compare it to the asymptotic variance of the estimator based on our proposed method under several scenarios. We find that the estimator based on our proposed method is more efficient than the HQF estimator under a linear model without an intercept that links the posttest responses and the pretest responses. When there is an intercept, our proposed model is as efficient as the HQF method. When there is misspecification of the working models, our proposed method based on kernel imputation is most efficient. While the treatment effect is of primary interest for the analysis of pretest-posttest sample data, testing the difference of the two distribution functions for the treatment and the control groups is also an important problem. For two independent samples, the nonparametric Mann-Whitney test has been a standard tool for testing the difference of two distribution functions. Owen (2001) presented an EL formulation of the Mann-Whitney test but the computational procedures are heavy due to the use of a U-statistic in the constraints. We develop empirical likelihood based methods for the Mann-Whitney test to incorporate the two unique features of pretest-posttest studies: (i) the availability of baseline information for both groups; and (ii) the missing by design structure of the data. Our proposed methods combine the standard Mann-Whitney test with the empirical likelihood method of Huang, Qin and Follmann (2008), the imputation-based empirical likelihood method of Chen, Wu and Thompson (2014a), and the jackknife empirical likelihood (JEL) method of Jing, Yuan and Zhou (2009). The JEL method provides a major relief on computational burdens with the constrained maximization problems. We also develop bootstrap calibration methods for the proposed EL-based Mann-Whitney test when the corresponding EL ratio statistic does not have a standard asymptotic chi-square distribution. We conduct simulation studies to compare the finite sample performances of the proposed methods. Our results show that the Mann-Whitney test based on the Huang, Qin and Follmann estimators and the test based on the two-sample JEL method perform very well. In addition, incorporating the baseline information for the test makes the test more powerful. Finally, we consider the EL method for the pretest-posttest studies when the design and data collection involve complex surveys. We consider both stratification and inverse probability weighting via propensity scores to balance the distributions of the baseline covariates between two treatment groups. We use a pseudo empirical likelihood approach to make inference of the treatment effect. The proposed methods are illustrated through an application using data from the International Tobacco Control (ITC) Policy Evaluation Project Four Country (4C) Survey.