415 results on '"62F03"'
Search Results
2. Statistical Inference for Chi-square Statistics or F-Statistics Based on Multiple Imputation
- Author
-
Wang, Binhuan, Fang, Yixin, and Jin, Man
- Subjects
Statistics - Methodology ,62F03 - Abstract
Missing data is a common issue in medical, psychiatry, and social studies. In literature, Multiple Imputation (MI) was proposed to multiply impute datasets and combine analysis results from imputed datasets for statistical inference using Rubin's rule. However, Rubin's rule only works for combined inference on statistical tests with point and variance estimates and is not applicable to combine general F-statistics or Chi-square statistics. In this manuscript, we provide a solution to combine F-test statistics from multiply imputed datasets, when the F-statistic has an explicit fractional form (that is, both the numerator and denominator of the F-statistic are reported). Then we extend the method to combine Chi-square statistics from multiply imputed datasets. Furthermore, we develop methods for two commonly applied F-tests, Welch's ANOVA and Type-III tests of fixed effects in mixed effects models, which do not have the explicit fractional form. SAS macros are also developed to facilitate applications., Comment: 21 pages
- Published
- 2024
3. Types of Stickiness in BHV Phylogenetic Tree Spaces and Their Degree
- Author
-
Lammers, Lars, Van, Do Tran, Nye, Tom M. W., and Huckemann, Stephan F.
- Subjects
Mathematics - Statistics Theory ,62F03 - Abstract
It has been observed that the sample mean of certain probability distributions in Billera-Holmes-Vogtmann (BHV) phylogenetic spaces is confined to a lower-dimensional subspace for large enough sample size. This non-standard behavior has been called stickiness and poses difficulties in statistical applications when comparing samples of sticky distributions. We extend previous results on stickiness to show the equivalence of this sampling behavior to topological conditions in the special case of BHV spaces. Furthermore, we propose to alleviate statistical comparision of sticky distributions by including the directional derivatives of the Fr\'echet function: the degree of stickiness., Comment: 8 Pages, 1 Figure, conference submission to GSI 2023
- Published
- 2023
4. Subexponential-Time Algorithms for Sparse PCA.
- Author
-
Ding, Yunzi, Kunisky, Dmitriy, Wein, Alexander S., and Bandeira, Afonso S.
- Subjects
- *
POLYNOMIAL time algorithms , *THRESHOLDING algorithms , *ALGORITHMS , *SEARCH algorithms , *INTERPOLATION algorithms , *RANDOM matrices , *RANDOM graphs - Abstract
We study the computational cost of recovering a unit-norm sparse principal component x ∈ R n planted in a random matrix, in either the Wigner or Wishart spiked model (observing either W + λ x x ⊤ with W drawn from the Gaussian orthogonal ensemble, or N independent samples from N (0 , I n + β x x ⊤) , respectively). Prior work has shown that when the signal-to-noise ratio (λ or β N / n , respectively) is a small constant and the fraction of nonzero entries in the planted vector is ‖ x ‖ 0 / n = ρ , it is possible to recover x in polynomial time if ρ ≲ 1 / n . While it is possible to recover x in exponential time under the weaker condition ρ ≪ 1 , it is believed that polynomial-time recovery is impossible unless ρ ≲ 1 / n . We investigate the precise amount of time required for recovery in the "possible but hard" regime 1 / n ≪ ρ ≪ 1 by exploring the power of subexponential-time algorithms, i.e., algorithms running in time exp (n δ) for some constant δ ∈ (0 , 1) . For any 1 / n ≪ ρ ≪ 1 , we give a recovery algorithm with runtime roughly exp (ρ 2 n) , demonstrating a smooth tradeoff between sparsity and runtime. Our family of algorithms interpolates smoothly between two existing algorithms: the polynomial-time diagonal thresholding algorithm and the exp (ρ n) -time exhaustive search algorithm. Furthermore, by analyzing the low-degree likelihood ratio, we give rigorous evidence suggesting that the tradeoff achieved by our algorithms is optimal. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Post-selection Inference in Multiverse Analysis (PIMA): an inferential framework based on the sign flipping score test
- Author
-
Girardi, Paolo, Vesely, Anna, Lakens, Daniël, Altoè, Gianmarco, Pastore, Massimiliano, Calcagnì, Antonio, and Finos, Livio
- Subjects
Statistics - Methodology ,Statistics - Applications ,62F03 ,G.3 - Abstract
When analyzing data researchers make some decisions that are either arbitrary, based on subjective beliefs about the data generating process, or for which equally justifiable alternative choices could have been made. This wide range of data-analytic choices can be abused, and has been one of the underlying causes of the replication crisis in several fields. Recently, the introduction of multiverse analysis provides researchers with a method to evaluate the stability of the results across reasonable choices that could be made when analyzing data. Multiverse analysis is confined to a descriptive role, lacking a proper and comprehensive inferential procedure. Recently, specification curve analysis adds an inferential procedure to multiverse analysis, but this approach is limited to simple cases related to the linear model, and only allows researchers to infer whether at least one specification rejects the null hypothesis, but not which specifications should be selected. In this paper we present a Post-selection Inference approach to Multiverse Analysis (PIMA) which is a flexible and general inferential approach that accounts for all possible models, i.e., the multiverse of reasonable analyses. The approach allows for a wide range of data specifications (i.e. pre-processing) and any generalized linear model; it allows testing the null hypothesis of a given predictor not being associated with the outcome, by merging information from all reasonable models of multiverse analysis, and provides strong control of the family-wise error rate such that it allows researchers to claim that the null-hypothesis can be rejected for each specification that shows a significant effect. The inferential proposal is based on a conditional resampling procedure. To be continued..., Comment: 37 pages, 2 figures
- Published
- 2022
6. Flexible control of the median of the false discovery proportion
- Author
-
Hemerik, Jesse, Solari, Aldo, and Goeman, Jelle J
- Subjects
Statistics - Methodology ,62F03 - Abstract
We introduce a multiple testing procedure that controls the median of the proportion of false discoveries (FDP) in a flexible way. The procedure only requires a vector of p-values as input and is comparable to the Benjamini-Hochberg method, which controls the mean of the FDP. Our method allows freely choosing one or several values of alpha after seeing the data -- unlike Benjamini-Hochberg, which can be very liberal when alpha is chosen post hoc. We prove these claims and illustrate them with simulations. Our procedure is inspired by a popular estimator of the total number of true hypotheses. We adapt this estimator to provide simultaneously median unbiased estimators of the FDP, valid for finite samples. This simultaneity allows for the claimed flexibility. Our approach does not assume independence. The time complexity of our method is linear in the number of hypotheses, after sorting the p-values.
- Published
- 2022
7. Fundamental Frequency and its Harmonics Model: A Robust Method of Estimation.
- Author
-
Kundu, Debasis
- Subjects
- *
LEAST squares , *NONLINEAR equations , *ASYMPTOTIC normality , *PARAMETER estimation - Abstract
In this paper we have proposed a novel robust method of estimation of the unknown parameters of a fundamental frequency and its harmonics model. Although the least squares estimators (LSEs) or the periodogram type estimators are the most efficient estimators, it is well known that they are not robust. In presence of outliers the LSEs are known to be not efficient. In presence of outliers, robust estimators like least absolute deviation estimators (LADEs) or Huber's M-estimators (HMEs) may be used. But implementation of the LADEs or HMEs are quite challenging, particularly if the number of component is large. Finding initial guesses in the higher dimensions is always a non-trivial issue. Moreover, theoretical properties of the robust estimators can be established under stronger assumptions than what are needed for the LSEs. In this paper we have proposed novel weighted least squares estimators (WLSEs) which are more robust compared to the LSEs or periodogram estimators in presence of outliers. The proposed WLSEs can be implemented very conveniently in practice. It involves in solving only one non-linear equation. We have established the theoretical properties of the proposed WLSEs. Extensive simulations suggest that in presence of outliers, the WLSEs behave better than the LSEs, periodogram estimators, LADEs and HMEs. The performance of the WLSEs depend on the weight function, and we have discussed how to choose the weight function. We have analyzed one synthetic data set to show how the proposed method can be used in practice. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. On Determining the Distribution of a Goodness-of-Fit Test Statistic
- Author
-
van der Merwe, Sean
- Subjects
Statistics - Methodology ,62F03 - Abstract
We consider the problem of goodness-of-fit testing for a model that has at least one unknown parameter that cannot be eliminated by transformation. Examples of such problems can be as simple as testing whether a sample consists of independent Gamma observations, or whether a sample consists of independent Generalised Pareto observations given a threshold. Over time the approach to determining the distribution of a test statistic for such a problem has moved towards on-the-fly calculation post observing a sample. Modern approaches include the parametric bootstrap and posterior predictive checks. We argue that these approaches are merely approximations to integrating over the posterior predictive distribution that flows naturally from a given model. Further, we attempt to demonstrate that shortcomings which may be present in the parametric bootstrap, especially in small samples, can be reduced through the use of objective Bayes techniques, in order to more reliably produce a test with the correct size., Comment: Bayes, Distribution, Gamma, GPD, Hypothesis Testing, Objective Bayes, p-value, Predictive Posterior, Simulation
- Published
- 2021
9. Extension of the Lagrange multiplier test for error cross-section independence to large panels with non normal errors
- Author
-
Li, Zhaoyuan and Yao, Jianfeng
- Subjects
Economics - Econometrics ,Statistics - Methodology ,62F03 - Abstract
This paper reexamines the seminal Lagrange multiplier test for cross-section independence in a large panel model where both the number of cross-sectional units n and the number of time series observations T can be large. The first contribution of the paper is an enlargement of the test with two extensions: firstly the new asymptotic normality is derived in a simultaneous limiting scheme where the two dimensions (n, T) tend to infinity with comparable magnitudes; second, the result is valid for general error distribution (not necessarily normal). The second contribution of the paper is a new test statistic based on the sum of the fourth powers of cross-section correlations from OLS residuals, instead of their squares used in the Lagrange multiplier statistic. This new test is generally more powerful, and the improvement is particularly visible against alternatives with weak or sparse cross-section dependence. Both simulation study and real data analysis are proposed to demonstrate the advantages of the enlarged Lagrange multiplier test and the power enhanced test in comparison with the existing procedures., Comment: 39 papers, 8 figures
- Published
- 2021
10. A comprehensive empirical power comparison of univariate goodness-of-fit tests for the Laplace distribution
- Author
-
Desgagné, Alain, de Micheaux, Pierre Lafaye, and Ouimet, Frédéric
- Subjects
Statistics - Methodology ,62F03 - Abstract
In this paper we present the results from an empirical power comparison of 40 goodness-of-fit tests for the univariate Laplace distribution, carried out using Monte Carlo simulations with sample sizes $n = 20, 50, 100, 200$, significance levels $\alpha = 0.01, 0.05, 0.10$, and 400 alternatives consisting of asymmetric and symmetric light/heavy-tailed distributions taken as special cases from 11 models. In addition to the unmatched scope of our study, an interesting contribution is the proposal of an innovative design for the selection of alternatives. The 400 alternatives consist of 20 specific cases of 20 submodels drawn from the main 11 models. For each submodel, the 20 specific cases corresponded to parameter values chosen to cover the full power range. An analysis of the results leads to a recommendation of the best tests for five different groupings of the alternative distributions. A real-data example is also presented, where an appropriate test for the goodness-of-fit of the univariate Laplace distribution is applied to weekly log-returns of Amazon stock over a recent four-year period., Comment: 48 pages, 2 figures, 38 tables
- Published
- 2020
- Full Text
- View/download PDF
11. A stationary Weibull process and its applications.
- Author
-
Kundu, Debasis
- Subjects
- *
CUMULATIVE distribution function , *GOLD sales & prices , *MAXIMUM likelihood statistics , *WEIBULL distribution , *STATIONARY processes , *DISTRIBUTION (Probability theory) , *MARKOV processes - Abstract
In this paper we introduce a discrete-time and continuous state-space Markov stationary process { X n ; n = 1 , 2 , ... } , where X n has a two-parameter Weibull distribution, X n 's are dependent and there is a positive probability that X n = X n + 1 . The motivation came from the gold price data where there are several instances for which X n = X n + 1 . Hence, the existing methods cannot be used to analyze this data. We derive different properties of the proposed Weibull process. It is observed that the joint cumulative distribution function of X n and X n + 1 has a very convenient copula structure. Hence, different dependence properties and dependence measures can be obtained. The maximum likelihood estimators cannot be obtained in explicit forms, we have proposed a simple profile likelihood method to compute these estimators. We have used this model to analyze two synthetic data sets and one gold price data set of the Indian market, and it is observed that the proposed model fits quite well with the data set. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
12. Alternating minimization algorithm with a probability generating function-based distance measure
- Author
-
Rajakrishnan, Ranusha, Ong, Seng Huat, and Ng, Choung Min
- Published
- 2024
- Full Text
- View/download PDF
13. Should we test the model assumptions before running a model-based test?
- Author
-
Shamsudheen, M. Iqbal and Hennig, Christian
- Subjects
Statistics - Methodology ,62F03 - Abstract
Statistical methods are based on model assumptions, and it is statistical folklore that a method's model assumptions should be checked before applying it. This can be formally done by running one or more misspecification tests of model assumptions before running a method that requires these assumptions; here we focus on model-based tests. A combined test procedure can be defined by specifying a protocol in which first model assumptions are tested and then, conditionally on the outcome, a test is run that requires or does not require the tested assumptions. Although such an approach is often taken in practice, much of the literature that investigated this is surprisingly critical of it. Our aim is to explore conditions under which model checking is advisable or not advisable. For this, we review results regarding such "combined procedures" in the literature, we review and discuss controversial views on the role of model checking in statistics, and we present a general setup in which we can show that preliminary model checking is advantageous, which implies conditions for making model checking worthwhile., Comment: 35 pages, 1 figure
- Published
- 2019
14. A technical note on divergence of the Wald statistic
- Author
-
Dufour, Jean-Marie, Renault, Eric, and Zinde-Walsh, Victoria
- Subjects
Mathematics - Statistics Theory ,62F03 - Abstract
The Wald test statistic has been shown to diverge (Dufour et al, 2013, 2017) under some conditions. This note links the divergence to eigenvalues of a polynomial matrix and establishes the divergence rate.
- Published
- 2019
15. Interpretable hypothesis tests
- Author
-
Coscrato, Victor, Esteves, Luís Gustavo, Izbicki, Rafael, and Stern, Rafael Bassi
- Subjects
Statistics - Methodology ,62F03 - Abstract
Although hypothesis tests play a prominent role in Science, their interpretation can be challenging. Three issues are (i) the difficulty in making an assertive decision based on the output of an hypothesis test, (ii) the logical contradictions that occur in multiple hypothesis testing, and (iii) the possible lack of practical importance when rejecting a precise hypothesis. These issues can be addressed through the use of agnostic tests and pragmatic hypotheses., Comment: 19 pages, 3 figures
- Published
- 2019
16. A goodness-of-fit test for elliptical distributions with diagnostic capabilities
- Author
-
Ducharme, Gilles R. and de Micheaux, Pierre Lafaye
- Subjects
Mathematics - Statistics Theory ,62F03 - Abstract
This paper develops a smooth test of goodness-of-fit for elliptical distributions. The test is adaptively omnibus, invariant to affine-linear transformations and has a convenient expression that can be broken into components. These components have diagnostic capabilities and can be used to identify specific departures. This helps in correcting the null model when the test rejects. As an example, the results are applied to the multivariate normal distribution for which the R package ECGofTestDx is available. It is shown that the proposed test strategy encompasses and generalizes a number of existing approaches. Some other cases are studied, such as the bivariate Laplace, logistic and Pearson type II distribution. A simulation experiment shows the usefulness of the diagnostic tools., Comment: 35 p. pre-print
- Published
- 2019
17. Only Closed Testing Procedures are Admissible for Controlling False Discovery Proportions
- Author
-
Goeman, Jelle, Hemerik, Jesse, and Solari, Aldo
- Subjects
Mathematics - Statistics Theory ,Statistics - Methodology ,62F03 - Abstract
We consider the class of all multiple testing methods controlling tail probabilities of the false discovery proportion, either for one random set or simultaneously for many such sets. This class encompasses methods controlling familywise error rate, generalized familywise error rate, false discovery exceedance, joint error rate, simultaneous control of all false discovery proportions, and others, as well as seemingly unrelated methods such as gene set testing in genomics and cluster inference methods in neuroimaging. We show that all such methods are either equivalent to a closed testing method, or are uniformly improved by one. Moreover, we show that a closed testing method is admissible as a method controlling tail probabilities of false discovery proportions if and only if all its local tests are admissible. This implies that, when designing such methods, it is sufficient to restrict attention to closed testing methods only. We demonstrate the practical usefulness of this design principle by constructing a uniform improvement of a recently proposed method.
- Published
- 2019
- Full Text
- View/download PDF
18. Testing for explosive bubbles: a review
- Author
-
Skrobotov Anton
- Subjects
rational bubble ,testing for explosive bubble ,explosive autoregression ,time-varying volatility ,right-tailed unit root testing ,62f03 ,62m10 ,62p20 ,91b84 ,Science (General) ,Q1-390 ,Mathematics ,QA1-939 - Abstract
This review discusses methods of testing for explosive bubbles in time series. A large number of recently developed testing methods under various assumptions about innovation of errors are covered. The review also considers the methods for dating explosive (bubble) regimes. Special attention is devoted to time-varying volatility in the errors. Moreover, the modelling of possible relationships between time series with explosive regimes is discussed.
- Published
- 2023
- Full Text
- View/download PDF
19. Generalized fiducial methods for testing the homogeneity of a three-sample problem with a mixture structure.
- Author
-
Ren, Pengcheng, Liu, Guanfu, and Pu, Xiaolong
- Subjects
- *
TEST methods , *HOMOGENEITY , *MIXTURES , *SAMPLE size (Statistics) - Abstract
Recently, the likelihood ratio (LR) test was proposed to test the homogeneity of a three-sample model with a mixture structure. Because of the presence of the mixture structure, the null limiting distribution of the LR test has a complicated supremum form, which leads to challenges in determining p-values. In addition, the LR test cannot control type-I errors well under small to moderate sample size. In this paper, we propose seven generalized fiducial methods to test the homogeneity of the three-sample model. Via simulation studies, we find that our methods perform significantly better than the LR test method in controlling the type-I errors under small to moderate sample size, while they have comparable powers in most cases. A halibut data example is used to illustrate the proposed methods. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
20. Agnostic tests can control the type I and type II errors simultaneously
- Author
-
Coscrato, Victor, Izbicki, Rafael, and Stern, Rafael Bassi
- Subjects
Mathematics - Statistics Theory ,62F03 - Abstract
Despite its common practice, statistical hypothesis testing presents challenges in interpretation. For instance, in the standard frequentist framework there is no control of the type II error. As a result, the non-rejection of the null hypothesis cannot reasonably be interpreted as its acceptance. We propose that this dilemma can be overcome by using agnostic hypothesis tests, since they can control the type I and II errors simultaneously. In order to make this idea operational, we show how to obtain agnostic hypothesis in typical models. For instance, we show how to build (unbiased) uniformly most powerful agnostic tests and how to obtain agnostic tests from standard p-values. Also, we present conditions such that the above tests can be made logically coherent. Finally, we present examples of consistent agnostic hypothesis tests., Comment: 21 pages, 8 figures
- Published
- 2018
21. A novel approach to compare the spectral densities of some uncorrelated cyclostationary time series
- Author
-
Mohammad Reza Mahmoudi, Maria Rayisyan, Reza Vaghefi, Shahab S. Band, and Amir H. Mosavi
- Subjects
62M10 ,62M15 ,62F03 ,62G07 ,62G20 ,62P20 ,Engineering (General). Civil engineering (General) ,TA1-2040 - Abstract
Our primary objective in this article is to compare the spectral densities of some cyclostationary time series. By using the limiting distributions of the discrete Fourier transform, a novel approach is introduced to determine whether the spectral densities of some uncorrelated cyclostationary time series are the same or not. Also, the ability of the proposed technique is examined by employing simulated and real datasets.
- Published
- 2022
- Full Text
- View/download PDF
22. Are discoveries spurious? Distributions of maximum spurious correlations and their applications
- Author
-
Fan, Jianqing, Shao, Qi-Man, and Zhou, Wen-Xin
- Subjects
Mathematical Sciences ,Statistics ,High dimension ,spurious correlation ,bootstrap ,false discovery ,math.ST ,stat.ME ,stat.TH ,62H10 ,62H20 ,62E17 ,62F03 ,Applied Mathematics ,Econometrics ,Statistics & Probability - Abstract
Over the last two decades, many exciting variable selection methods have been developed for finding a small group of covariates that are associated with the response from a large pool. Can the discoveries by such data mining approaches be spurious due to high dimensionality and limited sample size? Can our fundamental assumptions on exogeneity of covariates needed for such variable selection be validated with the data? To answer these questions, we need to derive the distributions of the maximum spurious correlations given certain number of predictors, namely, the distribution of the correlation of a response variable Y with the best s linear combinations of p covariates X, even when X and Y are independent. When the covariance matrix of X possesses the restricted eigenvalue property, we derive such distributions for both finite s and diverging s, using Gaussian approximation and empirical process techniques. However, such a distribution depends on the unknown covariance matrix of X. Hence, we use the multiplier bootstrap procedure to approximate the unknown distributions and establish the consistency of such a simple bootstrap approach. The results are further extended to the situation where residuals are from regularized fits. Our approach is then applied to construct the upper confidence limit for the maximum spurious correlation and testing exogeneity of covariates. The former provides a baseline for guarding against false discoveries due to data mining and the latter tests whether our fundamental assumptions for high-dimensional model selection are statistically valid. Our techniques and results are illustrated by both numerical examples and real data analysis.
- Published
- 2018
23. Some Permutationllay Symmetric Multiple Hypotheses Testing Rules Under Dependent Set up
- Author
-
Kundu, Anupam and Bhandari, Subir Kumar
- Subjects
Mathematics - Statistics Theory ,62F03 - Abstract
In this paper, our interest is in the problem of simultaneous hypothesis testing when the test statistics corresponding to the individual hypotheses are possibly correlated. Specifically, we consider the case when the test statistics together have a multivariate normal distribution (with equal correlation between each pair) with an unknown mean vector and our goal is to decide which components of the mean vector are zero and which are non-zero. This problem was taken up earlier in Bogdan et al. (2011) for the case when the test statistics are independent normals. Asymptotic optimality in a Bayesian decision theoretic sense was studied in this context, the optimal precodures were characterized and optimality of some well-known procedures were thereby established. The case under dependence was left as a challenging open problem. We have studied the problem both theoretically and through extensive simulations and have given some permutation invariant rules. Though in Bogdan et al. (2011), the asymptotic derivations were done in the context of sparsity of the non-zero means, our result does not require the assumption of sparsity and holds under a more general setup., Comment: 16 pages, 0 figures, 2 tables
- Published
- 2016
24. Selective inference after cross-validation
- Author
-
Loftus, Joshua R.
- Subjects
Statistics - Methodology ,62F03 - Abstract
This paper describes a method for performing inference on models chosen by cross-validation. When the test error being minimized in cross-validation is a residual sum of squares it can be written as a quadratic form. This allows us to apply the inference framework in Loftus et al. (2015) for models determined by quadratic constraints to the model that minimizes CV test error. Our only requirement on the model training pro- cedure is that its selection events are regions satisfying linear or quadratic constraints. This includes both Lasso and forward stepwise, which serve as our main examples throughout. We do not require knowledge of the error variance $\sigma^2$. The procedures described here are computationally intensive methods of selecting models adaptively and performing inference for the selected model. Implementations are available in an R package., Comment: 9 pages, 2 figures
- Published
- 2015
25. Stability and uniqueness of $p$-values for likelihood-based inference
- Author
-
DiCiccio, Thomas J., Kuffner, Todd A., Young, G. Alastair, and Zaretzki, Russell
- Subjects
Mathematics - Statistics Theory ,62F03 - Abstract
Likelihood-based methods of statistical inference provide a useful general methodology that is appealing, as a straightforward asymptotic theory can be applied for their implementation. It is important to assess the relationships between different likelihood-based inferential procedures in terms of accuracy and adherence to key principles of statistical inference, in particular those relating to conditioning on relevant ancillary statistics. An analysis is given of the stability properties of a general class of likelihood-based statistics, including those derived from forms of adjusted profile likelihood, and comparisons are made between inferences derived from different statistics. In particular, we derive a set of sufficient conditions for agreement to $O_{p}(n^{-1})$, in terms of the sample size $n$, of inferences, specifically $p$-values, derived from different asymptotically standard normal pivots. Our analysis includes inference problems concerning a scalar or vector interest parameter, in the presence of a nuisance parameter., Comment: Accepted for publication in Statistica Sinica
- Published
- 2015
26. Meta-analysis of ratios of sample variances
- Author
-
Prendergast, Luke A. and Staudte, Robert G.
- Subjects
Statistics - Methodology ,62F03 - Abstract
When conducting a meta-analysis of standardized mean differences (SMDs), it is common to assume equal variances in the two arms of each study. This leads to Cohen's $d$ estimates for which interpretation is simple. However, this simplicity should not be used as a justification for the assumption of equal variances in situations where evidence may suggest that it is incorrect. Until now, researchers have either used an $F$-test for each individual study as a justification for the equality of variances or perhaps even conveniently ignored such tools altogether. In this paper we propose using a meta-analysis of F-test statistics to estimate the ratio of variances prior to the combination of SMD's. This procedure allows some studies to be included that might otherwise be omitted by individual fixed level tests for unequal variances, sometimes occur even when the assumption of equal variances holds. The estimated ratio of variances, as well as associated confidence intervals, can be used as guidance as to whether the assumption of equal variances is violated. The estimators considered include variance stabilization transformations (VST) of the $F$-test statistics as well as MLE estimators. The VST approaches enable the use of QQ-plots to visually inspect for violations of equal variances while the MLE estimator easily allows for the introduction of a random effect. When there is evidence of unequal variances, this work provides a means to formally justify the use of less common methods such as log ratio of means when studies are measured on a different scale.
- Published
- 2014
- Full Text
- View/download PDF
27. Adaptive False Discovery Rate Control for Heterogeneous Data
- Author
-
Habiger, Joshua D.
- Subjects
Statistics - Methodology ,62F03 - Abstract
Efforts to develop more efficient multiple hypothesis testing procedures for false discovery rate (FDR) control have focused on incorporating an estimate of the proportion of true null hypotheses (such procedures are called adaptive) or exploiting heterogeneity across tests via some optimal weighting scheme. This paper combines these approaches using a weighted adaptive multiple decision function (WAMDF) framework. Optimal weights for a flexible random effects model are derived and a WAMDF that controls the FDR for arbitrary weighting schemes when test statistics are independent under the null hypotheses is given. Asymptotic and numerical assessment reveals that, under weak dependence, the proposed WAMDFs provide more efficient FDR control even if optimal weights are misspecified. The robustness and flexibility of the proposed methodology facilitates the development of more efficient, yet practical, FDR procedures for heterogeneous data. To illustrate, two different weighted adaptive FDR methods for heterogeneous sample sizes are developed and applied to data.
- Published
- 2014
28. A review of tests for exponentiality with Monte Carlo comparisons.
- Author
-
Ossai, Everestus O., Madukaife, Mbanefo S., and Oladugba, Abimibola V.
- Subjects
- *
MONTE Carlo method , *HAZARD function (Statistics) - Abstract
In this paper, 91 different tests for exponentiality are reviewed. Some of the tests are universally consistent while others are against some special classes of life distributions. Power performances of 40 of these different tests for exponentiality of datasets are compared through extensive Monte Carlo simulations. The comparisons are conducted for different sample sizes of 10, 25, 50 and 100 for different groups of distributions according to the shape of their hazard functions at 5 percent level of significance. Also, the techniques are applied to two real-world datasets and a measure of power is employed for the comparison of the tests. The results show that some tests which are very good under one group of alternative distributions are not so under another group. Also, some tests maintained relatively high power over all the groups of alternative distributions studied while some others maintained poor power performances over all the groups of alternative distributions. Again, the result obtained from real-world datasets agree completely with those of the simulation studies. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
29. Exact Asymptotics for the Scan Statistic and Fast Alternatives
- Author
-
Sharpnack, James and Arias-Castro, Ery
- Subjects
Mathematics - Statistics Theory ,62F03 - Abstract
We consider the problem of detecting a rectangle of activation in a grid of sensors in d-dimensions with noisy measurements. This has applications to massive surveillance projects and anomaly detection in large datasets in which one detects anomalously high measurements over rectangular regions, or more generally, blobs. Recently, the asymptotic distribution of a multiscale scan statistic was established in (Kabluchko, 2011) under the null hypothesis, using non-constant boundary crossing probabilities for locally-stationary Gaussian random fields derived in (Chan and Lai, 2006). Using a similar approach, we derive the exact asymptotic level and power of four variants of the scan statistic: an oracle scan that knows the dimensions of the activation rectangle; the multiscale scan statistic just mentioned; an adaptive variant; and an epsilon-net approximation to the latter, in the spirit of (Arias-Castro, 2005). This approximate scan runs in time near-linear in the size of the grid and achieves the same asymptotic power as the adaptive scan. We complement our theory with some numerical experiments.
- Published
- 2014
30. A modified $\chi^2$-test for uplift models with applications in marketing performance measurement
- Author
-
Michel, Rene, Schnakenburg, Igor, and von Martens, Tobias
- Subjects
Mathematics - Statistics Theory ,62F03 ,G.3 - Abstract
Uplift, essentially being the difference between two probabilities, is a central number in marketing performance measurement. A frequent question in applications is whether the uplifts of two campaigns are significantly different. In this article we present a new $\chi^2$-statistic which allows to answer this question by performing a statistical test. We show that this statistic is asymptotically $\chi^2$-distributed and demonstrate its application in a real life example. By running simulations with this new and alternative approaches, we find our suggested test to exhibit a better decisive power., Comment: 18 pages, 7 figures
- Published
- 2014
31. Estimation of False Discovery Proportion with Unknown Dependence
- Author
-
Fan, Jianqing and Han, Xu
- Subjects
Statistics - Methodology ,62F03 - Abstract
Large-scale multiple testing with highly correlated test statistics arises frequently in many scientific research. Incorporating correlation information in estimating false discovery proportion has attracted increasing attention in recent years. When the covariance matrix of test statistics is known, Fan, Han & Gu (2012) provided a consistent estimate of False Discovery Proportion (FDP) under arbitrary dependence structure. However, the covariance matrix is often unknown in many applications and such dependence information has to be estimated before estimating FDP (Efron, 2010). The estimation accuracy can greatly affect the convergence result of FDP or even violate its consistency. In the current paper, we provide methodological modification and theoretical investigations for estimation of FDP with unknown covariance. First we develop requirements for estimates of eigenvalues and eigenvectors such that we can obtain a consistent estimate of FDP. Secondly we give conditions on the dependence structures such that the estimate of FDP is consistent. Such dependence structures include sparse covariance matrices, which have been popularly considered in the contemporary random matrix theory. When data are sampled from an approximate factor model, which encompasses most practical situations, we provide a consistent estimate of FDP via exploiting this specific dependence structure. The results are further demonstrated by simulation studies and some real data applications., Comment: 39 pages, 7 figures
- Published
- 2013
32. Testing approximate normality of an estimator using the estimated MSE and bias with an application to the shape parameter of the generalized Pareto distribution
- Author
-
van Zyl, J. Martin
- Subjects
Statistics - Applications ,62F03 - Abstract
Often it is not easy to choose between estimators, based on the estimated MSE and bias using simulation studies. Normality in small samples and a variance of the estimator, which is correct and easy to calculate using a single sample, give the added advantage that hypotheses concerning the parameter can be tested in new samples. A procedure to check normality is proposed where previously published MSE and bias are used to perform a test for normality. A confidence interval for the index of the S&P500 index is found by applying the results to estimators of the generalized Pareto distribution.
- Published
- 2012
33. Appropriate Methodology of Statistical Tests According to Prior Probability and Required Objectivity
- Author
-
Konishi, Tomokazu
- Subjects
Statistics - Methodology ,62F03 - Abstract
In contrast to its common definition and calculation, interpretation of p-values diverges among statisticians. Since p-value is the basis of various methodologies, this divergence has led to a variety of test methodologies and evaluations of test results. This chaotic situation has complicated the application of tests and decision processes. Here, the origin of the divergence is found in the prior probability of the test. Effects of difference in Pr(H0 = true) on the character of p-values are investigated by comparing real microarray data and its artificial imitations as subjects of Student's t-tests. Also, the importance of the prior probability is discussed in terms of the applicability of Bayesian approaches. Suitable methodology is found in accordance with the prior probability and purpose of the test., Comment: 16 pages, 3 figures, and 1 table
- Published
- 2011
- Full Text
- View/download PDF
34. Ratio Estimators in Simple Random Sampling when Study Variable is an Attribute
- Author
-
Singh, Rajesh, Kumar, Mukesh, and Smarandache, Florentin
- Subjects
Mathematics - General Mathematics ,62F03 - Abstract
In this paper we have suggested a family of estimators for the population mean when study variable itself is qualitative in nature. Expressions for the bias and mean square error (MSE) of the suggested family have been obtained. An empirical study has been carried out to show the superiority of the constructed estimator over others., Comment: 9 pages
- Published
- 2010
35. Extending The Range of Application of Permutation Tests: the Expected Permutation p-value Approach
- Author
-
Commenges, Daniel
- Subjects
Statistics - Methodology ,62F03 ,62F40 - Abstract
The limitation of permutation tests is that they assume exchangeability. It is shown that in generalized linear models one can construct permutation tests from score statistics in particular cases. When under the null hypothesis the observations are not exchangeable, a representation in terms of Cox-Snell residuals allows to develop an approach based on an expected permutation p-value (Eppv); this is applied to the logistic regression model. A small simulation study and an illustration with real data are given., Comment: 15 pages, 1 table
- Published
- 2010
36. Dual divergence estimators and tests: robustness results
- Author
-
Toma, Aida and Broniatowski, Michel
- Subjects
Mathematics - Statistics Theory ,62F10 ,62F03 ,62G35 - Abstract
The class of dual $\phi$-divergence estimators (introduced in Broniatowski and Keziou (2009) is explored with respect to robustness through the influence function approach. For scale and location models, this class is investigated in terms of robustness and asymptotic relative efficiency. Some hypothesis tests based on dual divergence criterions are proposed and their robustness properties are studied. The empirical performances of these estimators and tests are illustrated by Monte Carlo simulation for both noncontaminated and contaminated data.
- Published
- 2009
37. Ratio Estimators in Simple Random Sampling Using Information on Auxiliary Attribute
- Author
-
Singh, Rajesh, Chauhan, Pankaj, Sawan, Nirmala, and Smarandache, Florentin
- Subjects
Mathematics - General Mathematics ,62F03 - Abstract
Some ratio estimators for estimating the population mean of the variable under study, which make use of information regarding the population proportion possessing certain attribute, are proposed. Under simple random sampling without replacement (SRSWOR) scheme, the expressions of bias and mean-squared error (MSE) up to the first order of approximation are derived. The results obtained have been illustrated numerically by taking some empirical population considered in the literature., Comment: 7 pages
- Published
- 2009
38. Optimum Statistical Test Procedure
- Author
-
Singh, Rajesh, Singh, Jayant, and Smarandache, Florentin
- Subjects
Mathematics - General Mathematics ,62F03 - Abstract
In this paper we obtain a test which minimizes the sum of the two error probabilities irrespective of whether $\sigma^2$ is known or unknown., Comment: 16 pages
- Published
- 2009
39. Parametric estimation and tests through divergences and duality technique
- Author
-
Broniatowski, Michel and Keziou, Amor
- Subjects
Mathematics - Statistics Theory ,62F03 ,62F10 ,62F30 - Abstract
We introduce estimation and test procedures through divergence optimization for discrete or continuous parametric models. This approach is based on a new dual representation for divergences. We treat point estimation and tests for simple and composite hypotheses, extending maximum likelihood technique. An other view at the maximum likelihood approach, for estimation and test, is given. We prove existence and consistency of the proposed estimates. The limit laws of the estimates and test statistics (including the generalized likelihood ratio one) are given both under the null and the alternative hypotheses, and approximation of the power functions is deduced. A new procedure of construction of confidence regions, when the parameter may be a boundary value of the parameter space, is proposed. Also, a solution to the irregularity problem of the generalized likelihood ratio test pertaining to the number of components in a mixture is given, and a new test is proposed, based on $\chi ^{2}$-divergence on signed finite measures and duality technique.
- Published
- 2008
40. How to define and test explanations in populations
- Author
-
Peter J. Veazie
- Subjects
62a01 ,62f03 ,Mathematics ,QA1-939 ,Probabilities. Mathematical statistics ,QA273-280 - Abstract
Solving applied social, economic, psychological, health care and public health problems can require an understanding of facts or phenomena related to populations of interest. Therefore, it can be useful to test whether an explanation of a phenomenon holds in a population. However, different definitions for the phrase “explain in a population” lead to different interpretations and methods of testing. In this paper, I present two definitions: The first is based on the number of members in the population that conform to the explanation’s implications; the second is based on the total magnitude of explanation-consistent effects in the population. I show that claims based on either definition can be tested using random coefficient models, but claims based on the second definition can also be tested using the more common, and simpler, population-level regression models. Consequently, this paper provides an understanding of the type of explanatory claims these common methods can test.
- Published
- 2019
- Full Text
- View/download PDF
41. A Note on Testing of Hypothesis
- Author
-
Singh, Rajesh, Singh, Jayant, and Smarandache, Florentin
- Subjects
Mathematics - General Mathematics ,62F03 - Abstract
In this paper, a problem of testing is discussed when the samples have been drawn from the normal distribution. The study of hypothesis testing is also extended to Baye's set up., Comment: 5 pages
- Published
- 2008
42. Computation of Power Loss in Likelihood Ratio Tests for Probability Densities Extended by Lehmann Alternatives
- Author
-
Soares, Lucas Gallindo Martins
- Subjects
Mathematics - Statistics Theory ,Mathematics - Probability ,Statistics - Methodology ,62B10 ,62F03 ,62E10 - Abstract
We compute the loss of power in likelihood ratio tests when we test the original parameter of a probability density extended by the first Lehmann alternative., Comment: 6 pages, one figure; Corrected minor typographic errors (pay special attention to one in equation 4)
- Published
- 2007
43. An Extension to Gaussian Semigroup and Some Applications
- Author
-
Liu, Guibao
- Subjects
Mathematics - Probability ,Quantitative Finance - Computational Finance ,60J35 ,35K05 ,62F03 - Abstract
We look at the semigroup generated by a system of heat equations. Applications to testing normality and option pricing are addressed., Comment: 16 pages, 4 figures
- Published
- 2006
44. Exact inequalities for sums of asymmetric random variables, with applications
- Author
-
Pinelis, Iosif
- Subjects
Mathematics - Probability ,Mathematics - Statistics Theory ,60E15 ,60G50 ,60G42 ,60G48 ,62F03 ,62F25 ,62G10 ,60G15 ,60E05 ,62E10 ,62G35 - Abstract
Let $\BS_1,...,\BS_n$ be independent identically distributed random variables each having the standardized Bernoulli distribution with parameter $p\in(0,1)$. Let $m_*(p):=(1+p+2p^2)/(2\sqrt{p-p^2}+4p^2)$ if $0
- Published
- 2006
- Full Text
- View/download PDF
45. Double and group acceptance sampling plan for truncated life test based on inverse log-logistic distribution.
- Author
-
Tripathi, Harsh, Dey, Sanku, and Saha, Mahendra
- Subjects
- *
ACCEPTANCE sampling , *SAMPLING (Process) , *CHARACTERISTIC functions , *SAMPLE size (Statistics) - Abstract
This paper introduces a double and group acceptance sampling plans based on time truncated lifetimes when the lifetime of an item follows the inverse log-logistic (ILL) distribution with known shape parameter. The operating characteristic function and average sample number (ASN) values of the double acceptance sampling inspection plan are provided. The values of the minimum number of groups and operating characteristic function for various quality levels are obtained for a group acceptance sampling inspection plan. A comparative study between single acceptance sampling inspection plan and double acceptance sampling inspection plan is carried out in terms of sample size. One simulated example and four real-life examples are discussed to show the applicability of the proposed double and group acceptance sampling inspection plans for ILL distributed quality parameters. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
46. Testing simultaneously different covariance block diagonal structures – the multi-sample case.
- Author
-
Marques, F. J. and Coelho, C. A.
- Subjects
- *
LIKELIHOOD ratio tests , *COVARIANCE matrices , *GAMMA distributions - Abstract
In this work a likelihood ratio test which allows to test simultaneously if, several covariance matrices are equal and block diagonal with different specific structures in the diagonal blocks, is developed. The distribution of the likelihood ratio statistic is studied and the expression of its hth null moment are derived. In order to make the test useful in practical terms, near-exact approximations are developed for the likelihood ratio statistic. A practical application to real data set together with numerical studies and simulations are provided in order illustrate the applicability of the test and also to assess the precision of the near-exact approximations developed. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
47. Asymptotic normality of the test statistics for the unified relative dispersion and relative variation indexes.
- Author
-
Touré, Aboubacar Y., Dossou-Gbété, Simplice, and Kokonendji, Célestin C.
- Subjects
- *
ASYMPTOTIC distribution , *BINOMIAL distribution , *ASYMPTOTIC normality , *POISSON distribution , *STATISTICS , *EXPONENTIAL families (Statistics) , *DISPERSION (Chemistry) - Abstract
Dispersion indexes with respect to the Poisson and binomial distributions are widely used to assess the conformity of the underlying distribution from an observed sample of the count with one or the other of these theoretical distributions. Recently, the exponential variation index has been proposed as an extension to nonnegative continuous data. This paper aims to gather to study the unified definition of these indexes with respect to the relative variability of a nonnegative natural exponential family of distributions through its variance function. We establish the strong consistency of the plug-in estimators of the indexes as well as their asymptotic normalities. Since the exact distributions of the estimators are not available in closed form, we consider the test of hypothesis relying on these estimators as test statistics with their asymptotic distributions. Simulation studies globally suggest good behaviours of these tests of hypothesis procedures. Applicable examples are analysed, including the lesser-known references such as negative binomial and inverse Gaussian, and improving the very usual case of the Poisson dispersion index. Concluding remarks are made with suggestions of possible extensions. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
48. Goodness-of-fit tests for the logistic location family.
- Author
-
Nikitin, Ya. Yu. and Ragozin, I. A.
- Subjects
- *
GOODNESS-of-fit tests , *U-statistics - Abstract
We construct two U-empirical tests for the logistic location family which are based on appropriate characterization of this family using independent exponential shifts. We study the limiting distributions and local Bahadur efficiency of corresponding test statistics under close alternatives. It turns out that the present tests are considerably more efficient than the recently proposed similar tests based on another characterization. The efficiency calculations are accompanied by the simulation of power for new tests together with the previous ones.Both efficiency and power turn out to be very high. Finally we consider the application of our tests to real data example. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
49. Quasi-binomial zero-inflated regression model suitable for variables with bounded support.
- Author
-
Gómez–Déniz, E., Gallardo, D. I., and Gómez, H. W.
- Subjects
- *
REGRESSION analysis , *BINOMIAL distribution , *POISSON distribution , *SPECIAL functions - Abstract
In recent years, a variety of regression models, including zero-inflated and hurdle versions, have been proposed to explain the case of a dependent variable with respect to exogenous covariates. Apart from the classical Poisson, negative binomial and generalised Poisson distributions, many proposals have appeared in the statistical literature, perhaps in response to the new possibilities offered by advanced software that now enables researchers to implement numerous special functions in a relatively simple way. However, we believe that a significant research gap remains, since very little attention has been paid to the quasi-binomial distribution, which was first proposed over fifty years ago. We believe this distribution might constitute a valid alternative to existing regression models, in situations in which the variable has bounded support. Therefore, in this paper we present a zero-inflated regression model based on the quasi-binomial distribution, taking into account the moments and maximum likelihood estimators, and perform a score test to compare the zero-inflated quasi-binomial distribution with the zero-inflated binomial distribution, and the zero-inflated model with the homogeneous model (the model in which covariates are not considered). This analysis is illustrated with two data sets that are well known in the statistical literature and which contain a large number of zeros. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
50. Dr C R Rao’s contributions to the advancement of economic science.
- Author
-
Kumar, T Krishna, Vinod, H D, and Deman, Suresh
- Abstract
In this paper, the authors review Dr C R Rao’s contributions to statistical foundations in economic science, and the importance of his work in advancing econometric modelling and statistical inference in celebration of his birth centenary. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.