673 results on '"62G09"'
Search Results
252. Bootstrap tests for nonparametric comparison of regression curves with dependent errors.
- Author
-
Vilar-Fernández, J. and González-Manteiga, W.
- Published
- 2007
- Full Text
- View/download PDF
253. About tests of the 'simplifying' assumption for conditional copulas
- Author
-
Alexis Derumigny and Jean-David Fermanian
- Subjects
Statistics and Probability ,62G05, 62G08, 62G09 ,Statistics::Theory ,Science (General) ,Computer science ,Copula (linguistics) ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,01 natural sciences ,010104 statistics & probability ,simplifying assumption ,Q1-390 ,0502 economics and business ,FOS: Mathematics ,QA1-939 ,Applied mathematics ,Statistics::Methodology ,conditional copula ,0101 mathematics ,bootstrap ,050205 econometrics ,Statistical hypothesis testing ,Test procedures ,Applied Mathematics ,05 social sciences ,Probability and statistics ,Limiting ,62g09 ,Modeling and Simulation ,62g08 ,62g05 ,Mathematics - Abstract
We discuss the so-called "simplifying assumption" of conditional copulas in a general framework. We introduce several tests of the latter assumption for non- and semiparametric copula models. Some related test procedures based on conditioning subsets instead of point-wise events are proposed. The limiting distribution of such test statistics under the null are approximated by several bootstrap schemes, most of them being new. We prove the validity of a particular semiparametric bootstrap scheme. Some simulations illustrate the relevance of our results., Comment: 48 pages, 7 figures
- Published
- 2017
254. Linearly interpolated FDH efficiency score for nonconvex frontiers
- Author
-
Jeong, Seok-Oh and Simar, Léopold
- Subjects
- *
NUMERICAL analysis , *CONFIDENCE intervals , *STATISTICAL hypothesis testing , *STATISTICAL sampling - Abstract
Abstract: This paper addresses the problem of estimating the monotone boundary of a nonconvex set in a full nonparametric and multivariate setup. This is particularly useful in the context of productivity analysis where the efficient frontier is the locus of optimal production scenarios. Then efficiency scores are defined by the distance of a firm from this efficient boundary. In this setup, the free disposal hull (FDH) estimator has been extensively used due to its flexibility and because it allows nonconvex attainable production sets. However, the nonsmoothness and discontinuities of the FDH is a drawback for conducting inference in finite samples. In particular, it is shown that the bootstrap of the FDH has poor performances and so is not useful in practice. Our estimator, the LFDH, is a linearized version of the FDH, obtained by linear interpolation of appropriate FDH-efficient vertices. It offers a continuous, smooth version of the FDH. We provide an algorithm for computing the estimator, and we establish its asymptotic properties. We also provide an easy way to approximate its asymptotic sampling distribution. The latter could offer bias-corrected estimator and confidence intervals of the efficiency scores. In a Monte Carlo study, we show that these approximations work well even in moderate sample sizes and that our LFDH estimator outperforms, both in bias and in MSE, the original FDH estimator. [Copyright &y& Elsevier]
- Published
- 2006
- Full Text
- View/download PDF
255. Mixed model prediction and small area estimation.
- Author
-
Jiang, Jiming and Lahiri, P.
- Abstract
Over the last three decades, mixed models have been frequently used in a wide range of small area applications. Such models offer great flexibilities in combining information from various sources, and thus are well suited for solving most small area estimation problems. The present article reviews major research developments in the classical inferential approach for linear and generalized linear mixed models that are relevant to different issues concerning small area estimation and related problems. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
256. Approximation to the distribution of LAD estimators for censored regression by random weighting method
- Author
-
Fang, Yixin and Zhao, Lincheng
- Subjects
- *
REGRESSION analysis , *STATISTICS , *DISTRIBUTION (Probability theory) , *LEAST absolute deviations (Statistics) - Abstract
Abstract: Powell (J. Econometrics 25 (1984) 303) considered censored regression model, and established the asymptotic normality of the least absolute deviation (LAD) estimator. But the asymptotic covariance matrices depend on the error density and are therefore difficult to estimate reliably. In the earlier papers, this difficulty may be solved by applying the bootstrap method (see, e.g., Hahn (J. Econometric Theory 11 (1995) 105); Bilias et al. (J. Econometrics 99 (2000) 373). In this paper we propose a random weighting method to approximate the distribution of the LAD estimator. The random weighting method was developed by Rubin (Ann. Statist. 9 (1981) 130), Lo (Ann. Statist. 15 (1987) 360), Tu and Zheng (Chinese J. Appl. Probab. Statist. 3 (1987) 340) with reference to some statistics such as the sample mean. Rao and Zhao (Sankhya 54 (1992) 323) applied random weighting method to approximate asymptotic distribution of M-estimators in regression models. In this paper we extend this method to the censored regression model. [Copyright &y& Elsevier]
- Published
- 2006
- Full Text
- View/download PDF
257. Testing the Equality of Multivariate Distributions Using the Bootstrap and Integrated Empirical Processes.
- Author
-
Jing, Ping and Wang, Jinfang
- Subjects
- *
MULTIVARIATE analysis , *ANALYSIS of variance , *MATHEMATICAL statistics , *STATISTICAL bootstrapping , *DISTRIBUTION (Probability theory) , *STATISTICAL sampling , *LIMIT theorems - Abstract
We introduce some projected integrated empirical processes for testing the equality of two multivariate distributions. The bootstrap is used for determining the approximate critical values. We show that the bootstrap test is consistent. A number-theoretic method is used for efficient computation of the bootstrap critical values. Some simulation results are also given. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
258. Hypothesis Testing Problems in an Unbalanced Longitudinal Ophthalmology Study.
- Author
-
Kim, Jonghyeon
- Subjects
- *
OPHTHALMOLOGY , *STATISTICAL hypothesis testing , *DISTRIBUTION (Probability theory) , *HYPOTHESIS , *ESTIMATION theory , *MULTIVARIATE analysis - Abstract
The analysis of clustered data in a longitudinal ophthalmology study is complicated by correlations between repeatedly measured visual outcomes of paired eyes in a participant and missing observations due to the loss of follow-up. In the present article we consider hypothesis testing problems in an ophthalmology study, where eligible eyes are randomized to two treatments (when two eyes of a participant are eligible, the paired eyes are assigned to different treatments), and vision function outcomes are repeatedly measured over time. A large sample-based nonparametric test statistic and a nonparametric Bootstrap test analog are proposed for testing an interaction effect of two factors and testing an effect of a eye-specific factor within a level of the other person-specific factor on visual function outcomes. Both test statistics allow for missing observations, correlations between repeatedly measured outcomes on individual eyes, and correlations between repeatedly measured outcomes on both eyes of each participant. A simulation study shows that these proposed test statistics maintain nominal significance levels approximately and comparable powers to each other, as well as higher powers than the naive test statistic ignoring correlations between repeated bilateral measurements of both eyes in the same person. For illustration, we apply the proposed test statistics to the changes of visual field defect score in the Advanced Glaucoma Intervention Study. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
259. Asymptotic properties of a two sample randomized test for partially dependent data
- Author
-
Rempala, Grzegorz A. and Looney, Stephen W.
- Subjects
- *
HYPOTHESIS , *STATISTICAL sampling , *MATHEMATICS , *STATISTICS - Abstract
Abstract: We are concerned with an issue of asymptotic validity of a non-parametric randomization test for the two sample location problem under the assumption of partially dependent observations, in which case the validity of the usual permutation -test breaks down. We show that a certain modification of the permutation group used in the randomization procedure yields an unconditional asymptotically valid test in the sense that its probability of Type I error tends to the nominal level with increasing sample sizes. We show that this unconditional test is equivalent to the one based on a linear combination of two- and one-sample -statistics and enjoys some optimal power properties. We also conduct a simulation study comparing our approach with that based on the Fisher''s method of combining -values. Finally, we present an example of application of the test in a medical study on functional status assessment at the end of life. [Copyright &y& Elsevier]
- Published
- 2006
- Full Text
- View/download PDF
260. Quasi-Random Sampling Importance Resampling.
- Author
-
PÉREZ, C. J., MARTÍN, J., RUFO, M. J., and ROJANO, C.
- Subjects
- *
ALGORITHMS , *ALGEBRA , *POLAR forms (Mathematics) , *REASONING , *JUDGMENT (Logic) , *LOGIC - Abstract
We propose two modifications of the sampling/importance resampling (SIR) algorithm introduced by Rubin (1988). They are based on the use of low-discrepancy point sets and sequences. The proposed algorithms yield more representative samples in the sense of the F-discrepancy that turns into better estimations of summary inferences. Although no theoretical proof is provided, an empirical study through a wide range of distributions shows that the proposed approaches improve the SIR algorithm. We include some examples which are illustrative in this sense. [ABSTRACT FROM AUTHOR]
- Published
- 2005
- Full Text
- View/download PDF
261. The deductive phase of statistical analysis via predictive simulations: test, validation and control of a linear model with autocorrelated errors representing a food process
- Author
-
Girard, Philippe and Parent, Eric
- Subjects
- *
ESTIMATION theory , *STATISTICAL sampling , *BAYESIAN analysis , *QUALITY control - Abstract
Statistical analysis consists of two phases: induction for model parameter estimation and deduction to make decisions on the basis of the statistical model. In the Bayesian context, predictive analysis is the key concept to perform the deductive phase. In that context, Monte-Carlo posterior simulations are shown to be extremely valuable tools to achieve for instance model selection and model checking. Example of predictive analysis by simulation is detailed for the linear model with Autocorrelated Errors which has been beforehand estimated by Gibbs sampling. Numerical illustrations are then given for a food process with data collected on line. Special attention is cast on the control of its anticipated behavior under uncertainty within Bayesian decision theory. [Copyright &y& Elsevier]
- Published
- 2004
- Full Text
- View/download PDF
262. On the behavior of Tukey's depth and median under symmetric stable distributions
- Author
-
Chen, Zhiqiang and Tyler, David E.
- Subjects
- *
MULTIVARIATE analysis , *MEDIAN (Mathematics) , *ROBUST statistics , *MATHEMATICAL statistics - Abstract
Some curious properties of Tukey''s depth and Tukey''s multivariate median are revealed by examining their behavior at multivariate distributions possessing independent and identically distributed symmetric stable marginals. In particular, (i) the shape of the contours for Tukey''s depth can be the same for large classes of distributions, (ii) the influence function of a linear combination of the components of Tukey''s median can be uniformly smaller than the influence function of the univariate median for the corresponding linear combination of the multivariate distribution, and (iii) the maximum bias under epsilon contamination for Tukey''s median can be smaller than the maximum bias of the median of some univariate projections of the data. [Copyright &y& Elsevier]
- Published
- 2004
- Full Text
- View/download PDF
263. Bootstrap approximations for Bayesian analysis of Geo/G/1 discrete-time queueing models
- Author
-
Conti, Pier Luigi
- Subjects
- *
ASYNCHRONOUS transfer mode , *TELECOMMUNICATION , *BAYESIAN analysis , *APPROXIMATION theory - Abstract
In this paper we consider a Bayesian nonparametric approach to the analysis of discrete-time queueing models. The main motivation consists in applications to telecommunications, and in particular to asynchronous transfer mode (ATM) systems. Attention is focused on the posterior distribution of the overflow rate. Since the exact distribution of such a quantity is not available in a closed form, an approximation based on “proper” Bayesian bootstrap is proposed, and its properties are studied. Some possible alternatives to proper Bayesian bootstrap are also discussed. Finally, an application to real data is provided. [Copyright &y& Elsevier]
- Published
- 2004
- Full Text
- View/download PDF
264. Jackknifing two-sample statistics
- Author
-
Schechtman, Edna and Wang, A&M</f> University, College Station, TX 77840, USA
- Subjects
- *
JACKKNIFE (Statistics) , *RESAMPLING (Statistics) , *ASYMPTOTES , *STATISTICS - Abstract
In this paper, a new simple method for jackknifing two-sample statistics is proposed. The method is based on a two-step procedure. In the first step, the point estimator is calculated by leaving one
X (orY ) out at a time. At the second step, the point estimator obtained in the first step is further jackknifed, leaving oneY (orX ) out at a time, resulting in a simple formula for the proposed point estimator. It is shown that by using the two-step procedure, the bias of the point estimator is reduced in terms of asymptotic order, fromO(n−1) up toO(n−2) , under certain regularity conditions. This conclusion is also confirmed empirically in terms of finite sample numerical examples via a small-scale simulation study. We also discuss the idea of asymptotic bias to obtain parallel results without imposing some conditions that may be difficult to check or too restrictive in practice. [Copyright &y& Elsevier]- Published
- 2004
- Full Text
- View/download PDF
265. On a new multivariate two-sample test
- Author
-
Baringhaus, L. and Franz, C.
- Subjects
- *
STATISTICS , *SIMULATION methods & models - Abstract
In this paper we propose a new test for the multivariate two-sample problem. The test statistic is the difference of the sum of all the Euclidean interpoint distances between the random variables from the two different samples and one-half of the two corresponding sums of distances of the variables within the same sample. The asymptotic null distribution of the test statistic is derived using the projection method and shown to be the limit of the bootstrap distribution. A simulation study includes the comparison of univariate and multivariate normal distributions for location and dispersion alternatives. For normal location alternatives the new test is shown to have power similar to that of the
t - andT2 -Test. [Copyright &y& Elsevier]- Published
- 2004
- Full Text
- View/download PDF
266. A robustification approach to stability and to uniform particle approximation of nonlinear filters: the example of pseudo-mixing signals
- Author
-
LeGland, François and Oudjane, Nadia
- Subjects
- *
ROBUST control , *MARKOV processes - Abstract
We propose a new approach to study the stability of the optimal filter w.r.t. its initial condition, by introducing a “robust” filter, which is exponentially stable and which approximates the optimal filter uniformly in time. The “robust” filter is obtained here by truncation of the likelihood function, and the robustification result is proved under the assumption that the Markov transition kernel satisfies a pseudo-mixing condition (weaker than the usual mixing condition), and that the observations are “sufficiently good”. This robustification approach allows us to prove also the uniform convergence of several particle approximations to the optimal filter, in some cases of nonergodic signals. [Copyright &y& Elsevier]
- Published
- 2003
- Full Text
- View/download PDF
267. On testing for homogeneity of the covariance matrices.
- Author
-
Xiaoning, Zhang, Ping, Jing, and Xiaoming, Ji
- Abstract
Testing equality of covariance matrix of k populations has long been an interesting issue in statistical inference. To overcome the sparseness of data points in a high-dimensional space and deal with the general cases, we suggest several projection pursuit type statistics. Some results on the limiting distributions of the statistics are obtained. Some properties of Bootstrap approximation are investigated. Furthermore, for computational reasons an approximation which is based on Number theoretic method for the statistics is adopted. Several simulation experiments are performed. [ABSTRACT FROM AUTHOR]
- Published
- 2001
- Full Text
- View/download PDF
268. Bootstrap tests for the equality of distributions.
- Author
-
Ping, Jing
- Abstract
Testing equality of two and k distributions has long been an interesting issue in statistical inference. To overcome the sparseness of data points in high-dimensional space and deal with the general cases, we suggest several projection pursuit type statistics. Some results on the limiting distributions of the statistics are obtained. Some properties of Bootstrap approximation are investigated. Furthermore, for computational reasons an approximation for the statistics the based on Number theoretic method is applied. Several simulation experiments are performed. [ABSTRACT FROM AUTHOR]
- Published
- 2000
- Full Text
- View/download PDF
269. Detecting relevant changes in the mean of nonstationary processes—A mass excess approach
- Author
-
Holger Dette and Weichi Wu
- Subjects
Statistics and Probability ,Series (mathematics) ,Locally stationary process ,change point analysis ,Estimator ,Asymptotic distribution ,Gaussian approximation ,rearrangement estimators ,Type (model theory) ,01 natural sciences ,relevant change points ,Combinatorics ,010104 statistics & probability ,Delta method ,Monotone polygon ,62G08 ,local linear estimation ,62G09 ,Test statistic ,62M10 ,62F05 ,Initial value problem ,0101 mathematics ,Statistics, Probability and Uncertainty ,Mathematics - Abstract
This paper considers the problem of testing if a sequence of means $(\mu_{t})_{t=1,\ldots ,n}$ of a nonstationary time series $(X_{t})_{t=1,\ldots ,n}$ is stable in the sense that the difference of the means $\mu_{1}$ and $\mu_{t}$ between the initial time $t=1$ and any other time is smaller than a given threshold, that is $|\mu_{1}-\mu_{t}|\leq c$ for all $t=1,\ldots ,n$. A test for hypotheses of this type is developed using a bias corrected monotone rearranged local linear estimator and asymptotic normality of the corresponding test statistic is established. As the asymptotic variance depends on the location of the roots of the equation $|\mu_{1}-\mu_{t}|=c$ a new bootstrap procedure is proposed to obtain critical values and its consistency is established. As a consequence we are able to quantitatively describe relevant deviations of a nonstationary sequence from its initial value. The results are illustrated by means of a simulation study and by analyzing data examples.
- Published
- 2019
- Full Text
- View/download PDF
270. Bootstrapping and sample splitting for high-dimensional, assumption-lean inference
- Author
-
Max G'Sell, Alessandro Rinaldo, and Larry Wasserman
- Subjects
Statistics and Probability ,62F40 ,Model selection ,Inference ,assumption-lean ,01 natural sciences ,Measure (mathematics) ,Regression ,010104 statistics & probability ,Bootstrapping (electronics) ,Dimension (vector space) ,Sample splitting ,Robustness (computer science) ,62J05 ,62G09 ,Linear regression ,regression ,0101 mathematics ,Statistics, Probability and Uncertainty ,bootstrap ,Algorithm ,62F35 ,62G20 ,Mathematics - Abstract
Several new methods have been recently proposed for performing valid inference after model selection. An older method is sample splitting: use part of the data for model selection and the rest for inference. In this paper, we revisit sample splitting combined with the bootstrap (or the Normal approximation). We show that this leads to a simple, assumption-lean approach to inference and we establish results on the accuracy of the method. In fact, we find new bounds on the accuracy of the bootstrap and the Normal approximation for general nonlinear parameters with increasing dimension which we then use to assess the accuracy of regression inference. We define new parameters that measure variable importance and that can be inferred with greater accuracy than the usual regression coefficients. Finally, we elucidate an inference-prediction trade-off: splitting increases the accuracy and robustness of inference but can decrease the accuracy of the predictions.
- Published
- 2019
271. Perturbation bootstrap in adaptive Lasso
- Author
-
Debraj Das, Karl Gregory, and Soumendra N. Lahiri
- Subjects
FOS: Computer and information sciences ,Alasso ,Statistics and Probability ,Statistics::Theory ,Studentized range ,Correctness ,Perturbation (astronomy) ,Mathematics - Statistics Theory ,Feature selection ,Statistics Theory (math.ST) ,second-order correctness ,Oracle ,naive perturbation bootstrap ,Methodology (stat.ME) ,62J07 ,62G09 ,Linear regression ,FOS: Mathematics ,Statistics::Methodology ,Applied mathematics ,Statistics - Methodology ,Mathematics ,62E20 ,Estimator ,modified perturbation bootstrap ,oracle ,Sample size determination ,Statistics, Probability and Uncertainty - Abstract
The Adaptive Lasso(Alasso) was proposed by Zou [\textit{J. Amer. Statist. Assoc. \textbf{101} (2006) 1418-1429}] as a modification of the Lasso for the purpose of simultaneous variable selection and estimation of the parameters in a linear regression model. Zou (2006) established that the Alasso estimator is variable-selection consistent as well as asymptotically Normal in the indices corresponding to the nonzero regression coefficients in certain fixed-dimensional settings. In an influential paper, Minnier, Tian and Cai [\textit{J. Amer. Statist. Assoc. \textbf{106} (2011) 1371-1382}] proposed a perturbation bootstrap method and established its distributional consistency for the Alasso estimator in the fixed-dimensional setting. In this paper, however, we show that this (naive) perturbation bootstrap fails to achieve second order correctness in approximating the distribution of the Alasso estimator. We propose a modification to the perturbation bootstrap objective function and show that a suitably studentized version of our modified perturbation bootstrap Alasso estimator achieves second-order correctness even when the dimension of the model is allowed to grow to infinity with the sample size. As a consequence, inferences based on the modified perturbation bootstrap will be more accurate than the inferences based on the oracle Normal approximation. We give simulation studies demonstrating good finite-sample properties of our modified perturbation bootstrap method as well as an illustration of our method on a real data set., Comment: 43 pages, 3 tables, 2 figures
- Published
- 2019
- Full Text
- View/download PDF
272. Bootstrap tuning in Gaussian ordered model selection
- Author
-
Vladimir Spokoiny and Niklas Willrich
- Subjects
Statistics and Probability ,Bootstrap calibration ,Model selection ,Gaussian ,Structure (category theory) ,Estimator ,Noise (electronics) ,Combinatorics ,Smallest accepted ,symbols.namesake ,oracle ,62G09 ,symbols ,62J15 ,propagation condition ,62G05 ,Statistics, Probability and Uncertainty ,Prior information ,Mathematics - Abstract
The paper focuses on the problem of model selection in linear Gaussian regression with unknown possibly inhomogeneous noise. For a given family of linear estimators $\{\widetilde{\boldsymbol{{\theta}}}_{m},m\in\mathscr{M}\}$, ordered by their variance, we offer a new “smallest accepted” approach motivated by Lepski’s device and the multiple testing idea. The procedure selects the smallest model which satisfies the acceptance rule based on comparison with all larger models. The method is completely data-driven and does not use any prior information about the variance structure of the noise: its parameters are adjusted to the underlying possibly heterogeneous noise by the so-called “propagation condition” using a wild bootstrap method. The validity of the bootstrap calibration is proved for finite samples with an explicit error bound. We provide a comprehensive theoretical study of the method, describe in details the set of possible values of the selected model $\widehat{m}\in\mathscr{M}$ and establish some oracle error bounds for the corresponding estimator $\widehat{\boldsymbol{{\theta}}}=\widetilde{\boldsymbol{{\theta}}}_{\widehat{m}}$.
- Published
- 2019
273. Predictive inference with the jackknife+
- Author
-
Emmanuel J. Candès, Rina Foygel Barber, Ryan J. Tibshirani, and Aaditya Ramdas
- Subjects
Statistics and Probability ,FOS: Computer and information sciences ,Statistics::Theory ,Stability (learning theory) ,leave-one-out ,02 engineering and technology ,Interval (mathematics) ,cross-validation ,01 natural sciences ,Cross-validation ,Methodology (stat.ME) ,010104 statistics & probability ,Predictive inference ,62G08 ,62G09 ,0202 electrical engineering, electronic engineering, information engineering ,Applied mathematics ,Statistics::Methodology ,Distribution-free ,0101 mathematics ,Statistics - Methodology ,Mathematics ,62F40 ,conformal inference ,stability ,Confidence interval ,jackknife ,Data point ,020201 artificial intelligence & image processing ,Statistics, Probability and Uncertainty ,Jackknife resampling ,Quantile - Abstract
This paper introduces the jackknife+, which is a novel method for constructing predictive confidence intervals. Whereas the jackknife outputs an interval centered at the predicted response of a test point, with the width of the interval determined by the quantiles of leave-one-out residuals, the jackknife+ also uses the leave-one-out predictions at the test point to account for the variability in the fitted regression function. Assuming exchangeable training samples, we prove that this crucial modification permits rigorous coverage guarantees regardless of the distribution of the data points, for any algorithm that treats the training points symmetrically. Such guarantees are not possible for the original jackknife and we demonstrate examples where the coverage rate may actually vanish. Our theoretical and empirical analysis reveals that the jackknife and the jackknife+ intervals achieve nearly exact coverage and have similar lengths whenever the fitting algorithm obeys some form of stability. Further, we extend the jackknife+ to $K$-fold cross validation and similarly establish rigorous coverage properties. Our methods are related to cross-conformal prediction proposed by Vovk (Ann. Math. Artif. Intell. 74 (2015) 9–28) and we discuss connections.
- Published
- 2019
274. Convolved subsampling estimation with applications to block bootstrap
- Author
-
Daniel J. Nordman, Dimitris N. Politis, and Johannes Tewes
- Subjects
Statistics and Probability ,Mathematics - Statistics Theory ,Sample (statistics) ,Statistics Theory (math.ST) ,Astrophysics::Cosmology and Extragalactic Astrophysics ,01 natural sciences ,Convolution ,010104 statistics & probability ,Consistency (statistics) ,62J05 ,Resampling ,62G09 ,FOS: Mathematics ,mixing ,0101 mathematics ,Statistic ,62G20 ,Mathematics ,Block (data storage) ,nonstationary ,Estimator ,humanities ,moving blocks ,Sampling distribution ,62M10 ,Primary 62G09, secondary 62G20, 62J05, 62M10 ,Statistics, Probability and Uncertainty ,Algorithm - Abstract
The block bootstrap approximates sampling distributions from dependent data by resampling data blocks. A fundamental problem is establishing its consistency for the distribution of a sample mean, as a prototypical statistic. We use a structural relationship with subsampling to characterize the bootstrap in a new and general manner. While subsampling and block bootstrap differ, the block bootstrap distribution of a sample mean equals that of a $k$-fold self-convolution of a subsampling distribution. Motivated by this, we provide simple necessary and sufficient conditions for a convolved subsampling estimator to produce a normal limit that matches the target of bootstrap estimation. These conditions may be linked to consistency properties of an original subsampling distribution, which are often obtainable under minimal assumptions. Through several examples, the results are shown to validate the block bootstrap for means under significantly weakened assumptions in many existing (and some new) dependence settings, which also addresses a standing conjecture of Politis, Romano and Wolf(1999). Beyond sample means, the convolved subsampling estimator may not match the block bootstrap, but instead provides a hybrid-resampling estimator of interest in its own right. For general statistics with normal limits, results also establish the consistency of convolved subsampling under minimal dependence conditions, including non-stationarity., Comment: 42 pages
- Published
- 2019
275. Goodness-of-fit tests for the functional linear model based on randomly projected empirical processes
- Author
-
Wenceslao González-Manteiga, Juan A. Cuesta-Albertos, Manuel Febrero-Bande, Eduardo García-Portugués, Ministerio de Economía, Industria y Competitividad (España), Universidade de Santiago de Compostela. Departamento de Estatística, Análise Matemática e Optimización, and Universidad de Cantabria
- Subjects
FOS: Computer and information sciences ,Checks ,Statistics and Probability ,goodness-of-fit ,Goodness-of-fit ,Estadística ,Methodology (stat.ME) ,Functional principal components ,Goodness of fit ,62J05 ,62G09 ,Covariate ,Applied mathematics ,Almost surely ,functional principal components ,functional data ,Statistics - Methodology ,Form ,Empirical process ,Mathematics ,Statistical hypothesis testing ,Weak convergence ,Linear model ,Functional linear model ,62G10, 62J05, 62G09 ,Functional data ,random projections ,Regression ,Random projections ,Statistics, Probability and Uncertainty ,functional linear model ,62G10 ,Curse of dimensionality - Abstract
We consider marked empirical processes indexed by a randomly projected functional covariate to construct goodness-of-fit tests for the functional linear model with scalar response. The test statistics are built from continuous functionals over the projected process, resulting in computationally efficient tests that exhibit root-n convergence rates and circumvent the curse of dimensionality. The weak convergence of the empirical process is obtained conditionally on a random direction, whilst the almost surely equivalence between the testing for significance expressed on the original and on the projected functional covariate is proved. The computation of the test in practice involves calibration by wild bootstrap resampling and the combination of several p-values, arising from different projections, by means of the false discovery rate method. The finite sample properties of the tests are illustrated in a simulation study for a variety of linear models, underlying processes, and alternatives. The software provided implements the tests and allows the replication of simulations and data applications., Paper: 23 pages, 4 figures, 1 table. Supplementary material: 17 pages, 4 figures, 3 tables
- Published
- 2019
- Full Text
- View/download PDF
276. Permutation $p$-value approximation via generalized Stolarsky invariance
- Author
-
Qingyuan Zhao, Hera Yu He, Art B. Owen, and Kinjal Basu
- Subjects
Statistics and Probability ,gene sets ,Generalization ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,01 natural sciences ,Square (algebra) ,Set (abstract data type) ,010104 statistics & probability ,Permutation ,Approximation error ,62G09 ,hypothesis testing ,FOS: Mathematics ,Mathematics - Combinatorics ,Applied mathematics ,p-value ,0101 mathematics ,Discrepancy ,11K38 ,Mathematics ,Statistical hypothesis testing ,quasi-Monte Carlo ,Invariance principle ,Combinatorics (math.CO) ,Statistics, Probability and Uncertainty ,62G10 - Abstract
It is common for genomic data analysis to use $p$-values from a large number of permutation tests. The multiplicity of tests may require very tiny $p$-values in order to reject any null hypotheses and the common practice of using randomly sampled permutations then becomes very expensive. We propose an inexpensive approximation to $p$-values for two sample linear test statistics, derived from Stolarsky’s invariance principle. The method creates a geometrically derived reference set of approximate $p$-values for each hypothesis. The average of that set is used as a point estimate $\hat{p}$ and our generalization of the invariance principle allows us to compute the variance of the $p$-values in that set. We find that in cases where the point estimate is small, the variance is a modest multiple of the square of that point estimate, yielding a relative error property similar to that of saddlepoint approximations. On a Parkinson’s disease data set, the new approximation is faster and more accurate than the saddlepoint approximation. We also obtain a simple probabilistic explanation of Stolarsky’s invariance principle.
- Published
- 2019
- Full Text
- View/download PDF
277. A statistical test of isomorphism between metric-measure spaces using the distance-to-a-measure signature
- Author
-
Claire Brécheteau, Understanding the Shape of Data (DATASHAPE), Inria Sophia Antipolis - Méditerranée (CRISAM), Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Inria Saclay - Ile de France, Institut National de Recherche en Informatique et en Automatique (Inria), Laboratoire de Mathématiques d'Orsay (LMO), Université Paris-Sud - Paris 11 (UP11)-Centre National de la Recherche Scientifique (CNRS), ANR-13-BS01-0008,TopData,Analyse Topologique des Données : Méthodes Statistiques et Estimation(2013), and European Project: 339025,EC:FP7:ERC,ERC-2013-ADG,GUDHI(2014)
- Subjects
Statistics and Probability ,Space (mathematics) ,(Gromov)-Wasserstein distances ,[INFO.INFO-CG]Computer Science [cs]/Computational Geometry [cs.CG] ,01 natural sciences ,Measure (mathematics) ,Square (algebra) ,010104 statistics & probability ,distance to a measure ,[MATH.MATH-ST]Mathematics [math]/Statistics [math.ST] ,62G09 ,Test statistic ,0101 mathematics ,Statistical hypothesis testing ,Mathematics ,Metric-measure spaces ,Discrete mathematics ,Distance-to-measure ,010102 general mathematics ,Function (mathematics) ,[STAT.TH]Statistics [stat]/Statistics Theory [stat.TH] ,[MATH.MATH-PR]Mathematics [math]/Probability [math.PR] ,Metric (mathematics) ,Isomorphism ,Statistics, Probability and Uncertainty ,Statistical test ,Subsampling ,62G10 - Abstract
MSC: Primary 62G10; secondary 62G09; International audience; We introduce the notion of DTM-signature, a measure on R that can be associated to any metric-measure space. This signature is based on the function distance to a measure (DTM) introduced in 2009 by Chazal, Cohen-Steiner and Mérigot. It leads to a pseudo-metric between metric-measure spaces, that is bounded above by the Gromov-Wasserstein distance. This pseudo-metric is used to build a statistical test of isomorphism between two metric-measure spaces, from the observation of two N-samples. The test is based on subsampling methods and comes with theoretical guarantees. It is proven to be of the correct level asymptotically. Also, when the measures are supported on compact subsets of R^d, rates of convergence are derived for the L1-Wasserstein distance between the distribution of the test statistic and its subsampling approximation. These rates depend on some parameter \rho> 1. In addition, we prove that the power is bounded above by exp(−CN^(1/ \rho)), with C proportional to the square of the aforementioned pseudo-metric between the metric-measure spaces. Under some geometrical assumptions, we also derive lower bounds for this pseudo-metric. An algorithm is proposed for the implementation of this statistical test, and its performance is compared to the performance of other methods through numerical experiments.
- Published
- 2019
- Full Text
- View/download PDF
278. Estimating transformation function
- Author
-
Dimitris N. Politis, Yunyi Zhang, Zexin Pan, and Jiazheng Liu
- Subjects
Statistics and Probability ,Transformation function ,Distribution (number theory) ,quantile process ,Estimator ,Asymptotic distribution ,Monotonic function ,Function (mathematics) ,Combinatorics ,Convergence of random variables ,Resampling ,62G09 ,62G05 ,Statistics, Probability and Uncertainty ,resampling method ,Random variable ,Mathematics ,62G10 - Abstract
In this paper, we propose an estimator for $g(x)$ under the model $Y_{i}=g(Z_{i}),\ i=1,2,...,n$ where $Z_{i},\ i=1,2,...$ are random variables with known distribution but unknown observed values, $Y_{i},\ i=1,2,...$ are observed data and $g(x)$ is an unknown strictly monotonically increasing function (we call $g(x)$ transformation function). We prove the almost sure convergence of the estimator and construct confidence intervals and bands when $Z_{i},i=1,2,...$ are i.i.d data based on their asymptotic distribution. Corresponding case when $Z_{i}$ being linear process is handled by resampling method. We also design the hypothesis test regarding whether $g(x)$ equals an expected transformation function or not. The finite sample performance is evaluated by applying the method to simulated data and an urban waste water treatment plant’s dataset.
- Published
- 2019
279. Nonparametric estimation of the lifetime and disease onset distributions for a survival-sacrifice model
- Author
-
Antonio Eduardo Gomes, Piet Groeneboom, and Jon A. Wellner
- Subjects
Statistics and Probability ,smooth functionals ,01 natural sciences ,Volterra integral equation ,Combinatorics ,010104 statistics & probability ,symbols.namesake ,Joint probability distribution ,62G09 ,62N01 ,0502 economics and business ,Statistics ,0101 mathematics ,EM algorithm ,050205 econometrics ,Cube root ,Mathematics ,Quadratic growth ,Conjecture ,survival sacrifice model ,Estimation theory ,self-consistency equation ,05 social sciences ,Estimator ,primal-dual interior point algorithm ,symbols ,MLE ,Statistics, Probability and Uncertainty ,Random variable - Abstract
In carcinogenicity experiments with animals where the tumor is not palpable it is common to observe only the time of death of the animal, the cause of death (the tumor or another independent cause, as sacrifice) and whether the tumor was present at the time of death. These last two indicator variables are evaluated after an autopsy. Defining the non-negative variables T1 (time of tumor onset), T2 (time of death from the tumor) and C (time of death from an unrelated cause), we observe (Y,Δ1,Δ2), where Y = min{T2,C},Δ1 =1 {T1≤C}, and Δ2 =1 {T2≤C}. The random variables T1 and T2 are independent of C and have a joint distribution such that P(T1 ≤ T2) = 1. Some authors call this model a “survival-sacrifice model”. [20] (generally to be denoted by LJP (1997)) proposed a Weighted Least Squares estimator for F1 (the marginal distribution function of T1), using the Kaplan-Meier estimator of F2 (the marginal distribution function of T2). The authors claimed that their estimator is more efficient than the MLE (maximum likelihood estimator) of F1 and that the Kaplan-Meier estimator is more efficient than the MLE of F2. However, we show that the MLE of F1 was not computed correctly, and that the (claimed) MLE estimate of F1 is even undefined in the case of active constraints. In our simulation study we used a primal-dual interior point algorithm to obtain the true MLE of F1. The results showed a better performance of the MLE of F1 over the weighted least squares estimator in LJP (1997) for points where F1 is close to F2. Moreover, application to the model, used in the simulation study of LJP (1997), showed smaller variances of the MLE estimators of the first and second moments for both F1 and F2, and sample sizes from 100 up to 5000, in comparison to the estimates, based on the weighted least squares estimator for F1, proposed in LJP (1997), and the Kaplan-Meier estimator for F2. R scripts are provided for computing the estimates either with the primal-dual interior point method or by the EM algorithm. In spite of the long history of the model in the biometrics literature (since about 1982), basic properties of the real maximum likelihood estimator (MLE) were still unknown. We give necessary and sfficient conditions for the MLE (Theorem 3.1), as an element of a cone, where the number of generators of the cone increases quadratically with sample size. From this and a self-consistency equation, turned into a Volterra integral equation, we derive the consistency of the MLE (Theorem 4.1). We conjecture that (under some natural conditions) one can extend the methods, used to prove consistency, to proving that the MLE is √n consistent for F2 and cube root n convergent for F1, but this has presently not yet been proved.
- Published
- 2019
- Full Text
- View/download PDF
280. Consistent nonparametric change point detection combining CUSUM and marked empirical processes
- Author
-
Natalie Neumeyer and Maria Mohr
- Subjects
Statistics and Probability ,Heteroscedasticity ,Statistics::Theory ,CUSUM ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,sequential empirical process ,symbols.namesake ,change point detection ,62G08 ,62G09 ,kernel estimation ,FOS: Mathematics ,Applied mathematics ,Statistics::Methodology ,Gaussian process ,distribution-free test ,Empirical process ,Mathematics ,Weak convergence ,Univariate ,Nonparametric statistics ,Primary 62M10, Secondary 62G08, 62G09, 62G10 ,cumulative sums ,Bootstrap ,Nonparametric regression ,heteroscedasticity ,nonparametric regression ,symbols ,62M10 ,Statistics, Probability and Uncertainty ,62G10 - Abstract
A weakly dependent time series regression model with multivariate covariates and univariate observations is considered, for which we develop a procedure to detect whether the nonparametric conditional mean function is stable in time against change point alternatives. Our proposal is based on a modified CUSUM type test procedure, which uses a sequential marked empirical process of residuals. We show weak convergence of the considered process to a centered Gaussian process under the null hypothesis of no change in the mean function and a stationarity assumption. This requires some sophisticated arguments for sequential empirical processes of weakly dependent variables. As a consequence we obtain convergence of Kolmogorov-Smirnov and Cram\'er-von Mises type test statistics. The proposed procedure acquires a very simple limiting distribution and nice consistency properties, features from which related tests are lacking. We moreover suggest a bootstrap version of the procedure and discuss its applicability in the case of unstable variances., Comment: 35 pages (including 5 pages of supplementary material), 4 figures, 2 tables
- Published
- 2019
- Full Text
- View/download PDF
281. Coarse-to-fine multiple testing strategies
- Author
-
Donald Geman, Kamel Lahouel, and Laurent Younes
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,hierarchical testing ,nutritional and metabolic diseases ,Methodology (stat.ME) ,permutation tests ,symbols.namesake ,Bonferroni correction ,Survivorship bias ,FWER ,62G09 ,Multiple comparisons problem ,Statistics ,symbols ,Multiple testing ,Fraction (mathematics) ,Statistics, Probability and Uncertainty ,Null hypothesis ,Random variable ,Gaussian network model ,Statistics - Methodology ,62G10 ,Mathematics ,Statistical hypothesis testing - Abstract
We analyze control of the familywise error rate (FWER) in a multiple testing scenario with a great many null hypotheses about the distribution of a high-dimensional random variable among which only a very small fraction are false, or "active". In order to improve power relative to conservative Bonferroni bounds, we explore a coarse-to-fine procedure adapted to a situation in which tests are partitioned into subsets, or "cells", and active hypotheses tend to cluster within cells. We develop procedures for a standard linear model with Gaussian data and a non-parametric case based on generalized permutation testing, and demonstrate considerably higher power than Bonferroni estimates at the same FWER when the active hypotheses do cluster. The main technical difficulty arises from the correlation between the test statistics at the individual and cell levels, which increases the likelihood of a hypothesis being falsely discovered when the cell that contains it is falsely discovered (survivorship bias). This requires sharp estimates of certain quadrant probabilities when a cell is inactive.
- Published
- 2019
- Full Text
- View/download PDF
282. Refinements of the Kiefer-Wolfowitz theorem and a test of concavity
- Author
-
Zheng Fang
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Mathematics - Statistics Theory ,Monotonic function ,Statistics Theory (math.ST) ,test of concavity/monotonicity ,Convexity ,Methodology (stat.ME) ,Combinatorics ,62G09 ,FOS: Mathematics ,62G07 ,62G05 ,Almost surely ,Statistics - Methodology ,62G20 ,Grenander estimator ,Mathematics ,Pointwise ,Null (mathematics) ,Estimator ,Kiefer-Wolfowitz theorem ,Empirical distribution function ,Bounded function ,Statistics, Probability and Uncertainty ,least concave majorant ,62G10 - Abstract
This paper studies estimation of and inference on a distribution function $F$ that is concave on the nonnegative half line and admits a density function $f$ with potentially unbounded support. When $F$ is strictly concave, we show that the supremum distance between the Grenander distribution estimator and the empirical distribution may still be of order $O(n^{-2/3}(\log n)^{2/3})$ almost surely, which reduces to an existing result of Kiefer and Wolfowitz when $f$ has bounded support. We further refine this result by allowing $F$ to be not strictly concave or even non-concave and instead requiring it be "asymptotically" strictly concave. Building on these results, we then develop a test of concavity of $F$ or equivalently monotonicity of $f$, which is shown to have asymptotically pointwise level control under the entire null as well as consistency under any fixed alternative. In fact, we show that our test has local size control and nontrivial local power against any local alternatives that do not approach the null too fast, which may be of interest given the irregularity of the problem. Extensions to settings involving testing concavity/convexity/monotonicity are discussed., Forthcoming in Electronic Journal of Statistics. Compared to the journal version, the difference is that this version contains additional simulation results, collected in Appendix C
- Published
- 2019
- Full Text
- View/download PDF
283. Bootstrapping the empirical distribution of a stationary process with change-point
- Author
-
Farid El Ktaibi and B. Gail Ivanoff
- Subjects
Statistics and Probability ,62G30 ,Stationary process ,Time series ,moving block bootstrap ,sequential empirical process ,01 natural sciences ,010104 statistics & probability ,Mixing (mathematics) ,62G09 ,0502 economics and business ,Applied mathematics ,0101 mathematics ,causal linear process ,050205 econometrics ,Mathematics ,Series (mathematics) ,05 social sciences ,Empirical distribution function ,Moment (mathematics) ,Distribution (mathematics) ,Bootstrapping (electronics) ,60F17 ,62M10 ,Statistics, Probability and Uncertainty ,Marginal distribution ,change-point ,62G10 - Abstract
When detecting a change-point in the marginal distribution of a stationary time series, bootstrap techniques are required to determine critical values for the tests when the pre-change distribution is unknown. In this paper, we propose a sequential moving block bootstrap and demonstrate its validity under a converging alternative. Furthermore, we demonstrate that power is still achieved by the bootstrap under a non-converging alternative. We follow the approach taken by Peligrad in [14], and avoid assumptions of mixing, association or near epoch dependence. These results are applied to a linear process and are shown to be valid under very mild conditions on the existence of any moment of the innovations and a corresponding condition of summability of the coefficients.
- Published
- 2019
284. New insights on permutation approach for hypothesis testing on functional data.
- Author
-
Corain, Livio, Melas, Viatcheslav, Pepelyshev, Andrey, and Salmaso, Luigi
- Abstract
The permutation approach for testing the equality of distributions and thereby comparing two populations of functional data has recently received increasing attention thanks to the flexibility of permutation tests to handle complex testing problems. The purpose of this work is to present some new insights in the context of nonparametric inference on functional data using the permutation approach, more specifically we formally show the equivalence of some permutation procedures proposed in the literature and we suggest the use of the permutation and combination-based approach within the basis function approximation layout. Validation of theoretical results is shown by simulation studies. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
285. Bootstrap techniques for measures of center for three-dimensional rotation data
- Author
-
L. Katie Will and Melissa A. Bingham
- Subjects
General Mathematics ,mean matrix ,Spatial average ,spatial average ,02 engineering and technology ,021001 nanoscience & nanotechnology ,Geodesy ,Rotation ,01 natural sciences ,010104 statistics & probability ,62G09 ,62P30 ,Center (algebra and category theory) ,0101 mathematics ,bootstrap ,0210 nano-technology ,3-D rotations ,Mathematics - Abstract
Bootstrapping is a nonparametric statistical technique that can be used to estimate the sampling distribution of a statistic of interest. This paper focuses on implementation of bootstrapping in a new setting, where the data of interest are 3-dimensional rotations. Two measures of center, the mean rotation and spatial average, are considered, and bootstrap confidence regions for these measures are proposed. The developed techniques are then used in a materials science application, where precision is explored for measurements of crystal orientations obtained via electron backscatter diffraction.
- Published
- 2016
- Full Text
- View/download PDF
286. Pointwise and uniform convergence of kernel density estimators using random bandwidths.
- Author
-
Dutta, Santanu and Goswami, Alok
- Subjects
- *
STOCHASTIC convergence , *KERNEL (Mathematics) , *RANDOM variables , *BANDWIDTHS , *MIXING , *MATHEMATICAL analysis , *NUMERICAL analysis - Abstract
Abstract: We obtain the rates of pointwise and uniform convergence of kernel density estimators using random bandwidths under i.i.d. as well as strongly mixing dependence assumptions. Pointwise rates are faster and not affected by the tail of the density. [Copyright &y& Elsevier]
- Published
- 2013
- Full Text
- View/download PDF
287. Normal limits, nonnormal limits, and the bootstrap for quantiles of dependent data
- Author
-
Sharipov, Olimjon Sh. and Wendler, Martin
- Subjects
- *
LIMITS (Mathematics) , *STATISTICAL bootstrapping , *QUANTILES , *DATA analysis , *MATHEMATICS theorems , *MATHEMATICAL analysis - Abstract
Abstract: We will show under very weak conditions on differentiability and dependence that the central limit theorem for quantiles holds and that the block bootstrap is weakly consistent. Under slightly stronger conditions, the bootstrap is strongly consistent. Without the differentiability condition, quantiles might have a nonnormal asymptotic distribution and the bootstrap might fail. [Copyright &y& Elsevier]
- Published
- 2013
- Full Text
- View/download PDF
288. Density estimation using bootstrap bandwidth selector
- Author
-
Bose, Arup and Dutta, Santanu
- Subjects
- *
DISTRIBUTION (Probability theory) , *STATISTICAL bootstrapping , *BANDWIDTHS , *STATISTICAL smoothing , *ASYMPTOTIC expansions , *PROBABILITY theory - Abstract
Abstract: Smoothing methods for density estimators struggle when the shape of the reference density differs markedly from the actual density. We propose a bootstrap bandwidth selector where no reference distribution is used. It performs reliably in difficult cases and asymptotically outperforms well known automatic bandwidths. [Copyright &y& Elsevier]
- Published
- 2013
- Full Text
- View/download PDF
289. Bayesian Neural Networks as a pricing model to reduce information costs in peer-to-peer online marketplaces
- Author
-
Universitat de Barcelona. Departament d'Econometria, Estadística i Economia Espanyola, Universitat Politècnica de Catalunya. Departament d'Estadística i Investigació Operativa, Puig Oriol, Xavier, Torra Porras, Salvador, Susagna Holgado, Marc, Universitat de Barcelona. Departament d'Econometria, Estadística i Economia Espanyola, Universitat Politècnica de Catalunya. Departament d'Estadística i Investigació Operativa, Puig Oriol, Xavier, Torra Porras, Salvador, and Susagna Holgado, Marc
- Abstract
The main purpose of this thesis is to disclose the potential of exploiting the synergies between Statistical Science and Machine Learning. In particular, we propose a specific feed-forward Bayesian Neural Network (BNN) as a parametric statistical model able to both yield better punctual predictions than linear models and handle uncertainty through more grounded intervals than the ones offered by bootstrapping conventional Neural Networks. On top of proposing a complete methodology (based on DoE for architecture selection and MCMC to conduct inference) to apply BNNs in real cases, we analyze, using theoretical arguments from Microeconomics, the positive effect on society that it would have to use BNNs as pricing model for peer-to-peer online marketplaces and, moreover, we implement them for the case of Airbnb in Barcelona
- Published
- 2018
290. Bootstrap methods for stationary functional time series
- Author
-
Shang, Han Lin
- Published
- 2016
- Full Text
- View/download PDF
291. Comments on: A random forest guided tour
- Author
-
Hooker, Giles and Mentch, Lucas
- Published
- 2016
- Full Text
- View/download PDF
292. Comments on: A random forest guided tour
- Author
-
Arlot, Sylvain and Genuer, Robin
- Published
- 2016
- Full Text
- View/download PDF
293. Goodness-of-fit tests in semiparametric transformation models
- Author
-
Colling, Benjamin and Van Keilegom, Ingrid
- Published
- 2016
- Full Text
- View/download PDF
294. Stochastically optimal bootstrap sample size for shrinkage-type statistics
- Author
-
Wei, Bei, Lee, Stephen M. S., and Wu, Xiyuan
- Published
- 2016
- Full Text
- View/download PDF
295. On weak dependence conditions: The case of discrete valued processes
- Author
-
Doukhan, Paul, Fokianos, Konstantinos, and Li, Xiaoyin
- Subjects
- *
DEPENDENCE (Statistics) , *MARKOV processes , *GEOMETRIC analysis , *MATHEMATICAL models , *MATHEMATICAL statistics , *CONTRACTIONS (Topology) - Abstract
Abstract: We investigate the relationship between weak dependence and mixing for discrete valued processes. We show that weak dependence implies mixing conditions under natural assumptions. The results specialize to the case of Markov processes. Several examples of integer valued processes are discussed and their weak dependence properties are investigated by means of a contraction principle. In fact, we show the stronger result that the mixing coefficients for infinite memory weakly dependent models decay geometrically fast. Hence, all integer values models that we consider have weak dependence coefficients which decay geometrically fast. [Copyright &y& Elsevier]
- Published
- 2012
- Full Text
- View/download PDF
296. Multiple imputation in principal component analysis.
- Author
-
Josse, Julie, Pagès, Jérôme, and Husson, François
- Abstract
The available methods to handle missing values in principal component analysis only provide point estimates of the parameters (axes and components) and estimates of the missing values. To take into account the variability due to missing values a multiple imputation method is proposed. First a method to generate multiple imputed data sets from a principal component analysis model is defined. Then, two ways to visualize the uncertainty due to missing values onto the principal component analysis results are described. The first one consists in projecting the imputed data sets onto a reference configuration as supplementary elements to assess the stability of the individuals (respectively of the variables). The second one consists in performing a principal component analysis on each imputed data set and fitting each obtained configuration onto the reference one with Procrustes rotation. The latter strategy allows to assess the variability of the principal component analysis parameters induced by the missing values. The methodology is then evaluated from a real data set. [ABSTRACT FROM AUTHOR]
- Published
- 2011
- Full Text
- View/download PDF
297. Tests of covariance matrix by using projection pursuit and bootstrap method.
- Author
-
Ping, Jing
- Abstract
Testing equality of covariance matrix has long been an interesting issue in statistics inference. To overcome the sparseness of data points in high-dimensional space and deal with the general cases, the author suggests several projection pursuit type statistics. Some results on the limiting distributions of the statistics are obtained. Some properties of bootstrap approximation are investigated. Furthermore, for computational reasons an approximation for the statistics based on number-theoretic method is applied. Several simulation experiments are performed. [ABSTRACT FROM AUTHOR]
- Published
- 1998
- Full Text
- View/download PDF
298. Bootstrap based tests for generalized negative binomial distribution.
- Author
-
Famoye, F.
- Abstract
Goodness of fit test statistics based on the empirical distribution function (EDF) are considered for the generalized negative binomial distribution. The small sample levels of the tests are found to be very close to the nominal significance levels. For small sample sizes, the tests are compared with respect to their simulated power of detecting some alternative hypotheses against a null hypothesis of generalized negative binomial distribution. The discrete Anderson—Darling test is the most powerful among the EDF tests. Two numerical examples are used to illustrate the application of the goodness of fit tests. [ABSTRACT FROM AUTHOR]
- Published
- 1998
- Full Text
- View/download PDF
299. Exact testing with random permutations
- Author
-
Jesse Hemerik and Jelle J. Goeman
- Subjects
0301 basic medicine ,Statistics and Probability ,Conditional monte carlo ,Resampling ,Computer science ,Existential quantification ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,01 natural sciences ,010104 statistics & probability ,03 medical and health sciences ,Permutation ,62G09 ,FOS: Mathematics ,Permutation test ,0101 mathematics ,Discrete mathematics ,Original Paper ,Mathematics::Combinatorics ,Nonparametric statistics ,Permutation group ,030104 developmental biology ,Nonparametric test ,Multiple comparisons problem ,Statistics, Probability and Uncertainty ,62G10 - Abstract
When permutation methods are used in practice, often a limited number of random permutations are used to decrease the computational burden. However, most theoretical literature assumes that the whole permutation group is used, and methods based on random permutations tend to be seen as approximate. There exists a very limited amount of literature on exact testing with random permutations, and only recently a thorough proof of exactness was given. In this paper, we provide an alternative proof, viewing the test as a “conditional Monte Carlo test” as it has been called in the literature. We also provide extensions of the result. Importantly, our results can be used to prove properties of various multiple testing procedures based on random permutations.
- Published
- 2018
300. Sieve bootstrap for functional time series
- Author
-
Paparoditis, Efstathios and Paparoditis, Efstathios [0000-0003-1958-781X]
- Subjects
Statistics and Probability ,Statistics::Theory ,Mathematics - Statistics Theory ,Context (language use) ,Statistics Theory (math.ST) ,01 natural sciences ,010104 statistics & probability ,symbols.namesake ,Karhunen–Loève expansion ,62G09 ,0502 economics and business ,FOS: Mathematics ,Applied mathematics ,0101 mathematics ,Finite set ,Fourier series ,principal components ,050205 econometrics ,Mathematics ,Central limit theorem ,Series (mathematics) ,05 social sciences ,spectral density operator ,Bootstrap ,Fourier transform ,62M15 ,Autoregressive model ,Principal component analysis ,symbols ,62M10 ,Statistics, Probability and Uncertainty - Abstract
A bootstrap procedure for functional time series is proposed which exploits a general vector autoregressive representation of the time series of Fourier coefficients appearing in the Karhunen–Loève expansion of the functional process. A double sieve-type bootstrap method is developed, which avoids the estimation of process operators and generates functional pseudo-time series that appropriately mimics the dependence structure of the functional time series at hand. The method uses a finite set of functional principal components to capture the essential driving parts of the infinite dimensional process and a finite order vector autoregressive process to imitate the temporal dependence structure of the corresponding vector time series of Fourier coefficients. By allowing the number of functional principal components as well as the autoregressive order used to increase to infinity (at some appropriate rate) as the sample size increases, consistency of the functional sieve bootstrap can be established. We demonstrate this by proving a basic bootstrap central limit theorem for functional finite Fourier transforms and by establishing bootstrap validity in the context of a fully functional testing problem. A novel procedure to select the number of functional principal components is introduced while simulations illustrate the good finite sample performance of the new bootstrap method proposed. 46 6B 3510 3538
- Published
- 2018
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.