308 results
Search Results
2. Cross-Validation, Risk Estimation, and Model Selection: Comment on a Paper by Rosset and Tibshirani
- Author
-
Stefan Wager
- Subjects
Statistics and Probability ,Estimation ,Computer science ,business.industry ,Model selection ,05 social sciences ,Machine learning ,computer.software_genre ,01 natural sciences ,Cross-validation ,Task (project management) ,010104 statistics & probability ,0502 economics and business ,Range (statistics) ,Artificial intelligence ,0101 mathematics ,Statistics, Probability and Uncertainty ,business ,computer ,050205 econometrics - Abstract
How best to estimate the accuracy of a predictive rule has been a longstanding question in statistics. Approaches to this task range from simple methods like Mallow’s Cp to algorithmic techniques l...
- Published
- 2020
3. Review Papers: Modeling Capture, Recapture, and Removal Statistics for Estimation of Demographic Parameters for Fish and Wildlife Populations: Past, Present, and Future
- Author
-
Kenneth H. Pollock
- Subjects
Statistics and Probability ,Estimation ,Mark and recapture ,Computer science ,Statistics ,Wildlife ,Econometrics ,Statistics, Probability and Uncertainty ,Census ,Catch per unit effort ,Robustness (economics) ,Speculation ,Bayesian inference - Abstract
In this article I review the modeling of capture, recapture, and removal statistics for the purpose of estimating demographic parameters of fish and wildlife populations. Topics considered include capture-recapture models, band or tag return models, removal and catch per unit effort models, selective removal or change-in-ratio models, radio-tagging survival models, and nest survival models. The purpose is to present important concepts in a general manner for the benefit of a wide audience of statisticians. I will not attempt to be comprehensive, and I indulge in speculation about future directions. I indicate the importance of different statistical tools to this subject, such as Bayesian inference, “boot strapping,” robustness studies, goodness-of-fit tests. I also emphasize connections to other application areas of statistics. Capture-recapture methods, for example, are being considered for estimation of a variety of elusive human populations, such as the homeless and people missed in the census...
- Published
- 1991
4. Review Papers: Recent Developments in Nonparametric Density Estimation
- Author
-
Alan Julian Izenman
- Subjects
Statistics and Probability ,Statistics::Theory ,Restricted maximum likelihood ,Nonparametric statistics ,Estimator ,Density estimation ,Nonparametric regression ,Statistics::Machine Learning ,Histogram ,Kernel (statistics) ,Pattern recognition (psychology) ,Statistics ,Econometrics ,Statistics::Methodology ,Statistics, Probability and Uncertainty ,Mathematics - Abstract
Advances in computation and the fast and cheap computational facilities now available to statisticians have had a significant impact upon statistical research, and especially the development of nonparametric data analysis procedures. In particular, theoretical and applied research on nonparametric density estimation has had a noticeable influence on related topics, such as nonparametric regression, nonparametric discrimination, and nonparametric pattern recognition. This article reviews recent developments in nonparametric density estimation and includes topics that have been omitted from review articles and books on the subject. The early density estimation methods, such as the histogram, kernel estimators, and orthogonal series estimators are still very popular, and recent research on them is described. Different types of restricted maximum likelihood density estimators, including order-restricted estimators, maximum penalized likelihood estimators, and sieve estimators, are discussed, where re...
- Published
- 1991
5. A Note on Chernoff and Lieberman's Generalized Probability Paper
- Author
-
G. W. Cran
- Subjects
Statistics and Probability ,Combinatorics ,Percentile ,Distribution (mathematics) ,Estimator ,Applied mathematics ,Graph paper ,Statistics, Probability and Uncertainty ,Invariant (mathematics) ,Scale parameter ,Plot (graphics) ,Mathematics - Abstract
The determination of plotting positions on probability graph paper so that the associated weighted least squares estimators of the scale parameter and the percentiles of a continuous distribution have certain properties is discussed. Necessary and sufficient conditions are given for an invariant optimal plot for percentile estimation. Also discussed is the derivation of ordered plotting positions.
- Published
- 1975
6. Statistics, Probability and Game Theory: Papers in Honor of David Blackwell
- Author
-
David Blackwell, Thomas S. Ferguson, James B. MacQueen, and Lloyd S. Shapley
- Subjects
Statistics and Probability ,Honor ,Sociology ,Statistics, Probability and Uncertainty ,Positive economics ,Game theory - Published
- 1999
7. Selected Papers of Hirotugu Akaike
- Author
-
null MTW, E. Parzen, K. Tanabe, and G. Kitagawa
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Published
- 1998
8. Bayesian Analysis in Econometrics and Statistics: The Zellner View and Papers
- Author
-
null MTW and Arnold Zellner
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Published
- 1998
9. Festschrift for Lucien Le Cam: Research Papers in Probability and Statistics
- Author
-
Grace L. Yang, Erik Torgersen, Karen Kafadar, and David Polland
- Subjects
Statistics and Probability ,Probability and statistics ,Statistics, Probability and Uncertainty ,Mathematical economics ,Mathematics - Published
- 1997
10. Selected Papers of C. R. Rao
- Author
-
null RJ, C. R. Rao, S. Das Gupta, J. K. Ghosh, S. K. Mitra, A. C. Mukhopadhyay, P. S. S. N. V. P. Rao, and Y. R. Sarma
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Published
- 1995
11. Approaches to Developing Questionnaires (Statistical Policy Working Paper 10)
- Author
-
Howard Schuman, Stanley Presser, Theresa J. DeMaio, and Paul B. Sheatsley
- Subjects
Statistics and Probability ,Questions and answers ,Management science ,business.industry ,Business ,Statistics, Probability and Uncertainty ,Public relations - Published
- 1985
12. Report on Statistical Uses of Administrative Records (Statistical Policy Working Paper 6 of the Office of Federal Statistical Policy and Standards)
- Author
-
Charles B. Nam
- Subjects
Statistics and Probability ,business.industry ,Accounting ,Business ,Data mining ,Statistics, Probability and Uncertainty ,computer.software_genre ,computer - Published
- 1983
13. Report on Statistical Disclosure and Disclosure-Avoidance Techniques. (Statistical Policy Working Paper 2)
- Author
-
Ivan P. Fellegi
- Subjects
Statistics and Probability ,Actuarial science ,Computer science ,Management science ,Statistics, Probability and Uncertainty - Published
- 1979
14. Contributions to Survey Sampling and Applied Statistics: Papers in Honor of H. O. Hartley
- Author
-
Herbert A. David and Ronald R. Regal
- Subjects
Statistics and Probability ,Honor ,Statistics ,Survey sampling ,Statistics, Probability and Uncertainty ,Mathematics - Published
- 1980
15. Essays in Statistical Science: Papers in Honour of P.A.P. Moran
- Author
-
Joseph Gani, E. A. Thompson, and E. J. Hannan
- Subjects
Statistics and Probability ,General Immunology and Microbiology ,Applied Mathematics ,Philosophy ,media_common.quotation_subject ,General Medicine ,General Biochemistry, Genetics and Molecular Biology ,Honour ,Sociology ,Statistics, Probability and Uncertainty ,Religious studies ,Moran's I ,General Agricultural and Biological Sciences ,Classics ,media_common - Published
- 1983
16. Statistical Policy Working Paper 1: Report on Statistics for Allocation of Funds
- Author
-
I. Richard Savage
- Subjects
Statistics and Probability ,Actuarial science ,Business ,Statistics, Probability and Uncertainty - Published
- 1980
17. Jacob Wolfowitz Selected Papers
- Author
-
U. Augustin, Somesh Das Gupta, L. Weiss, and J. Kiefer
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Published
- 1981
18. Studies in Probability and Statistics, Papers in Honour of Edwin J. G. Pitman
- Author
-
Stephen M. Stigler and E. J. Williams
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Published
- 1977
19. E. T. Jaynes: Papers on Probability, Statistics, and Statistical Physics
- Author
-
Carlos C. Rodriguez and R. D. Rosenkrantz
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Published
- 1985
20. Jack Carl Kiefer Collected Papers III: Design of Experiments
- Author
-
Jerome Sacks, Lawrence D. Brown, Ingram Olkin, Henry P. Wynn, Roger H. Farrell, and J. Kiefer
- Subjects
Statistics and Probability ,Operations research ,Computer science ,Design of experiments ,Art history ,Statistics, Probability and Uncertainty - Published
- 1987
21. Collected Papers of R. A. Fisher, Volumes III, IV, and V
- Author
-
I. J. Good and J. H. Bennet
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Published
- 1980
22. Off-Policy Confidence Interval Estimation with Confounded Markov Decision Process
- Author
-
Chengchun Shi, Jin Zhu, Shen Ye, Shikai Luo, Hongtu Zhu, and Rui Song
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Computer Science - Machine Learning ,Statistics - Machine Learning ,Machine Learning (stat.ML) ,Statistics, Probability and Uncertainty ,Machine Learning (cs.LG) - Abstract
This paper is concerned with constructing a confidence interval for a target policy's value offline based on a pre-collected observational data in infinite horizon settings. Most of the existing works assume no unmeasured variables exist that confound the observed actions. This assumption, however, is likely to be violated in real applications such as healthcare and technological industries. In this paper, we show that with some auxiliary variables that mediate the effect of actions on the system dynamics, the target policy's value is identifiable in a confounded Markov decision process. Based on this result, we develop an efficient off-policy value estimator that is robust to potential model misspecification and provide rigorous uncertainty quantification. Our method is justified by theoretical results, simulated and real datasets obtained from ridesharing companies. A Python implementation of the proposed procedure is available at https://github.com/Mamba413/cope.
- Published
- 2022
23. Permutation Tests at Nonparametric Rates
- Author
-
Bertanha, Marinho and Chung, EunYi
- Subjects
Methodology (stat.ME) ,FOS: Economics and business ,FOS: Computer and information sciences ,Statistics and Probability ,Econometrics (econ.EM) ,FOS: Mathematics ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty ,Statistics - Methodology ,Economics - Econometrics - Abstract
Classical two-sample permutation tests for equality of distributions have exact size in finite samples, but they fail to control size for testing equality of parameters that summarize each distribution. This paper proposes permutation tests for equality of parameters that are estimated at root-$n$ or slower rates. Our general framework applies to both parametric and nonparametric models, with two samples or one sample split into two subsamples. Our tests have correct size asymptotically while preserving exact size in finite samples when distributions are equal. They have no loss in local asymptotic power compared to tests that use asymptotic critical values. We propose confidence sets with correct coverage in large samples that also have exact coverage in finite samples if distributions are equal up to a transformation. We apply our theory to four commonly-used hypothesis tests of nonparametric functions evaluated at a point. Lastly, simulations show good finite sample properties, and two empirical examples illustrate our tests in practice., One PDF file contains main paper (35 pages) plus supplement (64 pages)
- Published
- 2022
24. Linear Hypothesis Testing in Linear Models With High-Dimensional Responses
- Author
-
Runze Li and Changcheng Li
- Subjects
Statistics and Probability ,05 social sciences ,Linear model ,High dimensional ,Linear hypothesis ,01 natural sciences ,Article ,010104 statistics & probability ,0502 economics and business ,Linear regression ,Applied mathematics ,0101 mathematics ,Statistics, Probability and Uncertainty ,Projection (set theory) ,050205 econometrics ,Mathematics - Abstract
In this paper, we propose a new projection test for linear hypotheses on regression coefficient matrices in linear models with high dimensional responses. We systematically study the theoretical properties of the proposed test. We first derive the optimal projection matrix for any given projection dimension to achieve the best power and provide an upper bound for the optimal dimension of projection matrix. We further provide insights into how to construct the optimal projection matrix. One- and two-sample mean problems can be formulated as special cases of linear hypotheses studied in this paper. We both theoretically and empirically demonstrate that the proposed test can outperform the existing ones for one- and two-sample mean problems. We conduct Monte Carlo simulation to examine the finite sample performance and illustrate the proposed test by a real data example.
- Published
- 2021
25. Discussion of Cui and Tchetgen Tchetgen (2020) and Qiu et al. (2020)
- Author
-
Hongming Pu and Bo Zhang
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty ,GeneralLiterature_MISCELLANEOUS - Abstract
We would like to congratulate the authors on two beautifully motivated and technically flawless papers and thank the editors for the opportunity to discuss both papers. We would like to make two po...
- Published
- 2021
26. Predictive Inference for Locally Stationary Time Series With an Application to Climate Data
- Author
-
Srinjoy Das and Dimitris N. Politis
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Series (mathematics) ,05 social sciences ,Prediction interval ,Linear prediction ,01 natural sciences ,Methodology (stat.ME) ,010104 statistics & probability ,Predictive inference ,13. Climate action ,0502 economics and business ,Kernel smoother ,General regression ,Applied mathematics ,0101 mathematics ,Statistics, Probability and Uncertainty ,Statistics - Methodology ,050205 econometrics ,Mathematics - Abstract
The Model-free Prediction Principle of Politis (2015) has been successfully applied to general regression problems, as well as problems involving stationary time series. However, with long time series, e.g. annual temperature measurements spanning over 100 years or daily financial returns spanning several years, it may be unrealistic to assume stationarity throughout the span of the dataset. In the paper at hand, we show how Model-free Prediction can be applied to handle time series that are only locally stationary, i.e., they can be assumed to be as stationary only over short time-windows. Surprisingly there is little literature on point prediction for general locally stationary time series even in model-based setups and there is no literature on the construction of prediction intervals of locally stationary time series. We attempt to fill this gap here as well. Both one-step-ahead point predictors and prediction intervals are constructed, and the performance of model-free is compared to model-based prediction using models that incorporate a trend and/or heteroscedasticity. Both aspects of the paper, model-free and model-based, are novel in the context of time-series that are locally (but not globally) stationary. We also demonstrate the application of our Model-based and Model-free prediction methods to speleothem climate data which exhibits local stationarity and show that our best model-free point prediction results outperform that obtained with the RAMPFIT algorithm previously used for analysis of this data.
- Published
- 2020
27. Spatial Variable Selection and An Application to Virginia Lyme Disease Emergence
- Author
-
Jie Li, David N. Gaines, Li Xu, Korine N. Kolivras, Xinwei Deng, Yili Hong, and Yimeng Xie
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Spatial variable ,biology ,030231 tropical medicine ,bacterial infections and mycoses ,medicine.disease ,biology.organism_classification ,Statistics - Applications ,01 natural sciences ,Virology ,3. Good health ,010104 statistics & probability ,03 medical and health sciences ,0302 clinical medicine ,Lyme disease ,Infectious disease (medical specialty) ,parasitic diseases ,medicine ,Applications (stat.AP) ,0101 mathematics ,Statistics, Probability and Uncertainty ,Borrelia burgdorferi ,Sensu stricto - Abstract
Lyme disease is an infectious disease that is caused by a bacterium called Borrelia burgdorferi sensu stricto. In the United States, Lyme disease is one of the most common infectious diseases. The major endemic areas of the disease are New England, Mid-Atlantic, East-North Central, South Atlantic, and West North-Central. Virginia is on the front-line of the disease's diffusion from the northeast to the south. One of the research objectives for the infectious disease community is to identify environmental and economic variables that are associated with the emergence of Lyme disease. In this paper, we use a spatial Poisson regression model to link the spatial disease counts and environmental and economic variables, and develop a spatial variable selection procedure to effectively identify important factors by using an adaptive elastic net penalty. The proposed methods can automatically select important covariates, while adjusting for possible spatial correlations of disease counts. The performance of the proposed method is studied and compared with existing methods via a comprehensive simulation study. We apply the developed variable selection methods to the Virginia Lyme disease data and identify important variables that are new to the literature. Supplementary materials for this paper are available online., Comment: 34 pages
- Published
- 2019
28. Machine intelligence for individualized decision making under a counterfactual world: A rejoinder
- Author
-
Yifan Cui and Eric J. Tchetgen Tchetgen
- Subjects
Statistics and Probability ,Counterfactual thinking ,Actuarial science ,Policy making ,Joint (building) ,Statistics, Probability and Uncertainty ,Unmeasured confounding ,Psychology ,Article - Abstract
This JASA rejoinder concerns the problem of individualized decision making under point, sign, and partial identification. The paper unifies various classical decision making strategies through a lower bound perspective proposed in Cui and Tchetgen Tchetgen (2020b) in the context of optimal treatment regimes under uncertainty due to unmeasured confounding. Building on this unified framework, the paper also provides a novel minimax solution (i.e., a rule that minimizes the maximum regret for so-called opportunists) for individualized decision making/policy assignment.
- Published
- 2021
29. A Dynamic Bayesian Model for Characterizing Cross-Neuronal Interactions During Decision-Making
- Author
-
Sam Behseta, Bo Zhou, Hernando Ombao, Babak Shahbaba, and David E. Moorman
- Subjects
Statistics and Probability ,Dynamic synchrony ,1.2 Psychological and socioeconomic processes ,Computer science ,Statistics & Probability ,Spike train ,Population ,Gaussian processes ,Bayesian inference ,Machine learning ,computer.software_genre ,Basic Behavioral and Social Science ,01 natural sciences ,Article ,Task (project management) ,Substance Misuse ,010104 statistics & probability ,03 medical and health sciences ,0302 clinical medicine ,Underpinning research ,Behavioral and Social Science ,Spike trains ,Econometrics ,0101 mathematics ,Set (psychology) ,education ,Prefrontal cortex ,Demography ,education.field_of_study ,business.industry ,Statistics ,Neurosciences ,Statistical model ,Neurological ,Artificial intelligence ,Statistics, Probability and Uncertainty ,Drug Abuse (NIDA only) ,Neural coding ,business ,computer ,030217 neurology & neurosurgery ,Decision-making - Abstract
The goal of this paper is to develop a novel statistical model for studying cross-neuronal spike train interactions during decision making. For an individual to successfully complete the task of decision-making, a number of temporally-organized events must occur: stimuli must be detected, potential outcomes must be evaluated, behaviors must be executed or inhibited, and outcomes (such as reward or no-reward) must be experienced. Due to the complexity of this process, it is likely the case that decision-making is encoded by the temporally-precise interactions between large populations of neurons. Most existing statistical models, however, are inadequate for analyzing such a phenomenon because they provide only an aggregated measure of interactions over time. To address this considerable limitation, we propose a dynamic Bayesian model which captures the time-varying nature of neuronal activity (such as the time-varying strength of the interactions between neurons). The proposed method yielded results that reveal new insight into the dynamic nature of population coding in the prefrontal cortex during decision making. In our analysis, we note that while some neurons in the prefrontal cortex do not synchronize their firing activity until the presence of a reward, a different set of neurons synchronize their activity shortly after stimulus onset. These differentially synchronizing sub-populations of neurons suggests a continuum of population representation of the reward-seeking task. Secondly, our analyses also suggest that the degree of synchronization differs between the rewarded and non-rewarded conditions. Moreover, the proposed model is scalable to handle data on many simultaneously-recorded neurons and is applicable to analyzing other types of multivariate time series data with latent structure. Supplementary materials (including computer codes) for our paper are available online.
- Published
- 2016
30. Statistical Inferences for Complex Dependence of Multimodal Imaging Data
- Author
-
Jinyuan Chang, Jing He, Jian Kang, and Mingcong Wu
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,FOS: Mathematics ,Mathematics - Statistics Theory ,Applications (stat.AP) ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty ,Statistics - Applications ,Statistics - Methodology - Abstract
Statistical analysis of multimodal imaging data is a challenging task, since the data involves high-dimensionality, strong spatial correlations and complex data structures. In this paper, we propose rigorous statistical testing procedures for making inferences on the complex dependence of multimodal imaging data. Motivated by the analysis of multi-task fMRI data in the Human Connectome Project (HCP) study, we particularly address three hypothesis testing problems: (a) testing independence among imaging modalities over brain regions, (b) testing independence between brain regions within imaging modalities, and (c) testing independence between brain regions across different modalities. Considering a general form for all the three tests, we develop a global testing procedure and a multiple testing procedure controlling the false discovery rate. We study theoretical properties of the proposed tests and develop a computationally efficient distributed algorithm. The proposed methods and theory are general and relevant for many statistical problems of testing independence structure among the components of high-dimensional random vectors with arbitrary dependence structures. We also illustrate our proposed methods via extensive simulations and analysis of five task fMRI contrast maps in the HCP study.
- Published
- 2023
31. Fair Policy Targeting
- Author
-
Viviano, Davide and Bradic, Jelena
- Subjects
Methodology (stat.ME) ,FOS: Economics and business ,FOS: Computer and information sciences ,Statistics and Probability ,Statistics - Machine Learning ,Econometrics (econ.EM) ,FOS: Mathematics ,Mathematics - Statistics Theory ,Machine Learning (stat.ML) ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty ,Statistics - Methodology ,Economics - Econometrics - Abstract
One of the major concerns of targeting interventions on individuals in social welfare programs is discrimination: individualized treatments may induce disparities across sensitive attributes such as age, gender, or race. This paper addresses the question of the design of fair and efficient treatment allocation rules. We adopt the non-maleficence perspective of first do no harm: we select the fairest allocation within the Pareto frontier. We cast the optimization into a mixed-integer linear program formulation, which can be solved using off-the-shelf algorithms. We derive regret bounds on the unfairness of the estimated policy function and small sample guarantees on the Pareto frontier under general notions of fairness. Finally, we illustrate our method using an application from education economics.
- Published
- 2023
32. Optimal Design of Experiments on Riemannian Manifolds
- Author
-
Hang Li and Enrique Del Castillo
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,FOS: Mathematics ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Mathematics::Differential Geometry ,Statistics, Probability and Uncertainty ,Statistics - Methodology - Abstract
The theory of optimal design of experiments has been traditionally developed on an Euclidean space. In this paper, new theoretical results and an algorithm for finding the optimal design of an experiment located on a Riemannian manifold are provided. It is shown that analogously to the results in Euclidean spaces, D-optimal and G-optimal designs are equivalent on manifolds, and we provide a lower bound for the maximum prediction variance of the response evaluated over the manifold. In addition, a converging algorithm that finds the optimal experimental design on manifold data is proposed. Numerical experiments demonstrate the importance of considering the manifold structure in a designed experiment when present, and the superiority of the proposed algorithm.
- Published
- 2022
33. Spherical Regression Models Using Projective Linear Transformations
- Author
-
Eric Klassen, Wei Wu, Anuj Srivastava, and Michael Rosenthal
- Subjects
Statistics and Probability ,Unit sphere ,Statistics::Theory ,Regression analysis ,Geometry ,Linear map ,Linear predictor function ,Statistics::Methodology ,Applied mathematics ,Projective linear group ,Statistics, Probability and Uncertainty ,Rotation (mathematics) ,Random variable ,Mathematics ,Variable (mathematics) - Abstract
This article studies the problem of modeling relationship between two spherical (or directional) random variables in a regression setup. Here the predictor and the response variables are constrained to be on a unit sphere and, due to this nonlinear condition, the standard Euclidean regression models do not apply. Several past papers have studied this problem, termed spherical regression, by modeling the response variable with a von Mises-Fisher (VMF) density with the mean given by a rotation of the predictor variable. The few papers that go beyond rigid rotations are limited to one- or two-dimensional spheres. This article extends the mean transformations to a larger group—the projective linear group of transformations—on unit spheres of arbitrary dimensions, while keeping the VMF density to model the noise. It develops a Newton–Raphson algorithm on the special linear group for estimating the MLE of regression parameter and establishes its asymptotic properties when the sample-size becomes large. Through ...
- Published
- 2014
34. Fisher-Pitman Permutation Tests Based on Nonparametric Poisson Mixtures with Application to Single Cell Genomics
- Author
-
Zhen Miao, Weihao Kong, Ramya Korlakai Vinayak, Wei Sun, and Fang Han
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,Computer Science - Machine Learning ,Statistics - Machine Learning ,Machine Learning (stat.ML) ,Statistics, Probability and Uncertainty ,Statistics - Methodology ,Machine Learning (cs.LG) - Abstract
This paper investigates the theoretical and empirical performance of Fisher-Pitman-type permutation tests for assessing the equality of unknown Poisson mixture distributions. Building on nonparametric maximum likelihood estimators (NPMLEs) of the mixing distribution, these tests are theoretically shown to be able to adapt to complicated unspecified structures of count data and also consistent against their corresponding ANOVA-type alternatives; the latter is a result in parallel to classic claims made by Robinson (Robinson, 1973). The studied methods are then applied to a single-cell RNA-seq data obtained from different cell types from brain samples of autism subjects and healthy controls; empirically, they unveil genes that are differentially expressed between autism and control subjects yet are missed using common tests. For justifying their use, rate optimality of NPMLEs is also established in settings similar to nonparametric Gaussian (Wu and Yang, 2020a) and binomial mixtures (Tian et al., 2017; Vinayak et al., 2019)., 52 pages
- Published
- 2022
35. An Algebraic Estimator for Large Spectral Density Matrices
- Author
-
Matteo Farnè, Matteo Barigozzi, Barigozzi, Matteo, and Farne, Matteo
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,FOS: Mathematics ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty ,Large spectral density matrix, Generalized Dynamic Factor Model, Dynamic rank, Sparsity ,Statistics - Methodology - Abstract
We propose a new estimator of high-dimensional spectral density matrices, called ALgebraic Spectral Estimator (ALSE), under the assumption of an underlying low rank plus sparse structure, as typically assumed in dynamic factor models. The ALSE is computed by minimizing a quadratic loss under a nuclear norm plus l1 norm constraint to control the latent rank and the residual sparsity pattern. The loss function requires as input the classical smoothed periodogram estimator and two threshold parameters, the choice of which is thoroughly discussed. We prove consistency of ALSE as both the dimension p and the sample size T diverge to infinity, as well as the recovery of latent rank and residual sparsity pattern with probability one. We then propose the UNshrunk ALgebraic Spectral Estimator (UNALSE), which is designed to minimize the Frobenius loss with respect to the pre-estimator while retaining the optimality of the ALSE. When applying UNALSE to a standard US quarterly macroeconomic dataset, we find evidence of two main sources of comovements: a real factor driving the economy at business cycle frequencies, and a nominal factor driving the higher frequency dynamics. The paper is also complemented by an extensive simulation exercise.
- Published
- 2022
36. Selective Inference for Hierarchical Clustering
- Author
-
Gao, Lucy L., Bien, Jacob, and Witten, Daniela
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,Statistics - Machine Learning ,Machine Learning (stat.ML) ,Statistics, Probability and Uncertainty ,Statistics - Methodology - Abstract
Classical tests for a difference in means control the type I error rate when the groups are defined a priori. However, when the groups are instead defined via clustering, then applying a classical test yields an extremely inflated type I error rate. Notably, this problem persists even if two separate and independent data sets are used to define the groups and to test for a difference in their means. To address this problem, in this paper, we propose a selective inference approach to test for a difference in means between two clusters. Our procedure controls the selective type I error rate by accounting for the fact that the choice of null hypothesis was made based on the data. We describe how to efficiently compute exact p-values for clusters obtained using agglomerative hierarchical clustering with many commonly-used linkages. We apply our method to simulated data and to single-cell RNA-sequencing data., Comment: Final accepted version
- Published
- 2022
37. An Empirical Bayes Approach to Shrinkage Estimation on the Manifold of Symmetric Positive-Definite Matrices
- Author
-
Chun-Hao Yang, Hani Doss, and Baba C. Vemuri
- Subjects
Statistics and Probability ,FOS: Mathematics ,Statistics::Methodology ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty - Abstract
The James-Stein estimator is an estimator of the multivariate normal mean and dominates the maximum likelihood estimator (MLE) under squared error loss. The original work inspired great interest in developing shrinkage estimators for a variety of problems. Nonetheless, research on shrinkage estimation for manifold-valued data is scarce. In this paper, we propose shrinkage estimators for the parameters of the Log-Normal distribution defined on the manifold of $N \times N$ symmetric positive-definite matrices. For this manifold, we choose the Log-Euclidean metric as its Riemannian metric since it is easy to compute and is widely used in applications. By using the Log-Euclidean distance in the loss function, we derive a shrinkage estimator in an analytic form and show that it is asymptotically optimal within a large class of estimators including the MLE, which is the sample Fr\'echet mean of the data. We demonstrate the performance of the proposed shrinkage estimator via several simulated data experiments. Furthermore, we apply the shrinkage estimator to perform statistical inference in diffusion magnetic resonance imaging problems., Comment: 54 pages, 5 figures
- Published
- 2022
38. Statistically Efficient Advantage Learning for Offline Reinforcement Learning in Infinite Horizons
- Author
-
Chengchun Shi, Shikai Luo, Yuan Le, Hongtu Zhu, and Rui Song
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Computer Science - Machine Learning ,Statistics - Machine Learning ,Machine Learning (stat.ML) ,Statistics, Probability and Uncertainty ,Machine Learning (cs.LG) - Abstract
We consider reinforcement learning (RL) methods in offline domains without additional online data collection, such as mobile health applications. Most of existing policy optimization algorithms in the computer science literature are developed in online settings where data are easy to collect or simulate. Their generalizations to mobile health applications with a pre-collected offline dataset remain unknown. The aim of this paper is to develop a novel advantage learning framework in order to efficiently use pre-collected data for policy optimization. The proposed method takes an optimal Q-estimator computed by any existing state-of-the-art RL algorithms as input, and outputs a new policy whose value is guaranteed to converge at a faster rate than the policy derived based on the initial Q-estimator. Extensive numerical experiments are conducted to back up our theoretical findings. A Python implementation of our proposed method is available at https://github.com/leyuanheart/SEAL.
- Published
- 2022
39. Adaptive Thresholding for Sparse Covariance Matrix Estimation
- Author
-
Weidong Liu and T. Tony Cai
- Subjects
Statistics and Probability ,FOS: Computer and information sciences ,Mathematical optimization ,62H12, 62F12 ,Covariance matrix ,05 social sciences ,Matrix norm ,Contrast (statistics) ,Estimator ,Covariance ,01 natural sciences ,Thresholding ,Methodology (stat.ME) ,010104 statistics & probability ,Rate of convergence ,0502 economics and business ,DNA microarray experiment ,0101 mathematics ,Statistics, Probability and Uncertainty ,Algorithm ,Statistics - Methodology ,050205 econometrics ,Mathematics - Abstract
In this paper we consider estimation of sparse covariance matrices and propose a thresholding procedure which is adaptive to the variability of individual entries. The estimators are fully data driven and enjoy excellent performance both theoretically and numerically. It is shown that the estimators adaptively achieve the optimal rate of convergence over a large class of sparse covariance matrices under the spectral norm. In contrast, the commonly used universal thresholding estimators are shown to be sub-optimal over the same parameter spaces. Support recovery is also discussed. The adaptive thresholding estimators are easy to implement. Numerical performance of the estimators is studied using both simulated and real data. Simulation results show that the adaptive thresholding estimators uniformly outperform the universal thresholding estimators. The method is also illustrated in an analysis on a dataset from a small round blue-cell tumors microarray experiment. A supplement to this paper which contains additional technical proofs is available online., Comment: To appear in Journal of the American Statistical Association
- Published
- 2011
40. Improved Small Domain Estimation via Compromise Regression Weights
- Author
-
Nicholas C. Henderson, Ravi Varadhan, and Thomas A. Louis
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,Statistics, Probability and Uncertainty ,Statistics - Methodology - Abstract
Shrinkage estimates of small domain parameters typically utilize a combination of a noisy "direct" estimate that only uses data from a specific small domain and a more stable regression estimate. When the regression model is misspecified, estimation performance for the noisier domains can suffer due to substantial shrinkage towards a poorly estimated regression surface. In this paper, we introduce a new class of robust, empirically-driven regression weights that target estimation of the small domain means under potential misspecification of the global regression model. Our regression weights are a convex combination of the model-based weights associated with the best linear unbiased predictor (BLUP) and those associated with the observed best predictor (OBP). The compromise parameter in this convex combination is found by minimizing a novel, unbiased estimate of the mean-squared prediction error for the small domain means, and we label the associated small domain estimates the "compromise best predictor", or CBP. Using a data-adaptive mixture for the regression weights enables the CBP to possess the robustness of the OBP while retaining the main advantages of the EBLUP whenever the regression model is correct. We demonstrate the use of the CBP in an application estimating gait speed in older adults.
- Published
- 2022
41. Robust Data-Driven Inference for Density-Weighted Average Derivatives
- Author
-
Richard K. Crump, Michael Jansson, and Matias D. Cattaneo
- Subjects
Statistics and Probability ,Mathematical optimization ,Mean squared error ,Bandwidth (signal processing) ,Monte Carlo method ,Robust statistics ,Estimator ,Inference ,Confidence interval ,Bandwidth (computing) ,Variety (universal algebra) ,Statistics, Probability and Uncertainty ,Algorithm ,Weighted arithmetic mean ,Mathematics - Abstract
This paper presents a novel data-driven bandwidth selector compatible with the small bandwidth asymptotics developed in Cattaneo, Crump, and Jansson (2009) for density-weighted average derivatives. The new bandwidth selector is of the plug-in variety, and is obtained based on a mean squared error expansion of the estimator of interest. An extensive Monte Carlo experiment shows a remarkable improvement in performance when the bandwidth-dependent robust inference procedures proposed by Cattaneo, Crump, and Jansson (2009) are coupled with this new data-driven bandwidth selector. The resulting robust data-driven confidence intervals compare favorably to the alternative procedures available in the literature. The online supplemental material to this paper contains further results from the simulation study.
- Published
- 2010
42. Comment
- Author
-
Armin Schwartzman
- Subjects
Statistics and Probability ,Normalization (statistics) ,False discovery rate ,Statistics ,Multiple comparisons problem ,Null distribution ,Econometrics ,Statistics, Probability and Uncertainty ,Standard normal table ,Mixture model ,Empirical distribution function ,Mathematics ,Statistical hypothesis testing - Abstract
In a series of recent articles, Bradley Efron has pointed out that in large-scale multiple testing problems, the observed distribution of the test statistics often does not match the theoretical null distribution (Efron, 2004, 2007a,b, 2008). The correction, which he termed “empirical null”, has been a subject of controversy in the statistical community. In Efron (2007a), he made the case that, even when the theoretical model is correct, the observed distribution of the test statistics can look different from the theoretical null distribution simply because of correlation between the test statistics. Efron’s present article represents an important step forward in the understanding of this problem. As opposed to previous papers, where the the effect of correlation was treated within the context of multiple testing problems and false discovery rates, Efron’s present article breaks through the confusion by separating these two concepts and focusing on the core issue, which is the behavior of a large collection of correlated normal variables. Only then, as applications, he presents the implications for false discovery rate analysis when the correlated normal variables are z-scores in a large-scale multiple testing problem. I think this separation is crucial and helps get us nearer a new theory of inference for high-dimensional data. In what follows, I present my own interpretation of Efron’s results on how correlation affects the empirical distribution of normal variables. As a shortcut, I work in the countinous domain directly rather than with histogram bins and avoid the inclusion of unnecessary constraints such as normalization, which applies very specifically to microarray data. For simplicity, I assume all the variables are standard normal rather than belonging to a mixture model, but may have an arbitrary correlation structure. Using these results, I hope to answer the question raised in Efron (2007a) of whether large-scale correlation can substantially widen the observed histogram, as in Figure 1 in that paper and Efron’s current one. The theoretical arguments below indicate that this is unlikely. On the contrary, large-scale correlation has to be mostly positive in order to satisfy positive definite contraints, and as a result, correlation will most likely narrow and shift the observed histogram rather than widen it. An important implication is that a wide histogram may be an indication of the presence of true signal, rather than an artifact of correlation. Having a model for the observed distribution under correlation, I will suggest a new way of fitting an empirical null, which may be more appropriate in cases where the observed widening is larger than could be explained by correlation alone. I will conclude commenting briefly on the possibility of performing this kind of analysis with χ2 variates rather than normal. I will try to keep the notation close to that of Efron, but some discrepancies in notation will be inevitable.
- Published
- 2010
43. Partial Correlation Estimation by Joint Sparse Regression Models
- Author
-
Ji Zhu, Nengfeng Zhou, Jie Peng, and Pei Wang
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Numerical linear algebra ,Estimation theory ,Covariance matrix ,Model selection ,Regression analysis ,computer.software_genre ,Article ,Methodology (stat.ME) ,ComputingMethodologies_PATTERNRECOGNITION ,Lasso (statistics) ,Statistics ,Statistics, Probability and Uncertainty ,Algorithm ,computer ,Statistics - Methodology ,Partial correlation ,Sparse matrix ,Mathematics - Abstract
In this paper, we propose a computationally efficient approach -- space(Sparse PArtial Correlation Estimation)-- for selecting non-zero partial correlations under the high-dimension-low-sample-size setting. This method assumes the overall sparsity of the partial correlation matrix and employs sparse regression techniques for model fitting. We illustrate the performance of space by extensive simulation studies. It is shown that space performs well in both non-zero partial correlation selection and the identification of hub variables, and also outperforms two existing methods. We then apply space to a microarray breast cancer data set and identify a set of hub genes which may provide important insights on genetic regulatory networks. Finally, we prove that, under a set of suitable assumptions, the proposed procedure is asymptotically consistent in terms of model selection and parameter estimation., Comment: A paper based on this report has been accepted for publication on Journal of the American Statistical Association(http://www.amstat.org/publications/JASA/)
- Published
- 2009
44. Learn From Thy Neighbor: Parallel-Chain and Regional Adaptive MCMC
- Author
-
Chao Yang, Radu V. Craiu, and Jeffrey S. Rosenthal
- Subjects
Statistics and Probability ,Theoretical computer science ,Markov chain ,Adaptive algorithm ,Monte Carlo method ,Markov chain Monte Carlo ,Random walk ,symbols.namesake ,Kernel method ,Metropolis–Hastings algorithm ,Econometrics ,symbols ,State space ,Statistics, Probability and Uncertainty ,Mathematics - Abstract
Starting with the seminal paper of Haario, Saksman and Tamminen (Haario et al. (2001)), a substantial amount of work has been done to validate adaptive Markov chain Monte Carlo algorithms. In this paper we focus on two practical aspects of adaptive Metropolis samplers. First, we draw attention to the deficient performance of standard adaptation when the target distribution is multi-modal. We propose a parallel chain adaptation strategy that incorporates multiple Markov chains which are run in parallel. Second, we note that
- Published
- 2009
45. Intervention and Causality
- Author
-
Casper J. Albers, Catriona M. Queen, and Psychometrics and Statistics
- Subjects
Statistics and Probability ,Conditional independence ,Operations research ,Computer science ,Management system ,Bayesian probability ,Econometrics ,Statistics, Probability and Uncertainty ,Flow network ,Traffic flow ,Traffic generation model ,Dynamic Bayesian network ,Network traffic simulation - Abstract
Real-time traffic flow data across entire networks can be used in a traffic management system to monitor current traffic flows so that traffic can be directed and managed efficiently. Reliable short-term forecasting models of traffic flows are crucial for the success of any traffic management system.\ud \ud The model proposed in this paper for forecasting traffic flows is a multivariate Bayesian dynamic model called the multiregression dynamic model (MDM). This model is an example of a dynamic Bayesian network and is designed to preserve the conditional independences and causal drive exhibited by the traffic flow series.\ud \ud Sudden changes can occur in traffic flow series in response to such events as traffic accidents or roadworks. A traffic management system is particularly useful at such times of change. To ensure that the associated forecasting model continues to produce reliable forecasts, despite the change, the MDM uses the technique of external intervention. This paper will demonstrate how intervention works in the MDM and how it can improve forecast performance at times of change.\ud \ud External intervention has also been used in the context of Bayesian networks to identify causal relationships between variables, and in dynamic Bayesian networks to identify lagged causal relationships between time series. This paper goes beyond the identification of lagged causal relationships previously addressed using intervention in dynamic Bayesian networks, to show how intervention in the MDM can be used to identify contemporaneous causal relationships between time series.
- Published
- 2009
46. To How Many Simultaneous Hypothesis Tests Can Normal, Student'stor Bootstrap Calibration Be Applied?
- Author
-
Jianqing Fan, Qiwei Yao, and Peter A. Hall
- Subjects
Statistics and Probability ,05 social sciences ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,01 natural sciences ,010104 statistics & probability ,Sampling distribution ,Student's t-distribution ,Sample size determination ,0502 economics and business ,Statistics ,FOS: Mathematics ,Test statistic ,jel:C1 ,Z-test ,HA Statistics ,0101 mathematics ,Statistics, Probability and Uncertainty ,Student's t-test ,Statistic ,050205 econometrics ,Mathematics ,Statistical hypothesis testing - Abstract
In the analysis of microarray data, and in some other contemporary statistical problems, it is not uncommon to apply hypothesis tests in a highly simultaneous way. The number, $\nu$ say, of tests used can be much larger than the sample sizes, $n$, to which the tests are applied, yet we wish to calibrate the tests so that the overall level of the simultaneous test is accurate. Often the sampling distribution is quite different for each test, so there may not be an opportunity for combining data across samples. In this setting, how large can $\nu$ be, as a function of $n$, before level accuracy becomes poor? In the present paper we answer this question in cases where the statistic under test is of Student's $t$ type. We show that if either Normal or Student's $t$ distribution is used for calibration then the level of the simultaneous test is accurate provided $\log\nu$ increases at a strictly slower rate than $n^{1/3}$ as $n$ diverges. On the other hand, if bootstrap methods are used for calibration then we may choose $\log\nu$ almost as large as $n\half$ and still achieve asymptotic level accuracy. The implications of these results are explored both theoretically and numerically., Comment: 25 pages paper
- Published
- 2007
47. Divide-and-Conquer: A Distributed Hierarchical Factor Approach to Modeling Large-Scale Time Series Data
- Author
-
Gao, Zhaoxing and Tsay, Ruey S.
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,FOS: Economics and business ,Statistics and Probability ,Econometrics (econ.EM) ,Statistics, Probability and Uncertainty ,Statistics - Methodology ,Economics - Econometrics - Abstract
This paper proposes a hierarchical approximate-factor approach to analyzing high-dimensional, large-scale heterogeneous time series data using distributed computing. The new method employs a multiple-fold dimension reduction procedure using Principal Component Analysis (PCA) and shows great promises for modeling large-scale data that cannot be stored nor analyzed by a single machine. Each computer at the basic level performs a PCA to extract common factors among the time series assigned to it and transfers those factors to one and only one node of the second level. Each 2nd-level computer collects the common factors from its subordinates and performs another PCA to select the 2nd-level common factors. This process is repeated until the central server is reached, which collects common factors from its direct subordinates and performs a final PCA to select the global common factors. The noise terms of the 2nd-level approximate factor model are the unique common factors of the 1st-level clusters. We focus on the case of 2 levels in our theoretical derivations, but the idea can easily be generalized to any finite number of hierarchies. We discuss some clustering methods when the group memberships are unknown and introduce a new diffusion index approach to forecasting. We further extend the analysis to unit-root nonstationary time series. Asymptotic properties of the proposed method are derived for the diverging dimension of the data in each computing unit and the sample size $T$. We use both simulated data and real examples to assess the performance of the proposed method in finite samples, and compare our method with the commonly used ones in the literature concerning the forecastability of extracted factors., Comment: 48 pages, 10 figures
- Published
- 2022
48. Covariate-Assisted Sparse Tensor Completion
- Author
-
Hilda S. Ibriga and Will Wei Sun
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,Computer Science - Machine Learning ,Statistics - Machine Learning ,FOS: Mathematics ,Machine Learning (stat.ML) ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty ,Statistics - Methodology ,Machine Learning (cs.LG) - Abstract
We aim to provably complete a sparse and highly-missing tensor in the presence of covariate information along tensor modes. Our motivation comes from online advertising where users click-through-rates (CTR) on ads over various devices form a CTR tensor that has about 96% missing entries and has many zeros on non-missing entries, which makes the standalone tensor completion method unsatisfactory. Beside the CTR tensor, additional ad features or user characteristics are often available. In this paper, we propose Covariate-assisted Sparse Tensor Completion (COSTCO) to incorporate covariate information for the recovery of the sparse tensor. The key idea is to jointly extract latent components from both the tensor and the covariate matrix to learn a synthetic representation. Theoretically, we derive the error bound for the recovered tensor components and explicitly quantify the improvements on both the reveal probability condition and the tensor recovery accuracy due to covariates. Finally, we apply COSTCO to an advertisement dataset consisting of a CTR tensor and ad covariate matrix, leading to 23% accuracy improvement over the baseline. An important by-product is that ad latent components from COSTCO reveal interesting ad clusters, which are useful for better ad targeting., To Appear in Journal of the American Statistical Association
- Published
- 2022
49. Self-supervised Metric Learning in Multi-View Data: A Downstream Task Perspective
- Author
-
Shulei Wang
- Subjects
Methodology (stat.ME) ,FOS: Computer and information sciences ,Statistics and Probability ,Computer Science - Machine Learning ,Statistics - Machine Learning ,FOS: Mathematics ,Machine Learning (stat.ML) ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty ,Statistics - Methodology ,Machine Learning (cs.LG) - Abstract
Self-supervised metric learning has been a successful approach for learning a distance from an unlabeled dataset. The resulting distance is broadly useful for improving various distance-based downstream tasks, even when no information from downstream tasks is utilized in the metric learning stage. To gain insights into this approach, we develop a statistical framework to theoretically study how self-supervised metric learning can benefit downstream tasks in the context of multi-view data. Under this framework, we show that the target distance of metric learning satisfies several desired properties for the downstream tasks. On the other hand, our investigation suggests the target distance can be further improved by moderating each direction’s weights. In addition, our analysis precisely characterizes the improvement by self-supervised metric learning on four commonly used downstream tasks: sample identification, two-sample testing, k-means clustering, and k-nearest neighbor classification. When the distance is estimated from an unlabeled dataset, we establish the upper bound on distance estimation’s accuracy and the number of samples sufficient for downstream task improvement. Finally, numerical experiments are presented to support the theoretical results in the paper.
- Published
- 2022
50. Efficient Fully Distribution-Free Center-Outward Rank Tests for Multiple-Output Regression and MANOVA
- Author
-
Marc Hallin, Daniel Hlubinka, and Šárka Hudecová
- Subjects
Statistics and Probability ,Distribution-free tests ,Multivariate ranks ,Multivariate signs ,Hjajek representation ,théorie et applications [Econométrie et méthodes statistiques] ,Statistics, Probability and Uncertainty - Abstract
Extending rank-based inference to a multivariate setting such as multiple-output regression or MANOVA with unspecified d-dimensional error density has remained an open problem for more than half a century. None of the many solutions proposed so far is enjoying the combination of distribution-freeness and efficiency that makes rank-based inference a successful tool in the univariate setting. A concept of center- outward multivariate ranks and signs based on measure transportation ideas has been introduced recently. Center-outward ranks and signs are not only distribution-free but achieve in dimension d > 1 the (essential) maximal ancillarity property of traditional univariate ranks. In the present case, we show that fully distribution-free testing procedures based on center-outward ranks can achieve parametric efficiency. We establish the Hajek representation and asymptotic normality results required in the construction of such tests in multiple-output regression and MANOVA models. Simulations and an empirical study demonstrate the excellent performance of the proposed procedures., info:eu-repo/semantics/published
- Published
- 2022
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.