40 results on '"Surya T. Tokdar"'
Search Results
2. Multiple objects evoke fluctuating responses in several regions of the visual pathway
- Author
-
Meredith N Schmehl, Valeria C Caruso, Yunran Chen, Na Young Jun, Shawn M Willett, Jeff T Mohl, Douglas A Ruff, Marlene Cohen, Akinori F Ebihara, Winrich A Freiwald, Surya T Tokdar, and Jennifer M Groh
- Subjects
neural representation ,neural code ,multiplexing ,object vision ,figure ground segregation ,visual system ,Medicine ,Science ,Biology (General) ,QH301-705.5 - Abstract
How neural representations preserve information about multiple stimuli is mysterious. Because tuning of individual neurons is coarse (e.g., visual receptive field diameters can exceed perceptual resolution), the populations of neurons potentially responsive to each individual stimulus can overlap, raising the question of how information about each item might be segregated and preserved in the population. We recently reported evidence for a potential solution to this problem: when two stimuli were present, some neurons in the macaque visual cortical areas V1 and V4 exhibited fluctuating firing patterns, as if they responded to only one individual stimulus at a time (Jun et al., 2022). However, whether such an information encoding strategy is ubiquitous in the visual pathway and thus could constitute a general phenomenon remains unknown. Here, we provide new evidence that such fluctuating activity is also evoked by multiple stimuli in visual areas responsible for processing visual motion (middle temporal visual area, MT), and faces (middle fundus and anterolateral face patches in inferotemporal cortex – areas MF and AL), thus extending the scope of circumstances in which fluctuating activity is observed. Furthermore, consistent with our previous results in the early visual area V1, MT exhibits fluctuations between the representations of two stimuli when these form distinguishable objects but not when they fuse into one perceived object, suggesting that fluctuating activity patterns may underlie visual object formation. Taken together, these findings point toward an updated model of how the brain preserves sensory information about multiple stimuli for subsequent processing and behavioral action.
- Published
- 2024
- Full Text
- View/download PDF
3. Coordinated multiplexing of information about separate objects in visual cortex
- Author
-
Na Young Jun, Douglas A Ruff, Lily E Kramer, Brittany Bowes, Surya T Tokdar, Marlene R Cohen, and Jennifer M Groh
- Subjects
noise correlations ,variability ,multiplexing ,population coding ,object vision ,Medicine ,Science ,Biology (General) ,QH301-705.5 - Abstract
Sensory receptive fields are large enough that they can contain more than one perceptible stimulus. How, then, can the brain encode information about each of the stimuli that may be present at a given moment? We recently showed that when more than one stimulus is present, single neurons can fluctuate between coding one vs. the other(s) across some time period, suggesting a form of neural multiplexing of different stimuli (Caruso et al., 2018). Here, we investigate (a) whether such coding fluctuations occur in early visual cortical areas; (b) how coding fluctuations are coordinated across the neural population; and (c) how coordinated coding fluctuations depend on the parsing of stimuli into separate vs. fused objects. We found coding fluctuations do occur in macaque V1 but only when the two stimuli form separate objects. Such separate objects evoked a novel pattern of V1 spike count (‘noise’) correlations involving distinct distributions of positive and negative values. This bimodal correlation pattern was most pronounced among pairs of neurons showing the strongest evidence for coding fluctuations or multiplexing. Whether a given pair of neurons exhibited positive or negative correlations depended on whether the two neurons both responded better to the same object or had different object preferences. Distinct distributions of spike count correlations based on stimulus preferences were also seen in V4 for separate objects but not when two stimuli fused to form one object. These findings suggest multiple objects evoke different response dynamics than those evoked by single stimuli, lending support to the multiplexing hypothesis and suggesting a means by which information about multiple objects can be preserved despite the apparent coarseness of sensory coding.
- Published
- 2022
- Full Text
- View/download PDF
4. Computer Emulation with Nonstationary Gaussian Processes.
- Author
-
Silvia Montagna and Surya T. Tokdar
- Published
- 2016
- Full Text
- View/download PDF
5. Coordinated multiplexing of information about separate objects in visual cortex
- Author
-
Brittany S. Bowes, Kramer Le, Marlene R. Cohen, Douglas A. Ruff, Surya T. Tokdar, Jennifer M. Groh, and Na Young Jun
- Subjects
genetic structures ,Sensory system ,Biology ,Stimulus (physiology) ,Multiplexing ,General Biochemistry, Genetics and Molecular Biology ,Correlation ,03 medical and health sciences ,0302 clinical medicine ,Time-division multiplexing ,medicine ,Animals ,Visual Cortex ,030304 developmental biology ,Neurons ,0303 health sciences ,General Immunology and Microbiology ,General Neuroscience ,Brain ,General Medicine ,Visual cortex ,medicine.anatomical_structure ,Receptive field ,Macaca ,Neuroscience ,030217 neurology & neurosurgery ,Coding (social sciences) - Abstract
Sensory receptive fields are large enough that they can contain more than one perceptible stimulus. How, then, can the brain encode information abouteachof the stimuli that may be present at a given moment? We recently showed that when more than one stimulus is present, single neurons can fluctuate between coding one vs. the other(s) across some time period, suggesting a form of neural multiplexing of different stimuli (Caruso et al., 2018). Here we investigate (a) whether such coding fluctuations occur in early visual cortical areas; (b) how coding fluctuations are coordinated across the neural population; and (c) how coordinated coding fluctuations depend on the parsing of stimuli into separate vs. fused objects. We found coding fluctuations do occur in macaque V1 but only when the two stimuli form separate objects. Such separate objects evoked a novel pattern of V1 spike count (“noise”) correlations involving distinct distributions of positive and negative values. This bimodal correlation pattern was most pronounced among pairs of neurons showing the strongest evidence for coding fluctuations or multiplexing. Whether a given pair of neurons exhibited positive or negative correlations depended on whether the two neurons both responded better to the same object or had different object preferences. Distinct distributions of spike count correlations based on stimulus preferences were also seen in V4 for separate objects but not when two stimuli fused to form one object. These findings suggest multiple objects evoke different response dynamics than those evoked by single stimuli, lending support to the multiplexing hypothesis and suggesting a means by which information about multiple objects can be preserved despite the apparent coarseness of sensory coding.Significance StatementHow the brain separates information about multiple objects despite overlap in the neurons responsive to each item is not well understood. Here we show that some neurons in V1 exhibit coding fluctuations in response to two objects, and that these coding fluctuations are coordinated at the population level in ways that are not observed for single objects. Broadly similar results were obtained in V4. These response dynamics lend support to the hypothesis that information about individual objects may be multiplexed across the neural population, preserving information about each item despite the coarseness of sensory coding.
- Published
- 2022
- Full Text
- View/download PDF
6. Joint Quantile Regression for Spatial Data
- Author
-
Surya T. Tokdar and Xu Chen
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Tail dependence ,Quantile regression ,Copula (probability theory) ,Methodology (stat.ME) ,Generative model ,Econometrics ,Statistics, Probability and Uncertainty ,Uncertainty quantification ,Spatial analysis ,Statistics - Methodology ,Smoothing ,Quantile ,Mathematics - Abstract
Linear quantile regression is a powerful tool to investigate how predictors may affect a response heterogeneously across different quantile levels. Unfortunately, existing approaches find it extremely difficult to adjust for any dependency between observation units, largely because such methods are not based upon a fully generative model of the data. For analyzing spatially indexed data, we address this difficulty by generalizing the joint quantile regression model of Yang and Tokdar (2017) and characterizing spatial dependence via a Gaussian or $t$ copula process on the underlying quantile levels of the observation units. A Bayesian semiparametric approach is introduced to perform inference of model parameters and carry out spatial quantile smoothing. An effective model comparison criteria is provided, particularly for selecting between different model specifications of tail heaviness and tail dependence. Extensive simulation studies and an application to particulate matter concentration in northeast US are presented to illustrate substantial gains in inference quality, accuracy and uncertainty quantification over existing alternatives., Comment: 30 pages, 10 figures
- Published
- 2021
- Full Text
- View/download PDF
7. Posterior consistency in conditional distribution estimation.
- Author
-
Debdeep Pati, David B. Dunson, and Surya T. Tokdar
- Published
- 2013
- Full Text
- View/download PDF
8. Author response: Coordinated multiplexing of information about separate objects in visual cortex
- Author
-
Na Young Jun, Douglas A Ruff, Lily E Kramer, Brittany Bowes, Surya T Tokdar, Marlene R Cohen, and Jennifer M Groh
- Published
- 2022
- Full Text
- View/download PDF
9. A Censored-Data Multiperiod Inventory Problem with Newsvendor Demand Distributions.
- Author
-
Arnab Bisi, Maqbool Dada, and Surya T. Tokdar
- Published
- 2011
- Full Text
- View/download PDF
10. Detection of bursts in extracellular spike trains using hidden semi-Markov point process models.
- Author
-
Surya T. Tokdar, Peiyi Xi, Ryan C. Kelly, and Robert E. Kass
- Published
- 2010
- Full Text
- View/download PDF
11. On the Empirical Bayes approach to the problem of multiple testing.
- Author
-
Malgorzata Bogdan, Jayanta K. Ghosh, Aleksandra Ochman, and Surya T. Tokdar
- Published
- 2007
- Full Text
- View/download PDF
12. Heavy-Tailed Density Estimation
- Author
-
Surya T. Tokdar, Sheng Jiang, and Erika L. Cunningham
- Subjects
Statistics and Probability ,Methodology (stat.ME) ,FOS: Computer and information sciences ,62G ,FOS: Mathematics ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Statistics, Probability and Uncertainty ,Statistics - Methodology - Abstract
A novel statistical method is proposed and investigated for estimating a heavy tailed density under mild smoothness assumptions. Statistical analyses of heavy-tailed distributions are susceptible to the problem of sparse information in the tail of the distribution getting washed away by unrelated features of a hefty bulk. The proposed Bayesian method avoids this problem by incorporating smoothness and tail regularization through a carefully specified semiparametric prior distribution, and is able to consistently estimate both the density function and its tail index at near minimax optimal rates of contraction. A joint, likelihood driven estimation of the bulk and the tail is shown to help improve uncertainty assessment in estimating the tail index parameter and offer more accurate and reliable estimates of the high tail quantiles compared to thresholding methods., Comment: Combined article with all technical details uploaded here to complement JASA publication
- Published
- 2022
- Full Text
- View/download PDF
13. Bayesian Analysis of Dynamic Linear Topic Models.
- Author
-
Chris Glynn, Surya T. Tokdar, David L. Banks, and Brian Howard
- Published
- 2015
14. Bayesian Test of Normality Versus a Dirichlet Process Mixture Alternative
- Author
-
Ryan Martin and Surya T. Tokdar
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Multivariate statistics ,Applied Mathematics ,media_common.quotation_subject ,Bayesian probability ,Nonparametric statistics ,Univariate ,Mathematics - Statistics Theory ,Bayes factor ,Statistics Theory (math.ST) ,Statistics - Computation ,Methodology (stat.ME) ,Normal distribution ,FOS: Mathematics ,Applied mathematics ,Statistics, Probability and Uncertainty ,Cluster analysis ,Computation (stat.CO) ,Statistics - Methodology ,Normality ,Mathematics ,media_common - Abstract
We propose a Bayesian test of normality for univariate or multivariate data against alternative nonparametric models characterized by Dirichlet process mixture distributions. The alternative models are based on the principles of embedding and predictive matching. They can be interpreted to offer random granulation of a normal distribution into a mixture of normals with mixture components occupying a smaller volume the farther they are from the distribution center. A scalar parametrization based on latent clustering is used to cover an entire spectrum of separation between the normal distributions and the alternative models. An efficient sequential importance sampler is developed to calculate Bayes factors. Simulations indicate the proposed test can detect non-normality without favoring the nonparametric alternative when normality holds., 24 pages, 5 figures, 1 table
- Published
- 2019
- Full Text
- View/download PDF
15. Analyzing second order stochasticity of neural spiking under stimuli-bundle exposure
- Author
-
Surya T. Tokdar, Jennifer M. Groh, Azeem Zaman, Valeria C. Caruso, Chris Glynn, Shawn M. Willett, and Jeff T. Mohl
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Quantitative Biology::Neurons and Cognition ,Computer science ,Stochastic process ,Spike train ,Stimulus (physiology) ,Bayesian inference ,Statistics - Applications ,Article ,Synthetic data ,Point process ,symbols.namesake ,Modeling and Simulation ,Bundle ,symbols ,Applications (stat.AP) ,Statistics, Probability and Uncertainty ,Biological system ,Gaussian process - Abstract
Conventional analysis of neuroscience data involves computing average neural activity over a group of trials and/or a period of time. This approach may be particularly problematic when assessing the response patterns of neurons to more than one simultaneously presented stimulus. In such cases, the brain must represent each individual component of the stimuli bundle, but trial-and-time-pooled averaging methods are fundamentally unequipped to address the means by which multi-item representation occurs. We introduce and investigate a novel statistical analysis framework that relates the firing pattern of a single cell, exposed to a stimuli bundle, to the ensemble of its firing patterns under each constituent stimulus. Existing statistical tools focus on what may be called "first order stochasticity" in trial-to-trial variation in the form of unstructured noise around a fixed firing rate curve associated with a given stimulus. Our analysis is based upon the theoretical premise that exposure to a stimuli bundle induces additional stochasticity in the cell's response pattern, in the form of a stochastically varying recombination of its single stimulus firing rate curves. We discuss challenges to statistical estimation of such "second order stochasticity" and address them with a novel dynamic admixture Poisson process (DAPP) model. DAPP is a hierarchical point process model that decomposes second order stochasticity into a Gaussian stochastic process and a random vector of interpretable features, and, facilitates borrowing of information on the latter across repeated trials through latent clustering. We present empirical evidence of the utility of the DAPP analysis with synthetic and real neural recordings., Comment: 26 pages, 7 figures
- Published
- 2021
- Full Text
- View/download PDF
16. Sensorimotor abilities predict on-field performance in professional baseball
- Author
-
Kyle Burris, Benjamin Ramger, Surya T. Tokdar, Sunith Suresh, Jerome P. Reiter, L. Gregory Appelbaum, and Kelly Vittetoe
- Subjects
Adult ,Adolescent ,Bayesian probability ,lcsh:Medicine ,Sample (statistics) ,Latent variable ,League ,Athletic Performance ,Baseball ,Field (computer science) ,Article ,Sensorimotor skills ,Nike ,03 medical and health sciences ,Young Adult ,0302 clinical medicine ,Theoretical ,Models ,Humans ,lcsh:Science ,Psychomotor learning ,Multidisciplinary ,lcsh:R ,Neurosciences ,030229 sport sciences ,Models, Theoretical ,030221 ophthalmology & optometry ,lcsh:Q ,Psychology ,Algorithms ,Psychomotor Performance ,Cognitive psychology - Abstract
Baseball players must be able to see and react in an instant, yet it is hotly debated whether superior performance is associated with superior sensorimotor abilities. In this study, we compare sensorimotor abilities, measured through 8 psychomotor tasks comprising the Nike Sensory Station assessment battery, and game statistics in a sample of 252 professional baseball players to evaluate the links between sensorimotor skills and on-field performance. For this purpose, we develop a series of Bayesian hierarchical latent variable models enabling us to compare statistics across professional baseball leagues. Within this framework, we find that sensorimotor abilities are significant predictors of on-base percentage, walk rate and strikeout rate, accounting for age, position, and league. We find no such relationship for either slugging percentage or fielder-independent pitching. The pattern of results suggests performance contributions from both visual-sensory and visual-motor abilities and indicates that sensorimotor screenings may be useful for player scouting.
- Published
- 2018
- Full Text
- View/download PDF
17. Joint Estimation of Quantile Planes Over Arbitrary Predictor Spaces
- Author
-
Yun Yang and Surya T. Tokdar
- Subjects
Statistics and Probability ,Mathematical optimization ,Estimation theory ,05 social sciences ,Statistical model ,Markov chain Monte Carlo ,Bayesian inference ,01 natural sciences ,Statistics::Computation ,Quantile regression ,010104 statistics & probability ,symbols.namesake ,0502 economics and business ,symbols ,Statistics::Methodology ,Applied mathematics ,0101 mathematics ,Statistics, Probability and Uncertainty ,Likelihood function ,Gaussian process ,050205 econometrics ,Quantile ,Mathematics - Abstract
In spite of the recent surge of interest in quantile regression, joint estimation of linear quantile planes remains a great challenge in statistics and econometrics. We propose a novel parameterization that characterizes any collection of noncrossing quantile planes over arbitrarily shaped convex predictor domains in any dimension by means of unconstrained scalar, vector and function valued parameters. Statistical models based on this parameterization inherit a fast computation of the likelihood function, enabling penalized likelihood or Bayesian approaches to model fitting. We introduce a complete Bayesian methodology by using Gaussian process prior distributions on the function valued parameters and develop a robust and efficient Markov chain Monte Carlo parameter estimation. The resulting method is shown to offer posterior consistency under mild tail and regularity conditions. We present several illustrative examples where the new method is compared against existing approaches and is found to offer better accuracy, coverage and model fit. Supplementary materials for this article are available online.
- Published
- 2017
- Full Text
- View/download PDF
18. Contributors
- Author
-
Taeryon Choi, James S. Clark, Erika Cunningham, Maria DeYoreo, J.-L. Dortet-Bernadet, Y. Fan, Michele Guindani, Maria Kalli, Nadja Klein, Thomas Kneib, Jeong Hwan Kook, Athanasios Kottas, Peter J. Lenk, Yadong Lu, Andrew A. Manderson, Giampiero Marra, Yinsen Miao, Kevin Murray, John T. Ormerod, Rosalba Radice, T. Rodrigues, Surya T. Tokdar, Berwin A. Turlach, and Marina Vannucci
- Published
- 2020
- Full Text
- View/download PDF
19. A vignette on model-based quantile regression: analysing excess zero response
- Author
-
Erika Cunningham, Surya T. Tokdar, and James S. Clark
- Subjects
R package ,Vignette ,Econometrics ,Tobit model ,Regression analysis ,Censoring (statistics) ,Mathematics ,Zero (linguistics) ,Quantile regression ,Quantile - Abstract
Quantile regression is widely seen as an ideal tool to understand complex predictor-response relations. Its biggest promise rests in its ability to quantify whether and how predictor effects vary across response quantile levels. But this promise has not been fully met due to a lack of statistical estimation methods that perform a rigorous, joint analysis of all quantile levels. This gap has been recently bridged by Yang and Tokdar [18] . Here we demonstrate how their joint quantile regression method, as encoded in the R package qrjoint , offers a comprehensive and model-based regression analysis framework. This chapter is an R vignette where we illustrate how to fit models, interpret coefficients, improve and compare models and obtain predictions under this framework. Our case study is an application to ecology where we analyse how the abundance of red maple trees depends on topographical and geographical features of the location. A complete absence of the species contributes excess zeros in the response data. We treat such excess zeros as left censoring in the spirit of a Tobit regression analysis. By utilising the generative nature of the joint quantile regression model, we not only adjust for censoring but also treat it as an object of independent scientific interest.
- Published
- 2020
- Full Text
- View/download PDF
20. Variable Selection Consistency of Gaussian Process Regression
- Author
-
Surya T. Tokdar and Sheng Jiang
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Smoothness (probability theory) ,62G08, 62G20 ,Posterior probability ,Feature selection ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Bayesian inference ,Methodology (stat.ME) ,Variable (computer science) ,symbols.namesake ,Consistency (statistics) ,FOS: Mathematics ,symbols ,Applied mathematics ,Statistics, Probability and Uncertainty ,Random variable ,Gaussian process ,Statistics - Methodology ,Mathematics - Abstract
Bayesian nonparametric regression under a rescaled Gaussian process prior offers smoothness-adaptive function estimation with near minimax-optimal error rates. Hierarchical extensions of this approach, equipped with stochastic variable selection, are known to also adapt to the unknown intrinsic dimension of a sparse true regression function. But it remains unclear if such extensions offer variable selection consistency, that is, if the true subset of important variables could be consistently learned from the data. It is shown here that variable consistency may indeed be achieved with such models at least when the true regression function has finite smoothness to induce a polynomially larger penalty on inclusion of false positive predictors. Our result covers the high-dimensional asymptotic setting where the predictor dimension is allowed to grow with the sample size. The proof utilizes Schwartz theory to establish that the posterior probability of wrong selection vanishes asymptotically. A necessary and challenging technical development involves providing sharp upper and lower bounds to small ball probabilities at all rescaling levels of the Gaussian process prior, a result that could be of independent interest.
- Published
- 2019
21. Sensitivity and specificity of a Bayesian single trial analysis for time varying neural signals
- Author
-
Jeff T. Mohl, Surya T. Tokdar, Jennifer M. Groh, and Valeria C. Caruso
- Subjects
0303 health sciences ,business.industry ,Bayesian probability ,Single stimulus ,Pattern recognition ,Stimulus (physiology) ,Correct response ,Article ,Synthetic data ,03 medical and health sciences ,Neural activity ,0302 clinical medicine ,Categorization ,Quantitative Biology - Neurons and Cognition ,FOS: Biological sciences ,Neurons and Cognition (q-bio.NC) ,Artificial intelligence ,Single trial ,business ,030217 neurology & neurosurgery ,030304 developmental biology ,Mathematics - Abstract
We recently reported the existence of fluctuations in neural signals that may permit neurons to code multiple simultaneous stimuli sequentially across time. This required deploying a novel statistical approach to permit investigation of neural activity at the scale of individual trials. Here we present tests using synthetic data to assess the sensitivity and specificity of this analysis. We fabricated datasets to match each of several potential response patterns derived from single-stimulus response distributions. In particular, we simulated dual stimulus trial spike counts that reflected fluctuating mixtures of the single stimulus spike counts, stable intermediate averages, single stimulus winner-take-all, or response distributions that were outside the range defined by the single stimulus responses (such as summation or suppression). We then assessed how well the analysis recovered the correct response pattern as a function of the number of simulated trials and the difference between the simulated responses to each "stimulus" alone. We found excellent recovery of the mixture, intermediate, and outside categories (>97% percent correct), and good recovery of the single/winner-take-all category (>90% correct) when the number of trials was >20 and the single-stimulus response rates were 50Hz and 20Hz respectively. Both larger numbers of trials and greater separation between the single stimulus firing rates improved categorization accuracy. These results provide a benchmark, and guidelines for data collection, for use of this method to investigate coding of multiple items at the individual-trial time scale., Accepted for publication in Neurons, Behavior, Data analysis, and Theory
- Published
- 2019
- Full Text
- View/download PDF
22. Bayesian Analysis of Dynamic Linear Topic Models
- Author
-
Brian E. Howard, David Banks, Surya T. Tokdar, and Chris Glynn
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Topic model ,MCMC ,Computer science ,Bayesian probability ,Inference ,Machine Learning (stat.ML) ,Machine learning ,computer.software_genre ,Pólya-Gamma ,01 natural sciences ,Machine Learning (cs.LG) ,Methodology (stat.ME) ,010104 statistics & probability ,symbols.namesake ,Statistics - Machine Learning ,0502 economics and business ,0101 mathematics ,topic model ,Statistics - Methodology ,050205 econometrics ,business.industry ,Applied Mathematics ,dynamic linear model ,05 social sciences ,Probabilistic logic ,Linear model ,Sampling (statistics) ,Markov chain Monte Carlo ,Dynamic topic model ,Computer Science - Learning ,symbols ,Artificial intelligence ,business ,computer - Abstract
Discovering temporal evolution of themes from a time-stamped collection of text poses a challenging statistical learning problem. Dynamic topic models offer a probabilistic modeling framework to decompose a corpus of text documents into “topics”, i.e., probability distributions over vocabulary terms, while simultaneously learning the temporal dynamics of the relative prevalence of these topics. We extend the dynamic topic model of Blei and Lafferty (2006) by fusing its multinomial factor model on topics with dynamic linear models that account for time trends and seasonality in topic prevalence. A Markov chain Monte Carlo (MCMC) algorithm that utilizes Polya-Gamma data augmentation is developed for posterior sampling. Conditional independencies in the model and sampling are made explicit, and our MCMC algorithm is parallelized where possible to allow for inference in large corpora. Our model and inference algorithm are validated with multiple synthetic examples, and we consider the applied problem of modeling trends in real estate listings from the housing website Zillow. We demonstrate in synthetic examples that sharing information across documents is critical for accurately estimating document-specific topic proportions. Analysis of the Zillow corpus demonstrates that the method is able to learn seasonal patterns and locally linear trends in topic prevalence.
- Published
- 2019
23. Evidence for time division multiplexing: Single neurons may encode simultaneous stimuli by switching between activity patterns
- Author
-
Jennifer M. Groh, Valeria C. Caruso, Shawn M. Willett, Surya T. Tokdar, Rolando Estrada, Akinori F. Ebihara, Winrich A. Freiwald, Jeff T. Mohl, Chris Glynn, Jungah Lee, and Azeem Zaman
- Subjects
Inferior colliculus ,0303 health sciences ,Visual perception ,Interleaving ,Local field potential ,Biology ,ENCODE ,Inferotemporal cortex ,03 medical and health sciences ,0302 clinical medicine ,Alternation (linguistics) ,Neuroscience ,030217 neurology & neurosurgery ,030304 developmental biology - Abstract
How the brain preserves information about multiple simultaneous items is poorly understood. We report that single neurons can represent multiple different stimuli by interleaving different signals across time. We record single units in an auditory region, the inferior colliculus, while monkeys localize 1 or 2 simultaneous sounds. During dual-sound trials, we find that some neurons fluctuate between firing rates observed for each single sound, either on a whole-trial or on a sub-trial timescale. These fluctuations are correlated in pairs of neurons, can be predicted by the state of local field potentials prior to sound onset, and, in one monkey, can predict which sound will be reported first. We find corroborating evidence of fluctuating activity patterns in a separate data set involving responses of inferotemporal cortex neurons to multiple visual stimuli. Alternation between activity patterns corresponding to each of multiple items may therefore be a general strategy to enhance the brain processing capacity, potentially linking such disparate phenomena as variable neural firing, neural oscillations, and limits in attentional/memory capacity.
- Published
- 2017
- Full Text
- View/download PDF
24. Adaptive Bayesian multivariate density estimation with Dirichlet mixtures
- Author
-
Surya T. Tokdar, Weining Shen, and Subhashis Ghosal
- Subjects
Statistics and Probability ,Mathematical optimization ,Applied Mathematics ,General Mathematics ,Inverse-Wishart distribution ,Concentration parameter ,Matrix t-distribution ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Agricultural and Biological Sciences (miscellaneous) ,Multivariate kernel density estimation ,Dirichlet distribution ,Normal-Wishart distribution ,symbols.namesake ,Generalized Dirichlet distribution ,Prior probability ,symbols ,FOS: Mathematics ,Applied mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Mathematics - Abstract
We show that rate-adaptive multivariate density estimation can be performed using Bayesian methods based on Dirichlet mixtures of normal kernels with a prior distribution on the kernel's covariance matrix parameter. We derive sufficient conditions on the prior specification that guarantee convergence to a true density at a rate that is optimal minimax for the smoothness class to which the true density belongs. No prior knowledge of smoothness is assumed. The sufficient conditions are shown to hold for the Dirichlet location mixture of normals prior with a Gaussian base measure and an inverse-Wishart prior on the covariance matrix parameter. Locally H\"older smoothness classes and their anisotropic extensions are considered. Our study involves several technical novelties, including sharp approximation of finitely differentiable multivariate densities by normal mixtures and a new sieve on the space of such densities., Comment: 29 pages
- Published
- 2013
- Full Text
- View/download PDF
25. Efficient Gaussian process regression for large datasets
- Author
-
David B. Dunson, Surya T. Tokdar, and Anjishnu Banerjee
- Subjects
Statistics and Probability ,Applied Mathematics ,General Mathematics ,Dimensionality reduction ,Random projection ,computer.software_genre ,Agricultural and Biological Sciences (miscellaneous) ,Article ,Nonparametric regression ,symbols.namesake ,Data point ,symbols ,Data mining ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Bayesian linear regression ,Spatial analysis ,computer ,Gaussian process ,Subspace topology ,Mathematics - Abstract
Gaussian processes are widely used in nonparametric regression, classification and spatiotemporal modelling, facilitated in part by a rich literature on their theoretical properties. However, one of their practical limitations is expensive computation, typically on the order of n-super-3 where n is the number of data points, in performing the necessary matrix inversions. For large datasets, storage and processing also lead to computational bottlenecks, and numerical stability of the estimates and predicted values degrades with increasing n. Various methods have been proposed to address these problems, including predictive processes in spatial data analysis and the subset-of-regressors technique in machine learning. The idea underlying these approaches is to use a subset of the data, but this raises questions concerning sensitivity to the choice of subset and limitations in estimating fine-scale structure in regions that are not well covered by the subset. Motivated by the literature on compressive sensing, we propose an alternative approach that involves linear projection of all the data points onto a lower-dimensional subspace. We demonstrate the superiority of this approach from a theoretical perspective and through simulated and real data examples. Copyright 2013, Oxford University Press.
- Published
- 2013
- Full Text
- View/download PDF
26. Bayesian Latent Factor Regression for Functional and Longitudinal Data
- Author
-
Silvia Montagna, Surya T. Tokdar, David B. Dunson, and Brian Neelon
- Subjects
Statistics and Probability ,Biometry ,Latent variable ,Article ,General Biochemistry, Genetics and Molecular Biology ,Structural equation modeling ,Statistics::Machine Learning ,symbols.namesake ,Statistics ,Econometrics ,Computer Simulation ,Longitudinal Studies ,Latent variable model ,Mathematics ,Models, Statistical ,General Immunology and Microbiology ,Applied Mathematics ,Bayes Theorem ,Regression analysis ,General Medicine ,Random effects model ,Latent class model ,Data Interpretation, Statistical ,symbols ,Regression Analysis ,Epidemiologic Methods ,General Agricultural and Biological Sciences ,Algorithms ,Factor regression model ,Gibbs sampling - Abstract
In studies involving functional data, it is commonly of interest to model the impact of predictors on the distribution of the curves, allowing flexible effects on not only the mean curve but also the distribution about the mean. Characterizing the curve for each subject as a linear combination of a high-dimensional set of potential basis functions, we place a sparse latent factor regression model on the basis coefficients. We induce basis selection by choosing a shrinkage prior that allows many of the loadings to be close to zero. The number of latent factors is treated as unknown through a highly-efficient, adaptive-blocked Gibbs sampler. Predictors are included on the latent variables level, while allowing different predictors to impact different latent factors. This model induces a framework for functional response regression in which the distribution of the curves is allowed to change flexibly with predictors. The performance is assessed through simulation studies and the methods are applied to data on blood pressure trajectories during pregnancy.
- Published
- 2012
- Full Text
- View/download PDF
27. Semiparametric inference in mixture models with predictive recursion marginal likelihood
- Author
-
Surya T. Tokdar and Ryan Martin
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Applied Mathematics ,General Mathematics ,Recursion (computer science) ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Density estimation ,Maximum likelihood sequence estimation ,Mixture model ,Agricultural and Biological Sciences (miscellaneous) ,Marginal likelihood ,Methodology (stat.ME) ,Dirichlet process ,Bayes' theorem ,ComputingMethodologies_PATTERNRECOGNITION ,FOS: Mathematics ,Econometrics ,Statistics::Methodology ,Applied mathematics ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Likelihood function ,Statistics - Methodology ,Mathematics - Abstract
Predictive recursion is an accurate and computationally efficient algorithm for nonparametric estimation of mixing densities in mixture models. In semiparametric mixture models, however, the algorithm fails to account for any uncertainty in the additional unknown structural parameter. As an alternative to existing profile likelihood methods, we treat predictive recursion as a filter approximation by fitting a fully Bayes model, whereby an approximate marginal likelihood of the structural parameter emerges and can be used for inference. We call this the predictive recursion marginal likelihood. Convergence properties of predictive recursion under model misspecification also lead to an attractive construction of this new procedure. We show pointwise convergence of a normalized version of this marginal likelihood function. Simulations compare the performance of this new approach with that of existing profile likelihood methods and with Dirichlet process mixtures in density estimation. Mixed-effects models and an empirical Bayes multiple testing application in time series analysis are also considered. Copyright 2011, Oxford University Press.
- Published
- 2011
- Full Text
- View/download PDF
28. Importance sampling: a review
- Author
-
Robert E. Kass and Surya T. Tokdar
- Subjects
Statistics and Probability ,Mathematical optimization ,Computer science ,business.industry ,Rejection sampling ,Monte Carlo method ,Slice sampling ,Sampling (statistics) ,Markov chain Monte Carlo ,Machine learning ,computer.software_genre ,symbols.namesake ,symbols ,Monte Carlo integration ,Monte Carlo method in statistical physics ,Artificial intelligence ,business ,computer ,Importance sampling - Abstract
We provide a short overview of importance sampling—a popular sampling tool used for Monte Carlo computing. We discuss its mathematical foundation and properties that determine its accuracy in Monte Carlo approximations. We review the fundamental developments in designing efficient importance sampling (IS) for practical use. This includes parametric approximation with optimization-based adaptation, sequential sampling with dynamic adaptation through resampling and population-based approaches that make use of Markov chain sampling. Copyright © 2009 John Wiley & Sons, Inc. For further resources related to this article, please visit the WIREs website.
- Published
- 2009
- Full Text
- View/download PDF
29. Towards a Faster Implementation of Density Estimation With Logistic Gaussian Process Priors
- Author
-
Surya T. Tokdar
- Subjects
Statistics and Probability ,business.industry ,Computation ,Gaussian ,Pattern recognition ,Markov chain Monte Carlo ,Density estimation ,symbols.namesake ,Bayes' theorem ,Prior probability ,symbols ,Discrete Mathematics and Combinatorics ,Artificial intelligence ,Imputation (statistics) ,Statistics, Probability and Uncertainty ,business ,Gaussian process ,Algorithm ,Mathematics - Abstract
A novel method is proposed to compute the Bayes estimate for a logistic Gaussian process prior for density estimation. The method gains speed by drawing samples from the posterior of a finite-dimensional surrogate prior, which is obtained by imputation of the underlying Gaussian process. We establish that imputation results in quite accurate computation. Simulation studies show that accuracy and high speed can be combined. This fact, along with known flexibility of the logistic Gaussian priors for modeling smoothness and recent results on their large support, makes these priors and the resulting density estimate very attractive.
- Published
- 2007
- Full Text
- View/download PDF
30. Posterior consistency of logistic Gaussian process priors in density estimation
- Author
-
Jayanta K. Ghosh and Surya T. Tokdar
- Subjects
Statistics and Probability ,Weak consistency ,Covariance function ,Applied Mathematics ,Posterior probability ,Strong consistency ,Covariance ,symbols.namesake ,Kernel method ,Statistics ,Prior probability ,symbols ,Applied mathematics ,Statistics, Probability and Uncertainty ,Gaussian process ,Mathematics - Abstract
We establish weak and strong posterior consistency of Gaussian process priors studied by Lenk [1988. The logistic normal distribution for Bayesian, nonparametric, predictive densities. J. Amer. Statist. Assoc. 83 (402), 509–516] for density estimation. Weak consistency is related to the support of a Gaussian process in the sup-norm topology which is explicitly identified for many covariance kernels. In fact we show that this support is the space of all continuous functions when the usual covariance kernels are chosen and an appropriate prior is used on the smoothing parameters of the covariance kernel. We then show that a large class of Gaussian process priors achieve weak as well as strong posterior consistency (under some regularity conditions) at true densities that are either continuous or piecewise continuous. © 2005 Elsevier B.V. All rights reserved.
- Published
- 2007
- Full Text
- View/download PDF
31. Posterior consistency in conditional distribution estimation
- Author
-
David B. Dunson, Debdeep Pati, and Surya T. Tokdar
- Subjects
Continuous stochastic process ,Statistics and Probability ,Numerical Analysis ,05 social sciences ,Conditional probability distribution ,01 natural sciences ,Article ,Dirichlet process ,010104 statistics & probability ,Monotone polygon ,Consistency (statistics) ,0502 economics and business ,Statistics ,Prior probability ,Applied mathematics ,Uncountable set ,Differentiable function ,0101 mathematics ,Statistics, Probability and Uncertainty ,050205 econometrics ,Mathematics - Abstract
A wide variety of priors have been proposed for nonparametric Bayesian estimation of conditional distributions, and there is a clear need for theorems providing conditions on the prior for large support, as well as posterior consistency. Estimation of an uncountable collection of conditional distributions across different regions of the predictor space is a challenging problem, which differs in some important ways from density and mean regression estimation problems. Defining various topologies on the space of conditional distributions, we provide sufficient conditions for posterior consistency focusing on a broad class of priors formulated as predictor-dependent mixtures of Gaussian kernels. This theory is illustrated by showing that the conditions are satisfied for a class of generalized stick-breaking process mixtures in which the stick-breaking lengths are monotone, differentiable functions of a continuous stochastic process. We also provide a set of sufficient conditions for the case where stick-breaking lengths are predictor independent, such as those arising from a fixed Dirichlet process prior.
- Published
- 2014
32. Minimax-optimal nonparametric regression in high dimensions
- Author
-
Surya T. Tokdar and Yun Yang
- Subjects
Statistics and Probability ,model selection ,Statistics::Theory ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Combinatorics ,Statistics::Machine Learning ,Adaptive estimation ,Kriging ,62G08 ,Additive function ,minimax risk ,FOS: Mathematics ,Statistics::Methodology ,Mathematics ,Smoothness (probability theory) ,62C20 ,Function (mathematics) ,Binary logarithm ,Minimax ,Regression ,high-dimensional regression ,Nonparametric regression ,nonparametric regression ,60G15 ,Statistics, Probability and Uncertainty - Abstract
Minimax $L_2$ risks for high-dimensional nonparametric regression are derived under two sparsity assumptions: (1) the true regression surface is a sparse function that depends only on $d=O(\log n)$ important predictors among a list of $p$ predictors, with $\log p=o(n)$; (2) the true regression surface depends on $O(n)$ predictors but is an additive function where each additive component is sparse but may contain two or more interacting predictors and may have a smoothness level different from other components. For either modeling assumption, a practicable extension of the widely used Bayesian Gaussian process regression method is shown to adaptively attain the optimal minimax rate (up to $\log n$ terms) asymptotically as both $n,p\to\infty$ with $\log p=o(n)$., Comment: Published at http://dx.doi.org/10.1214/14-AOS1289 in the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org)
- Published
- 2014
- Full Text
- View/download PDF
33. Simultaneous Linear Quantile Regression: A Semiparametric Bayesian Approach
- Author
-
Joseph B. Kadane and Surya T. Tokdar
- Subjects
Monotone Curves ,Statistics and Probability ,Statistics::Theory ,Multivariate statistics ,Mathematical optimization ,Applied Mathematics ,Univariate ,Linear Quantile Regression ,Bayesian Nonparametric Models ,Statistics::Computation ,Quantile regression ,Bayesian multivariate linear regression ,Covariate ,Bayesian Inference ,Statistics::Methodology ,Bayesian linear regression ,Gaussian Processes ,Bayesian average ,Joint Quantile Model ,Mathematics ,Quantile - Abstract
We introduce a semi-parametric Bayesian framework for a simultaneous analysis of linear quantile regression models. A simultaneous analysis is essential to attain the true potential of the quantile regression framework, but is computa- tionally challenging due to the associated monotonicity constraint on the quantile curves. For a univariate covariate, we present a simpler equivalent characterization of the monotonicity constraint through an interpolation of two monotone curves. The resulting formulation leads to a tractable likelihood function and is embedded within a Bayesian framework where the two monotone curves are modeled via lo- gistic transformations of a smooth Gaussian process. A multivariate extension is suggested by combining the full support univariate model with a linear projection of the predictors. The resulting single-index model remains easy to flt and provides substantial and measurable improvement over the flrst order linear heteroscedastic model. Two illustrative applications of the proposed method are provided.
- Published
- 2012
- Full Text
- View/download PDF
34. Impact of Beliefs About Atlantic Tropical Cyclone Detection on Conclusions About Trends in Tropical Cyclone Numbers
- Author
-
Joseph B. Kadane, Iris Grossmann, Anne-Sophie Charest, Surya T. Tokdar, and Mitchell J. Small
- Subjects
Statistics and Probability ,HURDAT ,Applied Mathematics ,Climatology ,tropical cyclone data ,Environmental science ,tropical cyclone detection ,Tropical cyclone ,Atlantic tropical cyclones - Abstract
Whether the number of tropical cyclones (TCs) has increased in the last 150 years has become a matter of intense debate. We investigate the efiects of beliefs about TC detection capacities in the North Atlantic on trends in TC num- bers since the 1870s. While raw data show an increasing trend of TC counts, the capability to detect TCs and to determine intensities and changes in intensity has also increased dramatically over the same period. We present a model of TC activ- ity that allows investigating the relationship between what one believes about the increase in detection and what one believes about TC trends. Previous work has used assumptions on TC tracks, detection capacities or the relationship between TC activity and various climate parameters to provide estimates of year-by-year missed TCs. These estimates and the associated conclusions about trends cover a wide range of possibilities. We build on previous work to investigate the sensitivity of these conclusions to the assumed priors about detection. Our analysis shows that any inference on TC count trends is strongly sensitive to one's speciflcation of prior beliefs about TC detection. Overall, we regard the evidence on the trend in North Atlantic TC numbers to be ambiguous.
- Published
- 2011
- Full Text
- View/download PDF
35. A nonparametric empirical Bayes framework for large-scale multiple testing
- Author
-
Ryan Martin and Surya T. Tokdar
- Subjects
Statistics and Probability ,FOS: Computer and information sciences ,Computer science ,Breast Neoplasms ,Statistics - Computation ,Methodology (stat.ME) ,Bayes' theorem ,Mixture distribution ,Humans ,Statistics::Methodology ,Computer Simulation ,Computation (stat.CO) ,Statistics - Methodology ,Oligonucleotide Array Sequence Analysis ,Likelihood Functions ,Leukemia ,Models, Statistical ,Nonparametric statistics ,Recursion (computer science) ,Bayes Theorem ,General Medicine ,Mixture model ,Thresholding ,Marginal likelihood ,Data Interpretation, Statistical ,Multiple comparisons problem ,Female ,Statistics, Probability and Uncertainty ,Algorithm ,Algorithms - Abstract
We propose a flexible and identifiable version of the two-groups model, motivated by hierarchical Bayes considerations, that features an empirical null and a semiparametric mixture model for the non-null cases. We use a computationally efficient predictive recursion marginal likelihood procedure to estimate the model parameters, even the nonparametric mixing distribution. This leads to a nonparametric empirical Bayes testing procedure, which we call PRtest, based on thresholding the estimated local false discovery rates. Simulations and real-data examples demonstrate that, compared to existing approaches, PRtest's careful handling of the non-null density can give a much better fit in the tails of the mixture distribution which, in turn, can lead to more realistic conclusions., 18 pages, 4 figures, 3 tables
- Published
- 2011
36. Bayesian density regression with logistic Gaussian process and subspace projection
- Author
-
Jayanta K. Ghosh, Surya T. Tokdar, and Yu Zhu
- Subjects
Statistics and Probability ,Computer science ,business.industry ,Applied Mathematics ,Dimensionality reduction ,Pattern recognition ,Feature selection ,Regression analysis ,Dimension Reduction ,symbols.namesake ,Semiparametric Model ,symbols ,Bayesian Inference ,Markov Chain Monte Carlo ,Artificial intelligence ,Gaussian Process ,Posterior Consistency ,Projection (set theory) ,Bayesian linear regression ,business ,Gaussian process ,Subspace topology ,Gibbs sampling - Abstract
We develop a novel Bayesian density regression model based on logistic Gaussian processes and subspace projection. Logistic Gaussian processes provide an attractive alternative to the popular stick-breaking processes for modeling a family of conditional densities that vary smoothly in the conditioning variable. Subspace projection offers dimension reduction of predictors through multiple linear combinations, offering an alternative to the zeroing out theme of variable selection. We illustrate that logistic Gaussian processes and subspace projection combine well to produce a computationally tractable and theoretically sound density regression procedure that offers good out of sample prediction, accurate estimation of subspace projection and satisfactory estimation of subspace dimensionality. We also demonstrate that subspace projection may lead to better prediction than variable selection when predictors are well chosen and possibly dependent on each other, each having a moderate influence on the response.
- Published
- 2010
- Full Text
- View/download PDF
37. Consistency of a recursive estimate of mixing distributions
- Author
-
Surya T. Tokdar, Jayanta K. Ghosh, and Ryan Martin
- Subjects
Statistics and Probability ,Approximation theory ,Weak topology ,Monte Carlo method ,Mathematics - Statistics Theory ,Density estimation ,Statistics Theory (math.ST) ,Mixture model ,recursive density estimation ,62G07 (Primary) 62G05, 62L20 (Secondary) ,Convergence of random variables ,Consistency (statistics) ,FOS: Mathematics ,62G07 ,Probability distribution ,Applied mathematics ,62G05 ,Statistics, Probability and Uncertainty ,62L20 ,Mixture models ,empirical Bayes ,Mathematics - Abstract
Mixture models have received considerable attention recently and Newton [Sankhy\={a} Ser. A 64 (2002) 306--322] proposed a fast recursive algorithm for estimating a mixing distribution. We prove almost sure consistency of this recursive estimate in the weak topology under mild conditions on the family of densities being mixed. This recursive estimate depends on the data ordering and a permutation-invariant modification is proposed, which is an average of the original over permutations of the data sequence. A Rao--Blackwell argument is used to prove consistency in probability of this alternative estimate. Several simulations are presented, comparing the finite-sample performance of the recursive estimate and a Monte Carlo approximation to the permutation-invariant alternative along with that of the nonparametric maximum likelihood estimate and a nonparametric Bayes estimate., Comment: Published in at http://dx.doi.org/10.1214/08-AOS639 the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org)
- Published
- 2009
38. Asymptotic properties of predictive recursion: Robustness and rate of convergence
- Author
-
Ryan Martin and Surya T. Tokdar
- Subjects
Statistics and Probability ,Mathematical optimization ,Weak convergence ,Nonparametric statistics ,Recursion (computer science) ,Kullback-Leibler projection ,Density estimation ,Mixture model ,Minimax ,Rate of convergence ,density estimation ,62G07 ,Applied mathematics ,62G05 ,62G35 ,Statistics, Probability and Uncertainty ,mixture models ,Almost supermartingale ,Mixing (physics) ,empirical Bayes ,62G20 ,Mathematics - Abstract
Here we explore general asymptotic properties of Predictive Recursion (PR) for nonparametric estimation of mixing distributions. We prove that, when the mixture model is mis-specified, the estimated mixture convergesalmost surely in total variation to the mixture that minimizes the Kullback-Leiblerdivergence,and a bound on the (Hellingercontrast)rate of convergence is obtained. Simulations suggest that this rate is nearly sharp in a minimax sense. Moreover, when the model is identifiable, almost sure weak convergence of the mixing distribution estimate follows. PR assumes that the support of the mixing distribution is known. To remove this requirement, we propose a generalization that incorporates a sequence of supports, increasing with the sample size, that combines the efficiency of PR with the flexibility of mixture sieves. Under mild conditions, we obtain a bound on the rate of convergence of these new estimates. AMS 2000 subject classifications: Primary 62G20; secondary 62G05, 62G07, 62G35. Keywords and phrases: Almost supermartingale, density estimation, empirical Bayes, Kullback-Leibler projection, mixture models.
- Published
- 2009
39. A comparison of the Benjamini-Hochberg procedure with some Bayesian rules for multiple testing
- Author
-
Małgorzata Bogdan, Surya T. Tokdar, and Jayanta K. Ghosh
- Subjects
False discovery rate ,Bayesian multiple testing ,62C12 ,nonparametric Bayes ,62C10 ,Mathematical statistics ,Bayesian probability ,Inference ,Mathematics - Statistics Theory ,Statistics Theory (math.ST) ,Bayes' theorem ,Benjamini-Hochberg Procedure ,Multiple comparisons problem ,Statistics ,FOS: Mathematics ,62C10 (Primary) 62C12 (Secondary) ,Algorithm ,empirical Bayes ,Parametric statistics ,Mathematics - Abstract
In the spirit of modeling inference for microarrays as multiple testing for sparse mixtures, we present a similar approach to a simplified version of quantitative trait loci (QTL) mapping. Unlike in case of microarrays, where the number of tests usually reaches tens of thousands, the number of tests performed in scans for QTL usually does not exceed several hundreds. However, in typical cases, the sparsity $p$ of significant alternatives for QTL mapping is in the same range as for microarrays. For methodological interest, as well as some related applications, we also consider non-sparse mixtures. Using simulations as well as theoretical observations we study false discovery rate (FDR), power and misclassification probability for the Benjamini-Hochberg (BH) procedure and its modifications, as well as for various parametric and nonparametric Bayes and Parametric Empirical Bayes procedures. Our results confirm the observation of Genovese and Wasserman (2002) that for small p the misclassification error of BH is close to optimal in the sense of attaining the Bayes oracle. This property is shared by some of the considered Bayes testing rules, which in general perform better than BH for large or moderate $p$'s., Published in at http://dx.doi.org/10.1214/193940307000000158 the IMS Collections (http://www.imstat.org/publications/imscollections.htm) by the Institute of Mathematical Statistics (http://www.imstat.org)
- Published
- 2008
40. Convergence and Consistency of Newton's Algorithm for Estimating Mixing Distribution
- Author
-
Surya T. Tokdar and Jayanta K. Ghosh
- Subjects
Mathematical optimization ,Distribution (mathematics) ,Consistency (statistics) ,Convergence (routing) ,Applied mathematics ,Mixing (physics) ,Mathematics - Published
- 2006
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.