4,198 results on '"Ding, Peng"'
Search Results
2. When is it worthwhile to jackknife? Breaking the quadratic barrier for Z-estimators
- Author
-
Lin, Licong, Su, Fangzhou, Mou, Wenlong, Ding, Peng, and Wainwright, Martin
- Subjects
Mathematics - Statistics Theory ,Statistics - Methodology - Abstract
Resampling methods are especially well-suited to inference with estimators that provide only "black-box'' access. Jackknife is a form of resampling, widely used for bias correction and variance estimation, that is well-understood under classical scaling where the sample size $n$ grows for a fixed problem. We study its behavior in application to estimating functionals using high-dimensional $Z$-estimators, allowing both the sample size $n$ and problem dimension $d$ to diverge. We begin showing that the plug-in estimator based on the $Z$-estimate suffers from a quadratic breakdown: while it is $\sqrt{n}$-consistent and asymptotically normal whenever $n \gtrsim d^2$, it fails for a broad class of problems whenever $n \lesssim d^2$. We then show that under suitable regularity conditions, applying a jackknife correction yields an estimate that is $\sqrt{n}$-consistent and asymptotically normal whenever $n\gtrsim d^{3/2}$. This provides strong motivation for the use of jackknife in high-dimensional problems where the dimension is moderate relative to sample size. We illustrate consequences of our general theory for various specific $Z$-estimators, including non-linear functionals in linear models; generalized linear models; and the inverse propensity score weighting (IPW) estimate for the average treatment effect, among others.
- Published
- 2024
3. Identifying and bounding the probability of necessity for causes of effects with ordinal outcomes
- Author
-
Zhang, Chao, Geng, Zhi, Li, Wei, and Ding, Peng
- Subjects
Mathematics - Statistics Theory - Abstract
Although the existing causal inference literature focuses on the forward-looking perspective by estimating effects of causes, the backward-looking perspective can provide insights into causes of effects. In backward-looking causal inference, the probability of necessity measures the probability that a certain event is caused by the treatment given the observed treatment and outcome. Most existing results focus on binary outcomes. Motivated by applications with ordinal outcomes, we propose a general definition of the probability of necessity. However, identifying the probability of necessity is challenging because it involves the joint distribution of the potential outcomes. We propose a novel assumption of monotonic incremental treatment effect to identify the probability of necessity with ordinal outcomes. We also discuss the testable implications of this key identification assumption. When it fails, we derive explicit formulas of the sharp large-sample bounds on the probability of necessity.
- Published
- 2024
4. Asymptotic theory for the quadratic assignment procedure
- Author
-
Shi, Lei and Ding, Peng
- Subjects
Statistics - Methodology ,62K15, 62J05, 62G05 - Abstract
The quadratic assignment procedure (QAP) is a popular tool for analyzing network data in medical and social sciences. To test the association between two network measurements represented by two symmetric matrices, QAP calculates the $p$-value by permuting the units, or equivalently, by simultaneously permuting the rows and columns of one matrix. Its extension to the regression setting, known as the multiple regression QAP, has also gained popularity, especially in psychometrics. However, the statistics theory for QAP has not been fully established in the literature. We fill the gap in this paper. We formulate the network models underlying various QAPs. We derive (a) the asymptotic sampling distributions of some canonical test statistics and (b) the corresponding asymptotic permutation distributions induced by QAP under strong and weak null hypotheses. Task (a) relies on applying the theory of U-statistics, and task (b) relies on applying the theory of double-indexed permutation statistics. The combination of tasks (a) and (b) provides a relatively complete picture of QAP. Overall, our asymptotic theory suggests that using properly studentized statistics in QAP is a robust choice in that it is finite-sample exact under the strong null hypothesis and preserves the asymptotic type one error rate under the weak null hypothesis., Comment: 74 pages, 2 figures
- Published
- 2024
5. With random regressors, least squares inference is robust to correlated errors with unknown correlation structure
- Author
-
Zhang, Zifeng, Ding, Peng, Zhou, Wen, and Wang, Haonan
- Subjects
Mathematics - Statistics Theory ,Statistics - Methodology - Abstract
Linear regression is arguably the most widely used statistical method. With fixed regressors and correlated errors, the conventional wisdom is to modify the variance-covariance estimator to accommodate the known correlation structure of the errors. We depart from the literature by showing that with random regressors, linear regression inference is robust to correlated errors with unknown correlation structure. The existing theoretical analyses for linear regression are no longer valid because even the asymptotic normality of the least-squares coefficients breaks down in this regime. We first prove the asymptotic normality of the t statistics by establishing their Berry-Esseen bounds based on a novel probabilistic analysis of self-normalized statistics. We then study the local power of the corresponding t tests and show that, perhaps surprisingly, error correlation can even enhance power in the regime of weak signals. Overall, our results show that linear regression is applicable more broadly than the conventional theory suggests, and further demonstrate the value of randomization to ensure robustness of inference.
- Published
- 2024
6. Hallu-PI: Evaluating Hallucination in Multi-modal Large Language Models within Perturbed Inputs
- Author
-
Ding, Peng, Wu, Jingyu, Kuang, Jun, Ma, Dan, Cao, Xuezhi, Cai, Xunliang, Chen, Shi, Chen, Jiajun, and Huang, Shujian
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Multimedia - Abstract
Multi-modal Large Language Models (MLLMs) have demonstrated remarkable performance on various visual-language understanding and generation tasks. However, MLLMs occasionally generate content inconsistent with the given images, which is known as "hallucination". Prior works primarily center on evaluating hallucination using standard, unperturbed benchmarks, which overlook the prevalent occurrence of perturbed inputs in real-world scenarios-such as image cropping or blurring-that are critical for a comprehensive assessment of MLLMs' hallucination. In this paper, to bridge this gap, we propose Hallu-PI, the first benchmark designed to evaluate Hallucination in MLLMs within Perturbed Inputs. Specifically, Hallu-PI consists of seven perturbed scenarios, containing 1,260 perturbed images from 11 object types. Each image is accompanied by detailed annotations, which include fine-grained hallucination types, such as existence, attribute, and relation. We equip these annotations with a rich set of questions, making Hallu-PI suitable for both discriminative and generative tasks. Extensive experiments on 12 mainstream MLLMs, such as GPT-4V and Gemini-Pro Vision, demonstrate that these models exhibit significant hallucinations on Hallu-PI, which is not observed in unperturbed scenarios. Furthermore, our research reveals a severe bias in MLLMs' ability to handle different types of hallucinations. We also design two baselines specifically for perturbed scenarios, namely Perturbed-Reminder and Perturbed-ICL. We hope that our study will bring researchers' attention to the limitations of MLLMs when dealing with perturbed inputs, and spur further investigations to address this issue. Our code and datasets are publicly available at https://github.com/NJUNLP/Hallu-PI., Comment: Acccepted by ACM MM 2024, 14 pages, 11 figures, 9 tables
- Published
- 2024
7. Identification and multiply robust estimation of causal effects via instrumental variables from an auxiliary heterogeneous population
- Author
-
Li, Wei, Liu, Jiapeng, Ding, Peng, and Geng, Zhi
- Subjects
Statistics - Methodology - Abstract
Evaluating causal effects in a primary population of interest with unmeasured confounders is challenging. Although instrumental variables (IVs) are widely used to address unmeasured confounding, they may not always be available in the primary population. Fortunately, IVs might have been used in previous observational studies on similar causal problems, and these auxiliary studies can be useful to infer causal effects in the primary population, even if they represent different populations. However, existing methods often assume homogeneity or equality of conditional average treatment effects between the primary and auxiliary populations, which may be limited in practice. This paper aims to remove the homogeneity requirement and establish a novel identifiability result allowing for different conditional average treatment effects across populations. We also construct a multiply robust estimator that remains consistent despite partial misspecifications of the observed data model and achieves local efficiency if all nuisance models are correct. The proposed approach is illustrated through simulation studies. We finally apply our approach by leveraging data from lower income individuals with cigarette price as a valid IV to evaluate the causal effect of smoking on physical functional status in higher income group where strong IVs are not available.
- Published
- 2024
8. Factorial Difference-in-Differences
- Author
-
Xu, Yiqing, Zhao, Anqi, and Ding, Peng
- Subjects
Statistics - Methodology ,Economics - Econometrics - Abstract
In many social science applications, researchers use the difference-in-differences (DID) estimator to establish causal relationships, exploiting cross-sectional variation in a baseline factor and temporal variation in exposure to an event that presumably may affect all units. This approach, which we term factorial DID (FDID), differs from canonical DID in that it lacks a clean control group unexposed to the event after the event occurs. In this paper, we clarify FDID as a research design in terms of its data structure, feasible estimands, and identifying assumptions that allow the DID estimator to recover these estimands. We frame FDID as a factorial design with two factors: the baseline factor, denoted by $G$, and the exposure level to the event, denoted by $Z$, and define the effect modification and causal interaction as the associative and causal effects of $G$ on the effect of $Z$, respectively. We show that under the canonical no anticipation and parallel trends assumptions, the DID estimator identifies only the effect modification of $G$ in FDID, and propose an additional factorial parallel trends assumption to identify the causal interaction. Moreover, we show that the canonical DID research design can be reframed as a special case of the FDID research design with an additional exclusion restriction assumption, thereby reconciling the two approaches. We extend this framework to allow conditionally valid parallel trends assumptions and multiple time periods, and clarify assumptions required to justify regression analysis under FDID. We illustrate these findings with empirical examples from economics and political science, and provide recommendations for improving practice and interpretation under FDID.
- Published
- 2024
9. Entropy-Reinforced Planning with Large Language Models for Drug Discovery
- Author
-
Liu, Xuefeng, Tien, Chih-chan, Ding, Peng, Jiang, Songhao, and Stevens, Rick L.
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Quantitative Biology - Quantitative Methods ,Statistics - Machine Learning - Abstract
The objective of drug discovery is to identify chemical compounds that possess specific pharmaceutical properties toward a binding target. Existing large language models (LLMS) can achieve high token matching scores in terms of likelihood for molecule generation. However, relying solely on LLM decoding often results in the generation of molecules that are either invalid due to a single misused token, or suboptimal due to unbalanced exploration and exploitation as a consequence of the LLMs prior experience. Here we propose ERP, Entropy-Reinforced Planning for Transformer Decoding, which employs an entropy-reinforced planning algorithm to enhance the Transformer decoding process and strike a balance between exploitation and exploration. ERP aims to achieve improvements in multiple properties compared to direct sampling from the Transformer. We evaluated ERP on the SARS-CoV-2 virus (3CLPro) and human cancer cell target protein (RTCB) benchmarks and demonstrated that, in both benchmarks, ERP consistently outperforms the current state-of-the-art algorithm by 1-5 percent, and baselines by 5-10 percent, respectively. Moreover, such improvement is robust across Transformer models trained with different objectives. Finally, to further illustrate the capabilities of ERP, we tested our algorithm on three code generation benchmarks and outperformed the current state-of-the-art approach as well. Our code is publicly available at: https://github.com/xuefeng-cs/ERP., Comment: Published in ICML2024
- Published
- 2024
10. Sensitivity Analysis for the Test-Negative Design
- Author
-
Kundu, Soumyabrata, Ding, Peng, Li, Xinran, and Wang, Jingshu
- Subjects
Statistics - Methodology - Abstract
The test-negative design has become popular for evaluating the effectiveness of post-licensure vaccines using observational data. In addition to its logistical convenience on data collection, the design is also believed to control for the differential health-care-seeking behavior between vaccinated and unvaccinated individuals, which is an important while often unmeasured confounder between the vaccination and infection. Hence, the design has been employed routinely to monitor seasonal flu vaccines and more recently to measure the COVID-19 vaccine effectiveness. Despite its popularity, the design has been questioned, in particular about its ability to fully control for the unmeasured confounding. In this paper, we explore deviations from a perfect test-negative design, and propose various sensitivity analysis methods for estimating the effect of vaccination measured by the causal odds ratio on the subpopulation of individuals with good health-care-seeking behavior. We start with point identification of the causal odds ratio under a test-negative design, considering two forms of assumptions on the unmeasured confounder. These assumptions then lead to two approaches for conducting sensitivity analysis, addressing the influence of the unmeasured confounding in different ways. Specifically, one approach investigates partial control for unmeasured confounder in the test-negative design, while the other examines the impact of unmeasured confounder on both vaccination and infection. Furthermore, these approaches can be combined to provide narrower bounds on the true causal odds ratio, and can be further extended to sharpen the bounds by restricting the treatment effect heterogeneity. Finally, we apply the proposed methods to evaluate the effectiveness of COVID-19 vaccines using observational data from test-negative designs.
- Published
- 2024
11. MANGO: A Benchmark for Evaluating Mapping and Navigation Abilities of Large Language Models
- Author
-
Ding, Peng, Fang, Jiading, Li, Peng, Wang, Kangrui, Zhou, Xiaochen, Yu, Mo, Li, Jing, Walter, Matthew R., and Mei, Hongyuan
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning ,Computer Science - Robotics - Abstract
Large language models such as ChatGPT and GPT-4 have recently achieved astonishing performance on a variety of natural language processing tasks. In this paper, we propose MANGO, a benchmark to evaluate their capabilities to perform text-based mapping and navigation. Our benchmark includes 53 mazes taken from a suite of textgames: each maze is paired with a walkthrough that visits every location but does not cover all possible paths. The task is question-answering: for each maze, a large language model reads the walkthrough and answers hundreds of mapping and navigation questions such as "How should you go to Attic from West of House?" and "Where are we if we go north and east from Cellar?". Although these questions are easy to humans, it turns out that even GPT-4, the best-to-date language model, performs poorly at answering them. Further, our experiments suggest that a strong mapping and navigation ability would benefit large language models in performing relevant downstream tasks, such as playing textgames. Our MANGO benchmark will facilitate future research on methods that improve the mapping and navigation capabilities of language models. We host our leaderboard, data, code, and evaluation program at https://mango.ttic.edu and https://github.com/oaklight/mango/., Comment: COLM 2024 camera-ready
- Published
- 2024
12. Two-phase rejective sampling and its asymptotic properties
- Author
-
Yang, Shu and Ding, Peng
- Subjects
Statistics - Methodology - Abstract
Rejective sampling improves design and estimation efficiency of single-phase sampling when auxiliary information in a finite population is available. When such auxiliary information is unavailable, we propose to use two-phase rejective sampling (TPRS), which involves measuring auxiliary variables for the sample of units in the first phase, followed by the implementation of rejective sampling for the outcome in the second phase. We explore the asymptotic design properties of double expansion and regression estimators under TPRS. We show that TPRS enhances the efficiency of the double expansion estimator, rendering it comparable to a regression estimator. We further refine the design to accommodate varying importance of covariates and extend it to multi-phase sampling. We start with the theory for the population mean and then extend the theory to parameters defined by general estimating equations. Our asymptotic results for TPRS immediately cover the existing single-phase rejective sampling, under which the asymptotic theory has not been fully established.
- Published
- 2024
13. Mesh-Free Method for Static Analyses of Carbon Nanotube-Reinforced Composite Plates
- Author
-
Ding, Peng-chu, Guo, Qin-qiang, Chang, Li-wu, Xu, Jun-feng, Li, Zhen, Yan, Shi-heng, and Han, Dong
- Published
- 2024
- Full Text
- View/download PDF
14. Quantifying Individual Risk for Binary Outcome: Bounds and Inference
- Author
-
Wu, Peng, Ding, Peng, Geng, Zhi, and Liu, Yue
- Subjects
Statistics - Methodology - Abstract
Understanding treatment heterogeneity is crucial for reliable decision-making in treatment evaluation and selection. While the conditional average treatment effect (CATE) is commonly used to capture treatment heterogeneity induced by covariates and design individualized treatment policies, it remains an averaging metric within subpopulations. This limitation prevents it from unveiling individual-level risks, potentially leading to misleading results. This article addresses this gap by examining individual risk for binary outcomes, specifically focusing on the fraction negatively affected (FNA) conditional on covariates -- a metric assessing the percentage of individuals experiencing worse outcomes with treatment compared to control. Under the strong ignorability assumption, FNA is unidentifiable, and we find that previous bounds are wide and practically unattainable except in certain degenerate cases. By introducing a plausible positive correlation assumption for the potential outcomes, we obtain significantly improved bounds compared to previous studies. We show that even with a positive and statistically significant CATE, the lower bound on FNA can be positive, i.e., in the best-case scenario many units will be harmed if receiving treatment. We establish a nonparametric sensitivity analysis framework for FNA using the Pearson correlation coefficient as the sensitivity parameter, thereby exploring the relationships among the correlation coefficient, FNA, and CATE. We also present a practical and tractable method for selecting the range of correlation coefficients. Furthermore, we propose flexible estimators for refined FNA bounds and prove their consistency and asymptotic normality.
- Published
- 2024
15. Linear Model and Extensions
- Author
-
Ding, Peng
- Subjects
Statistics - Methodology ,Statistics - Applications - Abstract
I developed the lecture notes based on my ``Linear Model'' course at the University of California Berkeley over the past seven years. This book provides an intermediate-level introduction to the linear model. It balances rigorous proofs and heuristic arguments. This book provides R code to replicate all simulation studies and case studies.
- Published
- 2023
16. Joint prediction of the capacity and temperature of Li-ion batteries by using ConvLSTM Network
- Author
-
Wang, Dong, Li, Jian, Ding, Peng, and Yao, Ning
- Published
- 2024
- Full Text
- View/download PDF
17. On long-term fatigue damage estimation for a floating offshore wind turbine using a surrogate model
- Author
-
Liu, Ding Peng, Ferri, Giulio, Heo, Taemin, Marino, Enzo, and Manuel, Lance
- Subjects
Statistics - Applications - Abstract
This study is concerned with the estimation of long-term fatigue damage for a floating offshore wind turbine. With the ultimate goal of efficient evaluation of fatigue limit states for floating offshore wind turbine systems, a detailed computational framework is introduced and used to develop a surrogate model using Gaussian process regression. The surrogate model, at first, relies only on a small subset of representative sea states and, then, is supplemented by the evaluation of additional sea states that leads to efficient convergence and accurate prediction of fatigue damage. A 5-MW offshore wind turbine supported by a semi-submersible floating platform is selected to demonstrate the proposed framework. The fore-aft bending moment at the turbine tower base and the fairlead tension in the windward mooring line are used for evaluation. Metocean data provide information on joint statistics of the wind and wave along with their relative likelihoods for the installation site in the Mediterranean Sea, near the coast of Sicily. \textcolor{black}{A coupled frequency-domain model} provides needed power spectra for the desired response processes. The proposed approach offers an efficient and accurate alternative to the exhaustive evaluation of a larger number of sea states and, as such, avoids excessive response simulations.
- Published
- 2023
- Full Text
- View/download PDF
18. Covariate adjustment in randomized experiments with missing outcomes and covariates
- Author
-
Zhao, Anqi, Ding, Peng, and Li, Fan
- Subjects
Statistics - Methodology - Abstract
Covariate adjustment can improve precision in analyzing randomized experiments. With fully observed data, regression adjustment and propensity score weighting are asymptotically equivalent in improving efficiency over unadjusted analysis. When some outcomes are missing, we consider combining these two adjustment methods with inverse probability of observation weighting for handling missing outcomes, and show that the equivalence between the two methods breaks down. Regression adjustment no longer ensures efficiency gain over unadjusted analysis unless the true outcome model is linear in covariates or the outcomes are missing completely at random. Propensity score weighting, in contrast, still guarantees efficiency over unadjusted analysis, and including more covariates in adjustment never harms asymptotic efficiency. Moreover, we establish the value of using partially observed covariates to secure additional efficiency by the missingness indicator method, which imputes all missing covariates by zero and uses the union of the completed covariates and corresponding missingness indicators as the new, fully observed covariates. Based on these findings, we recommend using regression adjustment in combination with the missingness indicator method if the linear outcome model or missing complete at random assumption is plausible and using propensity score weighting with the missingness indicator method otherwise.
- Published
- 2023
19. A decorrelation method for general regression adjustment in randomized experiments
- Author
-
Su, Fangzhou, Mou, Wenlong, Ding, Peng, and Wainwright, Martin J.
- Subjects
Statistics - Methodology ,Mathematics - Statistics Theory - Abstract
We study regression adjustment with general function class approximations for estimating the average treatment effect in the design-based setting. Standard regression adjustment involves bias due to sample re-use, and this bias leads to behavior that is sub-optimal in the sample size, and/or imposes restrictive assumptions. Our main contribution is to introduce a novel decorrelation-based approach that circumvents these issues. We prove guarantees, both asymptotic and non-asymptotic, relative to the oracle functions that are targeted by a given regression adjustment procedure. We illustrate our method by applying it to various high-dimensional and non-parametric problems, exhibiting improved sample complexity and weakened assumptions relative to known approaches., Comment: Fangzhou Su and Wenlong Mou contributed equally to this work
- Published
- 2023
20. A Wolf in Sheep's Clothing: Generalized Nested Jailbreak Prompts can Fool Large Language Models Easily
- Author
-
Ding, Peng, Kuang, Jun, Ma, Dan, Cao, Xuezhi, Xian, Yunsen, Chen, Jiajun, and Huang, Shujian
- Subjects
Computer Science - Computation and Language - Abstract
Large Language Models (LLMs), such as ChatGPT and GPT-4, are designed to provide useful and safe responses. However, adversarial prompts known as 'jailbreaks' can circumvent safeguards, leading LLMs to generate potentially harmful content. Exploring jailbreak prompts can help to better reveal the weaknesses of LLMs and further steer us to secure them. Unfortunately, existing jailbreak methods either suffer from intricate manual design or require optimization on other white-box models, which compromises either generalization or efficiency. In this paper, we generalize jailbreak prompt attacks into two aspects: (1) Prompt Rewriting and (2) Scenario Nesting. Based on this, we propose ReNeLLM, an automatic framework that leverages LLMs themselves to generate effective jailbreak prompts. Extensive experiments demonstrate that ReNeLLM significantly improves the attack success rate while greatly reducing the time cost compared to existing baselines. Our study also reveals the inadequacy of current defense methods in safeguarding LLMs. Finally, we analyze the failure of LLMs defense from the perspective of prompt execution priority, and propose corresponding defense strategies. We hope that our research can catalyze both the academic community and LLMs developers towards the provision of safer and more regulated LLMs. The code is available at https://github.com/NJUNLP/ReNeLLM., Comment: Acccepted by NAACL 2024, 18 pages, 7 figures, 13 tables
- Published
- 2023
21. Balancing Weights for Causal Inference in Observational Factorial Studies
- Author
-
Yu, Ruoqi and Ding, Peng
- Subjects
Statistics - Methodology - Abstract
Many scientific questions in biomedical, environmental, and psychological research involve understanding the effects of multiple factors on outcomes. While factorial experiments are ideal for this purpose, randomized controlled treatment assignment is generally infeasible in many empirical studies. Therefore, investigators must rely on observational data, where drawing reliable causal inferences for multiple factors remains challenging. As the number of treatment combinations grows exponentially with the number of factors, some treatment combinations can be rare or missing by chance in observed data, further complicating factorial effects estimation. To address these challenges, we propose a novel weighting method tailored to observational studies with multiple factors. Our approach uses weighted observational data to emulate a randomized factorial experiment, enabling simultaneous estimation of the effects of multiple factors and their interactions. Our investigations reveal a crucial nuance: achieving balance among covariates, as in single-factor scenarios, is necessary but insufficient for unbiasedly estimating factorial effects; balancing the factors is also essential in multi-factor settings. Moreover, we extend our weighting method to handle missing treatment combinations in observed data. Finally, we study the asymptotic behavior of the new weighting estimators and propose a consistent variance estimator, providing reliable inferences on factorial effects in observational studies.
- Published
- 2023
22. Algebraic and Statistical Properties of the Ordinary Least Squares Interpolator
- Author
-
Shen, Dennis, Song, Dogyoon, Ding, Peng, and Sekhon, Jasjeet S.
- Subjects
Mathematics - Statistics Theory ,Computer Science - Machine Learning ,Statistics - Methodology - Abstract
Deep learning research has uncovered the phenomenon of benign overfitting for overparameterized statistical models, which has drawn significant theoretical interest in recent years. Given its simplicity and practicality, the ordinary least squares (OLS) interpolator has become essential to gain foundational insights into this phenomenon. While properties of OLS are well established in classical, underparameterized settings, its behavior in high-dimensional, overparameterized regimes is less explored (unlike for ridge or lasso regression) though significant progress has been made of late. We contribute to this growing literature by providing fundamental algebraic and statistical results for the minimum $\ell_2$-norm OLS interpolator. In particular, we provide algebraic equivalents of (i) the leave-$k$-out residual formula, (ii) Cochran's formula, and (iii) the Frisch-Waugh-Lovell theorem in the overparameterized regime. These results aid in understanding the OLS interpolator's ability to generalize and have substantive implications for causal inference. Under the Gauss-Markov model, we present statistical results such as an extension of the Gauss-Markov theorem and an analysis of variance estimation under homoskedastic errors for the overparameterized regime. To substantiate our theoretical contributions, we conduct simulations that further explore the stochastic properties of the OLS interpolator.
- Published
- 2023
23. Principal Stratification with Continuous Post-Treatment Variables: Nonparametric Identification and Semiparametric Estimation
- Author
-
Lu, Sizhu, Jiang, Zhichao, and Ding, Peng
- Subjects
Statistics - Methodology ,Mathematics - Statistics Theory - Abstract
Post-treatment variables often complicate causal inference. They appear in many scientific problems, including noncompliance, truncation by death, mediation, and surrogate endpoint evaluation. Principal stratification is a strategy to address these challenges by adjusting for the potential values of the post-treatment variables, defined as the principal strata. It allows for characterizing treatment effect heterogeneity across principal strata and unveiling the mechanism of the treatment's impact on the outcome related to post-treatment variables. However, the existing literature has primarily focused on binary post-treatment variables, leaving the case with continuous post-treatment variables largely unexplored. This gap persists due to the complexity of infinitely many principal strata, which present challenges to both the identification and estimation of causal effects. We fill this gap by providing nonparametric identification and semiparametric estimation theory for principal stratification with continuous post-treatment variables. We propose to use working models to approximate the underlying causal effect surfaces and derive the efficient influence functions of the corresponding model parameters. Based on the theory, we construct doubly robust estimators and implement them in an R package.
- Published
- 2023
24. Causal inference in network experiments: regression-based analysis and design-based properties
- Author
-
Gao, Mengsi and Ding, Peng
- Subjects
Economics - Econometrics - Abstract
Investigating interference or spillover effects among units is a central task in many social science problems. Network experiments are powerful tools for this task, which avoids endogeneity by randomly assigning treatments to units over networks. However, it is non-trivial to analyze network experiments properly without imposing strong modeling assumptions. Previously, many researchers have proposed sophisticated point estimators and standard errors for causal effects under network experiments. We further show that regression-based point estimators and standard errors can have strong theoretical guarantees if the regression functions and robust standard errors are carefully specified to accommodate the interference patterns under network experiments. We first recall a well-known result that the Hajek estimator is numerically identical to the coefficient from the weighted-least-squares fit based on the inverse probability of the exposure mapping. Moreover, we demonstrate that the regression-based approach offers three notable advantages: its ease of implementation, the ability to derive standard errors through the same weighted-least-squares fit, and the capacity to integrate covariates into the analysis, thereby enhancing estimation efficiency. Furthermore, we analyze the asymptotic bias of the regression-based network-robust standard errors. Recognizing that the covariance estimator can be anti-conservative, we propose an adjusted covariance estimator to improve the empirical coverage rates. Although we focus on regression-based point estimators and standard errors, our theory holds under the design-based framework, which assumes that the randomness comes solely from the design of network experiments and allows for arbitrary misspecification of the regression models.
- Published
- 2023
25. Real Effect or Bias? Best Practices for Evaluating the Robustness of Real-World Evidence through Quantitative Sensitivity Analysis for Unmeasured Confounding
- Author
-
Faries, Douglas, Gao, Chenyin, Zhang, Xiang, Hazlett, Chad, Stamey, James, Yang, Shu, Ding, Peng, Shan, Mingyang, Sheffield, Kristin, and Dreyer, Nancy
- Subjects
Statistics - Methodology ,Statistics - Applications ,Primary 62 - Abstract
The assumption of no unmeasured confounders is a critical but unverifiable assumption required for causal inference yet quantitative sensitivity analyses to assess robustness of real-world evidence remains underutilized. The lack of use is likely in part due to complexity of implementation and often specific and restrictive data requirements required for application of each method. With the advent of sensitivity analyses methods that are broadly applicable in that they do not require identification of a specific unmeasured confounder, along with publicly available code for implementation, roadblocks toward broader use are decreasing. To spur greater application, here we present a best practice guidance to address the potential for unmeasured confounding at both the design and analysis stages, including a set of framing questions and an analytic toolbox for researchers. The questions at the design stage guide the research through steps evaluating the potential robustness of the design while encouraging gathering of additional data to reduce uncertainty due to potential confounding. At the analysis stage, the questions guide researchers to quantifying the robustness of the observed result and providing researchers with a clearer indication of the robustness of their conclusions. We demonstrate the application of the guidance using simulated data based on a real-world fibromyalgia study, applying multiple methods from our analytic toolbox for illustration purposes., Comment: 16 pages which includes 5 figures
- Published
- 2023
26. Power and sample size calculations for rerandomization
- Author
-
Branson, Zach, Li, Xinran, and Ding, Peng
- Subjects
Mathematical Sciences ,Statistics ,Covariate balance ,Design-based inference ,Dispersive ordering ,Experimental design ,Treatment effect heterogeneity ,Numerical and Computational Mathematics ,Econometrics ,Statistics & Probability - Abstract
Summary: Power analyses are an important aspect of experimental design, because they help determine how experiments are implemented in practice. It is common to specify a desired level of power and compute the sample size necessary to obtain that power. Such calculations are well known for completely randomized experiments, but there can be many benefits to using other experimental designs. For example, it has recently been established that rerandomization, where subjects are randomized until covariate balance is obtained, increases the precision of causal effect estimators. This work establishes the power of rerandomized treatment-control experiments, thereby allowing for sample size calculators. We find the surprising result that, while power is often greater under rerandomization than complete randomization, the opposite can occur for very small treatment effects. The reason is that inference under rerandomization can be relatively more conservative, in the sense that it can have a lower Type-I error at the same nominal significance level, and this additional conservativeness adversely affects power. This surprising result is due to treatment effect heterogeneity, a quantity often ignored in power analyses. We find that heterogeneity increases power for large effect sizes, but decreases power for small effect sizes.
- Published
- 2024
27. Application of antithrombotic drugs and risk factor analysis in ICU patients with lower gastrointestinal bleeding from MIMIC-IV
- Author
-
Ding Peng and Huihong Zhai
- Subjects
Antithrombotic drugs ,Lower gastrointestinal bleeding ,MIMIC ,Diseases of the digestive system. Gastroenterology ,RC799-869 - Abstract
Abstract Objective This study aims to assess the effects of antithrombotic therapy on the outcomes of lower gastrointestinal bleeding (LGIB) in ICU patients, focusing on in-hospital mortality, rebleeding, and length of hospital and ICU stays. Method This retrospective observational study utilized the MIMIC-IV 2.2 database, which includes 513 ICU patients with LGIB. Result The in-hospital mortality rate was 7.6%, and the rebleeding rate was 11.1%. The average Oakland risk score among the study population was 22.54. Multivariate Cox regression analysis identified the use of antiplatelet drugs as an independent protective factor for in-hospital mortality (HR = 0.37, 95% CI 0.15–0.90, p = 0.029). Patients on anticoagulants experienced significantly longer hospital stays (13.1 ± 12.2 days vs. 17.4 ± 12.6 days, p = 0.031) compared to those not using these drugs. Propensity score matching also supported these findings, indicating that antithrombotic therapy was associated with lower in-hospital mortality and longer hospital stays even after adjusting for factors like age, gender, and primary diagnosis. Conclusions Our analysis using various statistical methods, including propensity score matching and multivariate regression, confirms that use of antithrombotic drugs in 2.3 days, particularly antiplatelets, are associated with a lower risk of in-hospital mortality. However, they may increase the risk of rebleeding and extend hospital stays in certain subgroups.
- Published
- 2024
- Full Text
- View/download PDF
28. A randomization-based theory for preliminary testing of covariate balance in controlled trials
- Author
-
Zhao, Anqi and Ding, Peng
- Subjects
Statistics - Methodology ,Statistics - Applications - Abstract
Randomized trials balance all covariates on average and provide the gold standard for estimating treatment effects. Chance imbalances nevertheless exist more or less in realized treatment allocations and intrigue an important question: what should we do in case the treatment groups differ with respect to some important baseline characteristics? A common strategy is to conduct a {\it preliminary test} of the balance of baseline covariates after randomization, and invoke covariate adjustment for subsequent inference if and only if the realized allocation fails some prespecified criterion. Although such practice is intuitive and popular among practitioners, the existing literature has so far only evaluated its properties under strong parametric model assumptions in theory and simulation, yielding results of limited generality. To fill this gap, we examine two strategies for conducting preliminary test-based covariate adjustment by regression, and evaluate the validity and efficiency of the resulting inferences from the randomization-based perspective. As it turns out, the preliminary-test estimator based on the analysis of covariance can be even less efficient than the unadjusted difference in means, and risks anticonservative confidence intervals based on normal approximation even with the robust standard error. The preliminary-test estimator based on the fully interacted specification is on the other hand less efficient than its counterpart under the {\it always-adjust} strategy, and yields overconservative confidence intervals based on normal approximation. Based on theory and simulation, we echo the existing literature and do not recommend the preliminary-test procedure for covariate adjustment in randomized trials.
- Published
- 2023
29. Sensitivity Analysis for Unmeasured Confounding in Medical Product Development and Evaluation Using Real World Evidence
- Author
-
Ding, Peng, Fang, Yixin, Faries, Doug, Gruber, Susan, Lee, Hana, Lee, Joo-Yeon, Mishra-Kalyani, Pallavi, Shan, Mingyang, van der Laan, Mark, Yang, Shu, and Zhang, Xiang
- Subjects
Statistics - Methodology - Abstract
The American Statistical Association Biopharmaceutical Section (ASA BIOP) working group on real-world evidence (RWE) has been making continuous, extended effort towards a goal of supporting and advancing regulatory science with respect to non-interventional, clinical studies intended to use real-world data for evidence generation for the purpose of medical product development and evaluation (i.e., RWE studies). In 2023, the working group published a manuscript delineating challenges and opportunities in constructing estimands for RWE studies following a framework in ICH E9(R1) guidance on estimand and sensitivity analysis. As a follow-up task, we describe the other issue in RWE studies, sensitivity analysis. Focusing on the issue of unmeasured confounding, we review availability and applicability of sensitivity analysis methods for different types unmeasured confounding. We discuss consideration on the choice and use of sensitivity analysis for RWE studies. Updated version of this article will present how findings from sensitivity analysis could support regulatory decision-making using a real example., Comment: 17 pages, 2 figures
- Published
- 2023
30. Regression-Oriented Knowledge Distillation for Lightweight Ship Orientation Angle Prediction with Optical Remote Sensing Images
- Author
-
Shi, Zhan, Ding, Xin, Ding, Peng, Yang, Chun, Huang, Ru, and Song, Xiaoxuan
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
Ship orientation angle prediction (SOAP) with optical remote sensing images is an important image processing task, which often relies on deep convolutional neural networks (CNNs) to make accurate predictions. This paper proposes a novel framework to reduce the model sizes and computational costs of SOAP models without harming prediction accuracy. First, a new SOAP model called Mobile-SOAP is designed based on MobileNetV2, achieving state-of-the-art prediction accuracy. Four tiny SOAP models are also created by replacing the convolutional blocks in Mobile-SOAP with four small-scale networks, respectively. Then, to transfer knowledge from Mobile-SOAP to four lightweight models, we propose a novel knowledge distillation (KD) framework termed SOAP-KD consisting of a novel feature-based guidance loss and an optimized synthetic samples-based knowledge transfer mechanism. Lastly, extensive experiments on the FGSC-23 dataset confirm the superiority of Mobile-SOAP over existing models and also demonstrate the effectiveness of SOAP-KD in improving the prediction performance of four specially designed tiny models. Notably, by using SOAP-KD, the test mean absolute error of the ShuffleNetV2x1.0-based model is only 8% higher than that of Mobile-SOAP, but its number of parameters and multiply-accumulate operations (MACs) are respectively 61.6% and 60.8% less.
- Published
- 2023
31. Dietary allicin improves behavior, regulates intestinal microbial colonies, and improves behavioral resistance to handling stresses in the sea cucumber Apostichopus japonicus at high temperature
- Author
-
Huang, Xiyuan, Wang, Huiyan, Ding, Peng, Yang, Yunjie, Ding, Jun, and Zhao, Chong
- Published
- 2025
- Full Text
- View/download PDF
32. Resistance to temperature change, handling stress, and disease in small sea cucumbers Apostichopus japonicus in different color morphs
- Author
-
Ding, Peng, Wang, Xiajing, Wu, Hengye, Yu, Yushi, Li, Xiang, Ding, Jun, and Zhao, Chong
- Published
- 2025
- Full Text
- View/download PDF
33. Multifunctional PVA/PNIPAM conductive hydrogel sensors enabled human-machine interaction intelligent rehabilitation training
- Author
-
Zhao, Yanlong, Zhang, Xichong, Hao, Yilin, Zhao, Yinghe, Ding, Peng, Zhai, Wei, Dai, Kun, Zheng, Guoqiang, Liu, Chuntai, and Shen, Changyu
- Published
- 2024
- Full Text
- View/download PDF
34. The effects of PstR, a PadR family transcriptional regulatory factor, in Plesiomonas shigelloides are revealed by transcriptomics
- Author
-
Yan, Junxiang, Zhang, Zixu, Shi, Hongdan, Xue, Xinke, Li, Ang, Liu, Fenxia, Ding, Peng, Guo, Xi, and Cao, Boyang
- Published
- 2024
- Full Text
- View/download PDF
35. Association between screen time, homework and reading duration, sleep duration, social jetlag and mental health among Chinese children and adolescents
- Author
-
Li, Tingting, Liu, Xiaoling, Cao, Caiyun, Yang, Feng, Ding, Peng, Xu, Shaojun, Tao, Shuman, Wu, Xiaoyan, and Tao, Fangbiao
- Published
- 2024
- Full Text
- View/download PDF
36. Bioinformatics analysis of immune infiltration in human diabetic retinopathy and identification of immune-related hub genes and their ceRNA networks
- Author
-
Li, Jingru, Li, Chaozhong, Wu, Xinyu, Yu, Shuai, Sun, Guihu, Ding, Peng, Lu, Si, Zhang, Lijiao, Yang, Ping, Peng, Yunzhu, Fu, Jingyun, and Wang, Luqiao
- Published
- 2024
- Full Text
- View/download PDF
37. Intercultural interaction willingness: a PLS-PM approach to influencing factors and its mediation effect
- Author
-
Zheng, Haijian, Ding, Peng, Liu, Qian, and Xing, Lirong
- Published
- 2024
- Full Text
- View/download PDF
38. Prediction of delayed graft function by early salivary microbiota following kidney transplantation
- Author
-
Xiang, Xuyu, Peng, Bo, Liu, Kai, Wang, Tianyin, Ding, Peng, Zhu, Yi, Cheng, Ke, and Ming, Yingzi
- Published
- 2024
- Full Text
- View/download PDF
39. Mitochondria from osteolineage cells regulate myeloid cell-mediated bone resorption
- Author
-
Ding, Peng, Gao, Chuan, Zhou, Jian, Mei, Jialun, Li, Gan, Liu, Delin, Li, Hao, Liao, Peng, Yao, Meng, Wang, Bingqi, Lu, Yafei, Peng, Xiaoyuan, Jiang, Chenyi, Yin, Jimin, Huang, Yigang, Zheng, Minghao, Gao, Youshui, Zhang, Changqing, and Gao, Junjie
- Published
- 2024
- Full Text
- View/download PDF
40. Oropharyngeal swallowing hydrodynamics of thin and mildly thick liquids in an anatomically accurate throat-epiglottis model
- Author
-
Seifelnasr, Amr, Ding, Peng, Si, Xiuhua, Biondi, Andres, and Xi, Jinxiang
- Published
- 2024
- Full Text
- View/download PDF
41. Blocking the MIF-CD74 axis augments radiotherapy efficacy for brain metastasis in NSCLC via synergistically promoting microglia M1 polarization
- Author
-
Liu, Lichao, Wang, Jian, Wang, Ying, Chen, Lingjuan, Peng, Ling, Bin, Yawen, Ding, Peng, Zhang, Ruiguang, Tong, Fan, and Dong, Xiaorong
- Published
- 2024
- Full Text
- View/download PDF
42. Osteocyte mitochondria regulate angiogenesis of transcortical vessels
- Author
-
Liao, Peng, Chen, Long, Zhou, Hao, Mei, Jiong, Chen, Ziming, Wang, Bingqi, Feng, Jerry Q., Li, Guangyi, Tong, Sihan, Zhou, Jian, Zhu, Siyuan, Qian, Yu, Zong, Yao, Zou, Weiguo, Li, Hao, Zhang, Wenkan, Yao, Meng, Ma, Yiyang, Ding, Peng, Pang, Yidan, Gao, Chuan, Mei, Jialun, Zhang, Senyao, Zhang, Changqing, Liu, Delin, Zheng, Minghao, and Gao, Junjie
- Published
- 2024
- Full Text
- View/download PDF
43. CCDC92 promotes podocyte injury by regulating PA28α/ABCA1/cholesterol efflux axis in type 2 diabetic mice
- Author
-
Zuo, Fu-wen, Liu, Zhi-yong, Wang, Ming-wei, Du, Jun-yao, Ding, Peng-zhong, Zhang, Hao-ran, Tang, Wei, Sun, Yu, Wang, Xiao-jie, Zhang, Yan, Xie, Yu-sheng, Wu, Ji-chao, Liu, Min, Wang, Zi-ying, and Yi, Fan
- Published
- 2024
- Full Text
- View/download PDF
44. Study of Boson Stars with Wormhole
- Author
-
Ding, Peng-Bo, Ma, Tian-Xiang, and Wang, Yong-Qiang
- Subjects
General Relativity and Quantum Cosmology ,High Energy Physics - Theory - Abstract
In this paper, we reconsider the mixed system of BSs with wormholes at their center which performed by complex scalar field and phantom field and study a whole new condition about the potential. Both the symmetric and asymmetric solutions in the two asymptotically flat regions are obtained by using numerical method and we mainly explore the change of the results by varying the parameters of throats and potential. In ground state, we find there are multiple solutions at certain setting of parameters and with the increase of $\eta_0$ or decrease of $c$, the results gradually become single-valued functions and these two variables have similar influence to the curve shape of mass $M$ and charge $Q$, furthermore, the asymmetric solutions can turn into the solutions of symmetry at some frequency $\omega$ in certain $\eta_0$ and $c$. However, when it comes to excited state, the properties of solutions of symmetry is similar to the ground state while asymmetrical results exhibit altered characteristics. We also present the geometries of wormhole to investigate the property of this model., Comment: 23 pages, 17 figures
- Published
- 2023
45. A First Course in Causal Inference
- Author
-
Ding, Peng
- Subjects
Statistics - Methodology ,Statistics - Applications - Abstract
I developed the lecture notes based on my ``Causal Inference'' course at the University of California Berkeley over the past seven years. Since half of the students were undergraduates, my lecture notes only required basic knowledge of probability theory, statistical inference, and linear and logistic regressions., Comment: 1. Posted the companion R code and datasets on Harvard Dataverse https://doi.org/10.7910/DVN/ZX3VEV; 2. Corrected for many typos; 3. Added the Preface and many new homework problems
- Published
- 2023
46. Flexible sensitivity analysis for causal inference in observational studies subject to unmeasured confounding
- Author
-
Lu, Sizhu and Ding, Peng
- Subjects
Statistics - Methodology - Abstract
Causal inference with observational studies often suffers from unmeasured confounding, yielding biased estimators based on the unconfoundedness assumption. Sensitivity analysis assesses how the causal conclusions change with respect to different degrees of unmeasured confounding. Most existing sensitivity analysis methods work well for specific types of statistical estimation or testing strategies. We propose a flexible sensitivity analysis framework that can deal with commonly used inverse probability weighting, outcome regression, and doubly robust estimators simultaneously. It is based on the well-known parametrization of the selection bias as comparisons of the observed and counterfactual outcomes conditional on observed covariates. It is attractive for practical use because it only requires simple modifications of the standard estimators. Moreover, it naturally extends to many other causal inference settings, including the causal risk ratio or odds ratio, the average causal effect on the treated units, and studies with survival outcomes. We also develop an R package saci to implement our sensitivity analysis estimators.
- Published
- 2023
47. Boson star with parity-odd symmetry in wormhole spacetime
- Author
-
Yue, Yuan, Ding, Peng-Bo, and Wang, Yong-Qiang
- Subjects
General Relativity and Quantum Cosmology ,High Energy Physics - Theory - Abstract
In this paper, we revisit the model of bosonic matter in the form of a free complex scalar field with a nontrivial wormhole spacetime topology supported by a free phantom field. We obtain a new type of boson star with wormhole solutions, in which the complex scalar field possess full parity-odd symmetry with respect to the two asymptotically flat spacetime regions. When the size of the throat is small, The behavior of boson stars with wormhole approaches that of boson stars. When the size of the throat is intermediate, the typical spiraling dependence of the mass and the particle number on the frequency of the boson stars is replaced by a loop structure. However, as the size becomes relatively large, the loop structure will also disappear. In particular, The complex scalar field could form two boson stars with opposite phase differences with respect to the two spacetime regions in the limit of vanishing throat size. We analyze the properties of this new type of boson stars with wormhole and further show that the wormhole spacetime geometry., Comment: 9 pages, 5 figures
- Published
- 2023
48. When is the estimated propensity score better? High-dimensional analysis and bias correction
- Author
-
Su, Fangzhou, Mou, Wenlong, Ding, Peng, and Wainwright, Martin J.
- Subjects
Statistics - Methodology - Abstract
Anecdotally, using an estimated propensity score is superior to the true propensity score in estimating the average treatment effect based on observational data. However, this claim comes with several qualifications: it holds only if propensity score model is correctly specified and the number of covariates $d$ is small relative to the sample size $n$. We revisit this phenomenon by studying the inverse propensity score weighting (IPW) estimator based on a logistic model with a diverging number of covariates. We first show that the IPW estimator based on the estimated propensity score is consistent and asymptotically normal with smaller variance than the oracle IPW estimator (using the true propensity score) if and only if $n \gtrsim d^2$. We then propose a debiased IPW estimator that achieves the same guarantees in the regime $n \gtrsim d^{3/2}$. Our proofs rely on a novel non-asymptotic decomposition of the IPW error along with careful control of the higher order terms., Comment: Fangzhou Su and Wenlong Mou contributed equally to this work
- Published
- 2023
49. Forward selection and post-selection inference in factorial designs
- Author
-
Shi, Lei, Wang, Jingshen, and Ding, Peng
- Subjects
Statistics - Methodology ,62K15, 62J05, 62G05 - Abstract
Ever since the seminal work of R. A. Fisher and F. Yates, factorial designs have been an important experimental tool to simultaneously estimate the effects of multiple treatment factors. In factorial designs, the number of treatment combinations grows exponentially with the number of treatment factors, which motivates the forward selection strategy based on the sparsity, hierarchy, and heredity principles for factorial effects. Although this strategy is intuitive and has been widely used in practice, its rigorous statistical theory has not been formally established. To fill this gap, we establish design-based theory for forward factor selection in factorial designs based on the potential outcome framework. We not only prove a consistency property for the factor selection procedure but also discuss statistical inference after factor selection. In particular, with selection consistency, we quantify the advantages of forward selection based on asymptotic efficiency gain in estimating factorial effects. With inconsistent selection in higher-order interactions, we propose two strategies and investigate their impact on subsequent inference. Our formulation differs from the existing literature on variable selection and post-selection inference because our theory is based solely on the physical randomization of the factorial design and does not rely on a correctly specified outcome model., Comment: 80 pages, 4 figures
- Published
- 2023
50. Kernel-based off-policy estimation without overlap: Instance optimality beyond semiparametric efficiency
- Author
-
Mou, Wenlong, Ding, Peng, Wainwright, Martin J., and Bartlett, Peter L.
- Subjects
Mathematics - Statistics Theory ,Statistics - Methodology ,Statistics - Machine Learning - Abstract
We study optimal procedures for estimating a linear functional based on observational data. In many problems of this kind, a widely used assumption is strict overlap, i.e., uniform boundedness of the importance ratio, which measures how well the observational data covers the directions of interest. When it is violated, the classical semi-parametric efficiency bound can easily become infinite, so that the instance-optimal risk depends on the function class used to model the regression function. For any convex and symmetric function class $\mathcal{F}$, we derive a non-asymptotic local minimax bound on the mean-squared error in estimating a broad class of linear functionals. This lower bound refines the classical semi-parametric one, and makes connections to moduli of continuity in functional estimation. When $\mathcal{F}$ is a reproducing kernel Hilbert space, we prove that this lower bound can be achieved up to a constant factor by analyzing a computationally simple regression estimator. We apply our general results to various families of examples, thereby uncovering a spectrum of rates that interpolate between the classical theories of semi-parametric efficiency (with $\sqrt{n}$-consistency) and the slower minimax rates associated with non-parametric function estimation.
- Published
- 2023
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.