279 results on '"Adrian E. Raftery"'
Search Results
2. Probabilistic projections of increased heat stress driven by climate change
- Author
-
Lucas R. Vargas Zeppetello, Adrian E. Raftery, and David S. Battisti
- Subjects
Geology ,QE1-996.5 ,Environmental sciences ,GE1-350 - Abstract
Exposure to dangerous heat index levels will likely increase by 50-100% in the tropics and by a factor of 3-10 in the mid-latitudes by 2100, even if the Paris Agreement goal of limiting warming to 2°C is met, according to probabilistic projections of global warming.
- Published
- 2022
- Full Text
- View/download PDF
3. clustvarsel: A Package Implementing Variable Selection for Gaussian Model-Based Clustering in R
- Author
-
Luca Scrucca and Adrian E. Raftery
- Subjects
BIC ,model-based clustering ,subset selection ,Statistics ,HA1-4737 - Abstract
Finite mixture modeling provides a framework for cluster analysis based on parsimonious Gaussian mixture models. Variable or feature selection is of particular importance in situations where only a subset of the available variables provide clustering information. This enables the selection of a more parsimonious model, yielding more efficient estimates, a clearer interpretation and, often, improved clustering partitions. This paper describes the R package clustvarsel which performs subset selection for model-based clustering. An improved version of the Raftery and Dean (2006) methodology is implemented in the new release of the package to find the (locally) optimal subset of variables with group/cluster information in a dataset. Search over the solution space is performed using either a stepwise greedy search or a headlong algorithm. Adjustments for speeding up these algorithms are discussed, as well as a parallel implementation of the stepwise search. Usage of the package is presented through the discussion of several data examples.
- Published
- 2018
- Full Text
- View/download PDF
4. Probabilistic forecasting of maximum human lifespan by 2100 using Bayesian population projections
- Author
-
Michael Pearce and Adrian E. Raftery
- Subjects
Demography. Population. Vital events ,HB848-3697 - Abstract
Background: We consider the problem of quantifying the human lifespan using a statistical approach that probabilistically forecasts the maximum reported age at death (MRAD) through 2100. Objective: We seek to quantify the probability that any person attains various extreme ages, such as those above 120, by the year 2100. Methods: We use the exponential survival model for supercentenarians (people over age 110) of Rootzén and Zholud (2017) but extend the forecasting window, quantify population uncertainty using Bayesian population projections, and incorporate the most recent data from the International Database on Longevity (IDL) to obtain unconditional estimates of the distribution of MRAD this century in a fully Bayesian analysis. Results: We find that the exponential survival model for supercentenarians is consistent with the most recent IDL data and that projections of the population aged 110-114 through 2080 are sensible. We integrate over the posterior distributions of the exponential model parameter and uncertainty in the supercentenarian population projections to estimate an unconditional distribution of MRAD by 2100. Conclusions: Based on the Bayesian analysis, there is a greater than 99Š probability that the current MRAD of 122 will be broken by 2100. We estimate the probabilities that a person lives to at least age 126, 128, or 130 this century, as 89Š, 44Š, and 13Š, respectively. Contribution: We have updated the supercentenarian survival model of Rootzén and Zholud using the most recent IDL data, incorporated Bayesian population projections, and extended the forecasting window to create the first fully Bayesian and unconditional probabilistic projection of MRAD by 2100.
- Published
- 2021
- Full Text
- View/download PDF
5. bayesPop: Probabilistic Population Projections
- Author
-
Hana Ševčíková and Adrian E. Raftery
- Subjects
Bayesian hierarchical model ,population projections ,expression language ,population pyramid ,United Nations ,World Population Prospects ,Statistics ,HA1-4737 - Abstract
We describe bayesPop, an R package for producing probabilistic population projections for all countries. This uses probabilistic projections of total fertility and life expectancy generated by Bayesian hierarchical models. It produces a sample from the joint posterior predictive distribution of future age- and sex-specific population counts, fertility rates and mortality rates, as well as future numbers of births and deaths. It provides graphical ways of summarizing this information, including trajectory plots and various kinds of probabilistic population pyramids. An expression language is introduced which allows the user to produce the predictive distribution of a wide variety of derived population quantities, such as the median age or the old age dependency ratio. The package produces aggregated projections for sets of countries, such as UN regions or trading blocs. The methodology has been used by the United Nations to produce their most recent official population projections for all countries, published in the World Population Prospects.
- Published
- 2016
- Full Text
- View/download PDF
6. Probabilistic projection of subnational total fertility rates
- Author
-
Hana Sevcikova, Adrian E. Raftery, and Patrick Gerland
- Subjects
autoregressive model ,Bayesian hierarchical model ,correlation ,scaling model ,subnational projections ,total fertility rate (TFR) ,Demography. Population. Vital events ,HB848-3697 - Abstract
Background: We consider the problem of probabilistic projection of the total fertility rate (TFR) for subnational regions. Objective: We seek a method that is consistent with the UN's recently adopted Bayesian method for probabilistic TFR projections for all countries and works well for all countries. Methods: We assess various possible methods using subnational TFR data for 47 countries. Results: We find that the method that performs best in terms of out-of-sample predictive performance and also in terms of reproducing the within-country correlation in TFR is a method that scales each national trajectory from the national predictive posterior distribution by a region-specific scale factor that is allowed to vary slowly over time. Conclusions: Probabilistic projections of TFR for subnational units are best produced by scaling the national projection by a slowly time-varying region-specific scale factor. This supports the hypothesis of Watkins (1990, 1991) that within-country TFR converges over time in response to country-specific factors, and thus extends the Watkins hypothesis to the last50 years and to a much wider range of countries around the world. Contribution: We have developed a new method for probabilistic projection of subnational TFR that works well and outperforms other methods. This also sheds light on the extent to which within-country TFR converges over time.
- Published
- 2018
- Full Text
- View/download PDF
7. Bayesian projection of life expectancy accounting for the HIV/AIDS epidemic
- Author
-
Jessica Godwin and Adrian E. Raftery
- Subjects
Demography. Population. Vital events ,HB848-3697 - Abstract
Background: While probabilistic projection methods for projecting life expectancy exist, few account for covariates related to life expectancy. Generalized HIV/AIDS epidemics have a large, immediate negative impact on the life expectancy in a country, but this impact can be mitigated by widespread use of antiretroviral therapy (ART). Thus, projection methods for countries with generalized HIV/AIDS epidemics could be improved by accounting for HIV prevalence, the future course of the epidemic, and ART coverage. Methods: We extend the current Bayesian probabilistic life expectancy projection methods of Raftery et al. (2013) to account for HIV prevalence and adult ART coverage in countries with generalized HIV/AIDS epidemics. Results: We evaluate our method using out-of-sample validation. We find that the proposed method performs better than the method that does not account for HIV prevalence or ART coverage for projections of life expectancy in countries with a generalized epidemic, while projections for countries without an epidemic remain essentially unchanged. Conclusions: In general, our projections show rapid recovery to pre-epidemic life expectancy levels in the presence of widespread ART coverage. After the initial life expectancy recovery, we project a steady rise in life expectancy until the end of the century. Contribution: We develop a simple Bayesian hierarchical model for long-term projections of life expectancy while accounting for HIV/AIDS prevalence and coverage of ART. The method produces well-calibrated projections for countries with generalized HIV/AIDS epidemics up to 2100 while having limited data demands.
- Published
- 2017
- Full Text
- View/download PDF
8. Model-based Methods of Classification: Using the mclust Software in Chemometrics
- Author
-
Chris Fraley and Adrian E. Raftery
- Subjects
model-based clustering ,classification ,density estimation ,discriminant analysis ,mclust ,Statistics ,HA1-4737 - Abstract
Due to recent advances in methods and software for model-based clustering, and to the interpretability of the results, clustering procedures based on probability models are increasingly preferred over heuristic methods. The clustering process estimates a model for the data that allows for overlapping clusters, producing a probabilistic clustering that quantifies the uncertainty of observations belonging to components of the mixture. The resulting clustering model can also be used for some other important problems in multivariate analysis, including density estimation and discriminant analysis. Examples of the use of model-based clustering and classification techniques in chemometric studies include multivariate image analysis, magnetic resonance imaging, microarray image segmentation, statistical process control, and food authenticity. We review model-based clustering and related methods for density estimation and discriminant analysis, and show how the R package mclust can be applied in each instance.
- Published
- 2007
9. Probabilistic County-Level Population Projections
- Author
-
Crystal CY Yu, Hana Ševčíková, Adrian E. Raftery, and Sara R. Curran
- Subjects
Demography - Abstract
Population projections provide predictions of future population sizes for an area. Historically, most population projections have been produced using deterministic or scenario-based approaches and have not assessed uncertainty about future population change. Starting in 2015, however, the United Nations (UN) has produced probabilistic population projections for all countries using a Bayesian approach. There is also considerable interest in subnational probabilistic population projections, but the UN's national approach cannot be used directly for this purpose, because within-country correlations in fertility and mortality are generally larger than between-country ones, migration is not constrained in the same way, and there is a need to account for college and other special populations, particularly at the county level. We propose a Bayesian method for producing subnational population projections, including migration and accounting for college populations, by building on but modifying the UN approach. We illustrate our approach by applying it to the counties of Washington State and comparing the results with extant deterministic projections produced by Washington State demographers. Out-of-sample experiments show that our method gives accurate and well-calibrated forecasts and forecast intervals. In most cases, our intervals were narrower than the growth-based intervals issued by the state, particularly for shorter time horizons.
- Published
- 2023
- Full Text
- View/download PDF
10. Model-Based Clustering
- Author
-
Isobel Claire Gormley, Thomas Brendan Murphy, and Adrian E. Raftery
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Abstract
Clustering is the task of automatically gathering observations into homogeneous groups, where the number of groups is unknown. Through its basis in a statistical modeling framework, model-based clustering provides a principled and reproducible approach to clustering. In contrast to heuristic approaches, model-based clustering allows for robust approaches to parameter estimation and objective inference on the number of clusters, while providing a clustering solution that accounts for uncertainty in cluster membership. The aim of this article is to provide a review of the theory underpinning model-based clustering, to outline associated inferential approaches, and to highlight recent methodological developments that facilitate the use of model-based clustering for a broad array of data types. Since its emergence six decades ago, the literature on model-based clustering has grown rapidly, and as such, this review provides only a selection of the bibliography in this dynamic and impactful field.
- Published
- 2023
- Full Text
- View/download PDF
11. Contour Models for Physical Boundaries Enclosing Star-Shaped and Approximately Star-Shaped Polygons
- Author
-
Hannah M. Director and Adrian E. Raftery
- Subjects
Statistics and Probability ,Statistics, Probability and Uncertainty - Abstract
Boundaries on spatial fields divide regions with particular features from surrounding background areas. Methods to identify boundary lines from interpolated spatial fields are well established. Less attention has been paid to how to model sequences of connected spatial points. Such models are needed for physical boundaries. For example, in the Arctic ocean, large contiguous areas are covered by sea ice, or frozen ocean water. We define the ice edge contour as the ordered sequences of spatial points that connect to form a line around set(s) of contiguous grid boxes with sea ice present. Polar scientists need to describe how this contiguous area behaves in present and historical data and under future climate change scenarios. We introduce the Gaussian Star-shaped Contour Model (GSCM) for modelling boundaries represented as connected sequences of spatial points such as the sea ice edge. GSCMs generate sequences of spatial points via generating sets of distances in various directions from a fixed starting point. The GSCM can be applied to contours that enclose regions that are star-shaped polygons or approximately star-shaped polygons. Metrics are introduced to assess the extent to which a polygon deviates from star-shapedness. Simulation studies illustrate the performance of the GSCM in different situations.
- Published
- 2022
- Full Text
- View/download PDF
12. The Social Cost of Carbon: Advances in Long-Term Probabilistic Projections of Population, GDP, Emissions, and Discount Rates
- Author
-
Kevin Rennert, Brian C. Prest, William A. Pizer, Richard G. Newell, David Anthoff, Cora Kingdon, Lisa Rennels, Roger Cooke, Adrian E. Raftery, Hana Ševčíková, and Frank Errickson
- Subjects
Economics and Econometrics ,General Business, Management and Accounting - Published
- 2022
- Full Text
- View/download PDF
13. Model-Based Clustering, Classification, and Density Estimation Using mclust in R
- Author
-
Scrucca, Luca, primary, Fraley, Chris, additional, Murphy, T. Brendan, additional, and Adrian E., Raftery, additional
- Published
- 2023
- Full Text
- View/download PDF
14. Visualizing Gaussian Mixture Models
- Author
-
Luca Scrucca, Chris Fraley, T. Brendan Murphy, and Adrian E. Raftery
- Published
- 2023
- Full Text
- View/download PDF
15. Model-Based Clustering
- Author
-
Luca Scrucca, Chris Fraley, T. Brendan Murphy, and Adrian E. Raftery
- Published
- 2023
- Full Text
- View/download PDF
16. Finite Mixture Models
- Author
-
Luca Scrucca, Chris Fraley, T. Brendan Murphy, and Adrian E. Raftery
- Published
- 2023
- Full Text
- View/download PDF
17. Introduction
- Author
-
Luca Scrucca, Chris Fraley, T. Brendan Murphy, and Adrian E. Raftery
- Published
- 2023
- Full Text
- View/download PDF
18. The Lee–Carter method and probabilistic population forecasts
- Author
-
Adrian E. Raftery
- Subjects
Business and International Management - Published
- 2023
- Full Text
- View/download PDF
19. Integration of Multiple Data Sources for Gene Network Inference using Genetic Perturbation Data: Extended Abstract.
- Author
-
Xiao Liang, William Chad Young, Ling-Hong Hung, Adrian E. Raftery, and Ka Yee Yeung
- Published
- 2018
- Full Text
- View/download PDF
20. Effect of Model Space Priors on Statistical Inference with Model Uncertainty
- Author
-
Anupreet Porwal and Adrian E. Raftery
- Abstract
Bayesian model averaging (BMA) provides a coherent way to account for model uncertainty in statistical inference tasks. BMA requires specification of model space priors and parameter space priors. In this article we focus on comparing different model space priors in the presence of model uncertainty. We consider eight reference model space priors used in the literature and three adaptive parameter priors recommended by Porwal and Raftery [37]. We assess the performance of these combinations of prior specifications for variable selection in linear regression models for the statistical tasks of parameter estimation, interval estimation, inference, point and interval prediction. We carry out an extensive simulation study based on 14 real datasets representing a range of situations encountered in practice. We found that beta-binomial model space priors specified in terms of the prior probability of model size performed best on average across various statistical tasks and datasets, outperforming priors that were uniform across models. Recently proposed complexity priors performed relatively poorly.
- Published
- 2022
- Full Text
- View/download PDF
21. BIC extensions for order-constrained model selection
- Author
-
Joris Mulder, Adrian E. Raftery, and Department of Methodology and Statistics
- Subjects
FOS: Computer and information sciences ,model selection ,Sociology and Political Science ,Computer science ,Machine learning ,computer.software_genre ,01 natural sciences ,Article ,Methodology (stat.ME) ,010104 statistics & probability ,0504 sociology ,Bayesian information criterion ,European Values Study ,truncated priors ,0101 mathematics ,Social science research ,Statistics - Methodology ,business.industry ,Model selection ,05 social sciences ,050401 social sciences methods ,Bayes factor ,HYPOTHESES ,order constraints ,Order (business) ,BAYES FACTORS ,Artificial intelligence ,business ,computer ,Social Sciences (miscellaneous) - Abstract
The Schwarz or Bayesian information criterion (BIC) is one of the most widely used tools for model comparison in social science research. The BIC however is not suitable for evaluating models with order constraints on the parameters of interest. This paper explores two extensions of the BIC for evaluating order constrained models, one where a truncated unit information prior is used under the order-constrained model, and the other where a truncated local unit information prior is used. The first prior is centered around the maximum likelihood estimate and the latter prior is centered around a null value. Several analyses show that the order-constrained BIC based on the local unit information prior works better as an Occam's razor for evaluating order-constrained models and results in lower error probabilities. The methodology based on the local unit information prior is implemented in the R package `BICpack' which allows researchers to easily apply the method for order-constrained model selection. The usefulness of the methodology is illustrated using data from the European Values Study., 25 pages, 4, figures, 2 tables
- Published
- 2022
- Full Text
- View/download PDF
22. Probabilistic population forecasting: Short to very long-term
- Author
-
Hana Ševčíková and Adrian E. Raftery
- Subjects
education.field_of_study ,Greenhouse gas ,Social cost ,Population ,Bayesian probability ,Probabilistic logic ,Econometrics ,Economics ,World population ,Business and International Management ,Private sector ,education ,Term (time) - Abstract
Population forecasts are used by governments and the private sector for planning, with horizons up to about three generations (around 2100) for different purposes. The traditional methods are deterministic using scenarios, but probabilistic forecasts are desired to get an idea of accuracy, assess changes, and make decisions involving risks. In a significant breakthrough, since 2015, the United Nations has issued probabilistic population forecasts for all countries using a Bayesian methodology that we review here. Assessment of the social cost of carbon relies on long-term forecasts of carbon emissions, which in turn depend on even longer-range population and economic forecasts, to 2300. We extend the UN method to very-long range population forecasts by combining the statistical approach with expert review and elicitation. While the world population is projected to grow for the rest of this century, it will likely stabilize in the 22nd century and decline in the 23rd century.
- Published
- 2023
23. Probabilistic Forecasts of Arctic Sea Ice Thickness
- Author
-
Adrian E. Raftery, Cecilia M. Bitz, Hannah M. Director, and Peter A. Gao
- Subjects
Statistics and Probability ,geography ,geography.geographical_feature_category ,Calibration (statistics) ,Applied Mathematics ,Prediction interval ,Statistical model ,Agricultural and Biological Sciences (miscellaneous) ,Arctic ice pack ,Physics::Geophysics ,Gaussian random field ,Arctic ,Climatology ,Sea ice thickness ,Sea ice ,Environmental science ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Physics::Atmospheric and Oceanic Physics ,General Environmental Science - Abstract
In recent decades, warming temperatures have caused sharp reductions in the volume of sea ice in the Arctic Ocean. Predicting changes in Arctic sea ice thickness is vital in a changing Arctic for making decisions about shipping and resource management in the region. We propose a statistical spatio-temporal two-stage model for sea ice thickness and use it to generate probabilistic forecasts up to three months into the future. Our approach combines a contour model to predict the ice-covered region with a Gaussian random field to model ice thickness conditional on the ice-covered region. Using the most complete estimates of sea ice thickness currently available, we apply our method to forecast Arctic sea ice thickness. Point predictions and prediction intervals from our model offer comparable accuracy and improved calibration compared with existing forecasts. We show that existing forecasts produced by ensembles of deterministic dynamic models can have large errors and poor calibration. We also show that our statistical model can generate good forecasts of aggregate quantities such as overall and regional sea ice volume. Supplementary materials accompanying this paper appear on-line.
- Published
- 2021
- Full Text
- View/download PDF
24. Probabilistic Projection of Subnational Life Expectancy
- Author
-
Hana Ševčíková and Adrian E. Raftery
- Subjects
business.industry ,Statistics ,Probabilistic logic ,Economics ,Life expectancy ,Computer vision ,Artificial intelligence ,business ,Projection (set theory) ,HA1-4737 - Abstract
Projecting mortality for subnational units, or regions, is of great interest to practicing demographers. We seek a probabilistic method for projecting subnational life expectancy that is based on the national Bayesian hierarchical model used by the United Nations, and at the same time is easy to use. We propose three methods of this kind. Two of them are variants of simple scaling methods. The third method models life expectancy for a region as equal to national life expectancy plus a region-specific stochastic process which is a heteroskedastic first-order autoregressive process (AR(1)), with a variance that declines to a constant as life expectancy increases. We apply our models to data from 29 countries. In an out-of-sample comparison, the proposed methods outperformed other comparative methods and were well calibrated for individual regions. The AR (1) method performed best in terms of crossover patterns between regions. Although the methods work well for individual regions, there are some limitations when evaluating within-country variation. We identified four countries for which the AR(1) method either underestimated or overestimated the predictive between-region within-country standard deviation. However, none of the competing methods work better in this regard than the AR(1) method. In addition to providing the full distribution of subnational life expectancy, the methods can be used to obtain probabilistic forecasts of age-specific mortality rates.
- Published
- 2021
- Full Text
- View/download PDF
25. Comprehensive evidence implies a higher social cost of CO2
- Author
-
Kevin Rennert, Frank Errickson, Brian C. Prest, Lisa Rennels, Richard G. Newell, William Pizer, Cora Kingdon, Jordan Wingenroth, Roger Cooke, Bryan Parthum, David Smith, Kevin Cromar, Delavane Diaz, Frances C. Moore, Ulrich K. Müller, Richard J. Plevin, Adrian E. Raftery, Hana Ševčíková, Hannah Sheets, James H. Stock, Tammy Tan, Mark Watson, Tony E. Wong, and David Anthoff
- Subjects
Risk ,Multidisciplinary ,General Science & Technology ,Climate ,Uncertainty ,Climate Models ,Carbon Dioxide ,Basic Behavioral and Social Science ,Environmental Policy ,Climate Action ,Greenhouse Gases ,Socioeconomic Factors ,Delay Discounting ,Behavioral and Social Science ,Policy Making - Abstract
The social cost of carbon dioxide (SC-CO2) measures the monetized value of the damages to society caused by an incremental metric tonne of CO2 emissions and is a key metric informing climate policy. Used by governments and other decision-makers in benefit–cost analysis for over a decade, SC-CO2 estimates draw on climate science, economics, demography and other disciplines. However, a 2017 report by the US National Academies of Sciences, Engineering, and Medicine1 (NASEM) highlighted that current SC-CO2 estimates no longer reflect the latest research. The report provided a series of recommendations for improving the scientific basis, transparency and uncertainty characterization of SC-CO2 estimates. Here we show that improved probabilistic socioeconomic projections, climate models, damage functions, and discounting methods that collectively reflect theoretically consistent valuation of risk, substantially increase estimates of the SC-CO2. Our preferred mean SC-CO2 estimate is $185 per tonne of CO2 ($44–$413 per tCO2: 5%–95% range, 2020 US dollars) at a near-term risk-free discount rate of 2%, a value 3.6 times higher than the US government’s current value of $51 per tCO2. Our estimates incorporate updated scientific understanding throughout all components of SC-CO2 estimation in the new open-source Greenhouse Gas Impact Value Estimator (GIVE) model, in a manner fully responsive to the near-term NASEM recommendations. Our higher SC-CO2 values, compared with estimates currently used in policy evaluation, substantially increase the estimated benefits of greenhouse gas mitigation and thereby increase the expected net benefits of more stringent climate policies.
- Published
- 2022
26. Probabilistic forecasting of maximum human lifespan by 2100 using Bayesian population projections
- Author
-
Adrian E. Raftery and Michael Pearce
- Subjects
Projections of population growth ,education.field_of_study ,Statistics ,Bayesian probability ,Population ,Probabilistic logic ,Probabilistic forecasting ,education ,Projection (set theory) ,Survival analysis ,Demography ,Exponential function ,Mathematics - Abstract
Background: We consider the problem of quantifying the human lifespan using a statistical approach that probabilistically forecasts the maximum reported age at death (MRAD) through 2100. Objective: We seek to quantify the probability that any person attains various extreme ages, such as those above 120, by the year 2100. Methods: We use the exponential survival model for supercentenarians (people over age 110) of Rootzen and Zholud (2017) but extend the forecasting window, quantify population uncertainty using Bayesian population projections, and incorporate the most recent data from the International Database on Longevity (IDL) to obtain unconditional estimates of the distribution of MRAD this century in a fully Bayesian analysis. Results: We find that the exponential survival model for supercentenarians is consistent with the most recent IDL data and that projections of the population aged 110–114 through 2080 are sensible. We integrate over the posterior distributions of the exponential model parameter and uncertainty in the supercentenarian population projections to estimate an unconditional distribution of MRAD by 2100. Conclusions: Based on the Bayesian analysis, there is a greater than 99% probability that the current MRAD of 122 will be broken by 2100. We estimate the probabilities that a person lives to at least age 126, 128, or 130 this century, as 89%, 44%, and 13%, respectively. Contribution: We have updated the supercentenarian survival model of Rootzen and Zholud using the most recent IDL data, incorporated Bayesian population projections, and extended the forecasting window to create the first fully Bayesian and unconditional probabilistic projection of MRAD by 2100.
- Published
- 2021
- Full Text
- View/download PDF
27. Probabilistic forecasts of international bilateral migration flows
- Author
-
Nathan G. Welch and Adrian E. Raftery
- Subjects
Multidisciplinary ,Internationality ,Models, Statistical ,Human Migration ,Humans ,Bayes Theorem ,Emigration and Immigration ,Forecasting - Abstract
We propose a method for forecasting global human migration flows. A Bayesian hierarchical model is used to make probabilistic projections of the 39,800 bilateral migration flows among the 200 most populous countries. We generate out-of-sample forecasts for all bilateral flows for the 2015 to 2020 period, using models fitted to bilateral migration flows for five 5-y periods from 1990 to 1995 through 2010 to 2015. We find that the model produces well-calibrated out-of-sample forecasts of bilateral flows, as well as total country-level inflows, outflows, and net flows. The mean absolute error decreased by 61% using our method, compared to a leading model of international migration. Out-of-sample analysis indicated that simple methods for forecasting migration flows offered accurate projections of bilateral migration flows in the near term. Our method matched or improved on the out-of-sample performance using these simple deterministic alternatives, while also accurately assessing uncertainty. We integrate the migration flow forecasting model into a fully probabilistic population projection model to generate bilateral migration flow forecasts by age and sex for all flows from 2020 to 2025 through 2040 to 2045.
- Published
- 2022
28. Balancing Inferential Integrity and Disclosure Risk Via Model Targeted Masking and Multiple Imputation
- Author
-
Naisyin Wang, Bei Jiang, Russell Steele, and Adrian E. Raftery
- Subjects
Statistics and Probability ,Information privacy ,Computer science ,05 social sciences ,computer.software_genre ,01 natural sciences ,Masking (Electronic Health Record) ,Synthetic data ,010104 statistics & probability ,0502 economics and business ,Data mining ,0101 mathematics ,Statistics, Probability and Uncertainty ,computer ,050205 econometrics - Abstract
There is a growing expectation that data collected by government-funded studies should be openly available to ensure research reproducibility, which also increases concerns about data privacy. A st...
- Published
- 2021
- Full Text
- View/download PDF
29. Comparing methods for statistical inference with model uncertainty
- Author
-
Anupreet Porwal and Adrian E. Raftery
- Subjects
Multidisciplinary - Abstract
Probability models are used for many statistical tasks, notably parameter estimation, interval estimation, inference about model parameters, point prediction, and interval prediction. Thus, choosing a statistical model and accounting for uncertainty about this choice are important parts of the scientific process. Here we focus on one such choice, that of variables to include in a linear regression model. Many methods have been proposed, including Bayesian and penalized likelihood methods, and it is unclear which one to use. We compared 21 of the most popular methods by carrying out an extensive set of simulation studies based closely on real datasets that span a range of situations encountered in practical data analysis. Three adaptive Bayesian model averaging (BMA) methods performed best across all statistical tasks. These used adaptive versions of Zellner’s g-prior for the parameters, where the prior variance parameter g is a function of sample size or is estimated from the data. We found that for BMA methods implemented with Markov chain Monte Carlo, 10,000 iterations were enough. Computationally, we found two of the three best methods (BMA with g=√n and empirical Bayes-local) to be competitive with the least absolute shrinkage and selection operator (LASSO), which is often preferred as a variable selection technique because of its computational efficiency. BMA performed better than Bayesian model selection (in which just one model is selected).
- Published
- 2022
30. How Do Education and Family Planning Accelerate Fertility Decline?
- Author
-
Daphne H. Liu and Adrian E. Raftery
- Subjects
Secondary education ,Sociology and Political Science ,media_common.quotation_subject ,05 social sciences ,Fertility ,Articles ,Development ,Educational attainment ,Article ,Unmet needs ,Granger causality ,050902 family studies ,Family planning ,0502 economics and business ,Contraceptive prevalence ,Demographic economics ,050207 economics ,0509 other social sciences ,Psychology ,Demography ,media_common - Abstract
Education and family planning can both be influenced by policy and are thought to accelerate fertility decline. However, questions remain about the nature of these effects. Does the effect of education operate through increasing educational attainment of women or educational enrollment of children? At which educational level is the effect strongest? Does the effect of family planning operate through increasing contraceptive prevalence or reducing unmet need? Is education or family planning more important? We assessed the quantitative impact of education and family planning in high‐fertility settings using a regression framework inspired by Granger causality. We found that women's attainment of lower secondary education is key to accelerating fertility decline and found an accelerating effect of contraceptive prevalence for modern methods. We found the impact of contraceptive prevalence to be substantially larger than that of education. These accelerating effects hold in sub‐Saharan Africa, but with smaller effect sizes there than elsewhere.
- Published
- 2020
31. Probabilistic forecasting of the Arctic sea ice edge with contour modeling
- Author
-
M Hannah, Adrian E. Raftery, and Cecilia M. Bitz
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,010504 meteorology & atmospheric sciences ,Climate change ,Edge (geometry) ,Statistics - Applications ,01 natural sciences ,Physics::Geophysics ,Atmosphere ,010104 statistics & probability ,Sea ice ,Applications (stat.AP) ,0101 mathematics ,Physics::Atmospheric and Oceanic Physics ,0105 earth and related environmental sciences ,geography ,geography.geographical_feature_category ,Ensemble forecasting ,Lead (sea ice) ,13. Climate action ,Modeling and Simulation ,Climatology ,Seawater ,Astrophysics::Earth and Planetary Astrophysics ,Probabilistic forecasting ,Statistics, Probability and Uncertainty - Abstract
Sea ice, or frozen ocean water, freezes and melts every year in the Arctic. Forecasts of where sea ice will be located weeks to months in advance have become more important as the amount of sea ice declines due to climate change, for maritime planning and other uses. Typical sea ice forecasts are made with ensemble models, physics-based models of sea ice and the surrounding ocean and atmosphere. This paper introduces Mixture Contour Forecasting, a method to forecast sea ice probabilistically using a mixture of two distributions, one based on post-processed output from ensembles and the other on observed sea ice patterns in recent years. At short lead times, these forecasts are better calibrated than unadjusted dynamic ensemble forecasts and other statistical reference forecasts. To produce these forecasts, a statistical technique is introduced that directly models the sea ice edge contour, the boundary around the region that is ice-covered. Mixture Contour Forecasting and reference methods are evaluated for monthly sea ice forecasts for 2008-2016 at lead times ranging from 0.5-6.5 months using one of the European Centre for Medium-Range Weather Forecasts ensembles.
- Published
- 2021
- Full Text
- View/download PDF
32. Country-based rate of emissions reductions should increase by 80% beyond nationally determined contributions to meet the 2 °C target
- Author
-
Peiran R. Liu and Adrian E. Raftery
- Subjects
010504 meteorology & atmospheric sciences ,Global warming ,0211 other engineering and technologies ,02 engineering and technology ,01 natural sciences ,Agricultural economics ,Article ,General Earth and Planetary Sciences ,Environmental science ,021108 energy ,China ,Probabilistic framework ,0105 earth and related environmental sciences ,General Environmental Science - Abstract
The 2015 Paris Agreement aims to keep global warming by 2100 to below 2 °C, with 1.5 °C as a target. To that end, countries agreed to reduce their emissions by nationally determined contributions (NDCs). Using a fully statistically based probabilistic framework, we find that the probabilities of meeting their nationally determined contributions for the largest emitters are low, e.g. 2% for the USA and 16% for China. On current trends, the probability of staying below 2 °C of warming is only 5%, but if all countries meet their nationally determined contributions and continue to reduce emissions at the same rate after 2030, it rises to 26%. If the USA alone does not meet its nationally determined contribution, it declines to 18%. To have an even chance of staying below 2 °C, the average rate of decline in emissions would need to increase from the 1% per year needed to meet the nationally determined contributions, to 1.8% per year.
- Published
- 2021
33. Accounting for smoking in forecasting mortality and life expectancy
- Author
-
Yicheng Li and Adrian E. Raftery
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,business.industry ,Calibration (statistics) ,media_common.quotation_subject ,05 social sciences ,Accounting ,Statistics - Applications ,01 natural sciences ,Article ,3. Good health ,010104 statistics & probability ,Modeling and Simulation ,0502 economics and business ,Attributable risk ,Life expectancy ,Bayesian hierarchical modeling ,Applications (stat.AP) ,Quality (business) ,050207 economics ,0101 mathematics ,Statistics, Probability and Uncertainty ,business ,Psychology ,media_common - Abstract
Smoking is one of the main risk factors that has affected human mortality and life expectancy over the past century. Smoking accounts for a large part of the nonlinearities in the growth of life expectancy and of the geographic and sex differences in mortality. As Bongaarts (2006) and Janssen (2018) suggested, accounting for smoking could improve the quality of mortality forecasts due to the predictable nature of the smoking epidemic. We propose a new Bayesian hierarchical model to forecast life expectancy at birth for both sexes and for 69 countries with good data on smoking-related mortality. The main idea is to convert the forecast of the non-smoking life expectancy at birth (i.e., life expectancy at birth removing the smoking effect) into life expectancy forecast through the use of the age-specific smoking attributable fraction (ASSAF). We introduce a new age-cohort model for the ASSAF and a Bayesian hierarchical model for non-smoking life expectancy at birth. The forecast performance of the proposed method is evaluated by out-of-sample validation compared with four other commonly used methods for life expectancy forecasting. Improvements in forecast accuracy and model calibration based on the new method are observed.
- Published
- 2021
- Full Text
- View/download PDF
34. Will this be a Record-Breaking Century for Human Longevity?
- Author
-
Michael Pearce and Adrian E. Raftery
- Subjects
Statistics and Probability ,History ,Human longevity ,Environmental ethics - Abstract
The record for oldest human being was set in 1997 by Jeanne Calment of France at 122 years and 164 days. Michael Pearce and Adrian E. Raftery expect that record will be broken in the coming decades
- Published
- 2021
- Full Text
- View/download PDF
35. Model-Based Clustering, Classification, and Density Estimation Using Mclust in R
- Author
-
Luca Scrucca, Chris Fraley, T. Brendan Murphy, Adrian E. Raftery, Luca Scrucca, Chris Fraley, T. Brendan Murphy, and Adrian E. Raftery
- Subjects
- Estimation theory--Data processing, R (Computer program language), Cluster analysis--Data processing, Gaussian distribution--Data processing
- Abstract
Model-based clustering and classification methods provide a systematic statistical approach to clustering, classification, and density estimation via mixture modeling. The model-based framework allows the problems of choosing or developing an appropriate clustering or classification method to be understood within the context of statistical modeling. The mclust package for the statistical environment R is a widely adopted platform implementing these model-based strategies. The package includes both summary and visual functionality, complementing procedures for estimating and choosing models.Key features of the book: An introduction to the model-based approach and the mclust R package A detailed description of mclust and the underlying modeling strategies An extensive set of examples, color plots, and figures along with the R code for reproducing them Supported by a companion website, including the R code to reproduce the examples and figures presented in the book, errata, and other supplementary material Model-Based Clustering, Classification, and Density Estimation Using mclust in R is accessible to quantitatively trained students and researchers with a basic understanding of statistical methods, including inference and computing. In addition to serving as a reference manual for mclust, the book will be particularly useful to those wishing to employ these model-based techniques in research or applications in statistics, data science, clinical research, social science, and many other disciplines.
- Published
- 2023
36. ESTIMATING AND FORECASTING THE SMOKING-ATTRIBUTABLE MORTALITY FRACTION FOR BOTH GENDERS JOINTLY IN OVER 60 COUNTRIES
- Author
-
Yicheng Li and Adrian E. Raftery
- Subjects
FOS: Computer and information sciences ,0301 basic medicine ,Statistics and Probability ,Smoking attributable fraction ,Population ,double logistic curve ,Statistics - Applications ,01 natural sciences ,Article ,010104 statistics & probability ,03 medical and health sciences ,medicine ,Bayesian hierarchical modeling ,Applications (stat.AP) ,Fraction (mathematics) ,probabilislic projection ,0101 mathematics ,Risk factor ,Lung cancer ,education ,Bayesian hierarchical model ,education.field_of_study ,business.industry ,Mortality rate ,fungi ,Peto–Lopez method ,medicine.disease ,030104 developmental biology ,Modeling and Simulation ,Attributable risk ,Life expectancy ,Statistics, Probability and Uncertainty ,business ,Demography - Abstract
Smoking is one of the leading preventable threats to human health and a major risk factor for lung cancer, upper aerodigestive cancer and chronic obstructive pulmonary disease. Estimating and forecasting the smoking attributable fraction (SAF) of mortality can yield insights into smoking epidemics and also provide a basis for more accurate mortality and life expectancy projection. Peto et al. (Lancet 339 (1992) 1268–1278) proposed a method to estimate the SAF using the lung cancer mortality rate as an indicator of exposure to smoking in the population of interest. Here, we use the same method to estimate the all-age SAF (ASAF) for both genders for over 60 countries. We document a strong and cross-nationally consistent pattern of the evolution of the SAF over time. We use this as the basis for a new Bayesian hierarchical model to project future male and female ASAF from over 60 countries simultaneously. This gives forecasts as well as predictive distributions that can be used to find uncertainty intervals for any quantity of interest. We assess the model using out-of-sample predictive validation and find that it provides good forecasts and well-calibrated forecast intervals, comparing favorably with other methods.
- Published
- 2020
37. A Conversation about COVID-19 with Economists, Sociologists, Statisticians, and Operations Researchers
- Author
-
Adrian E. Raftery, Fiona Greig, Duncan Thomas, Jonathan P. Caulkins, David Banks, Sylvia Frühwirth-Schnatter, and Laura A. Albert
- Subjects
Coronavirus disease 2019 (COVID-19) ,business.industry ,media_common.quotation_subject ,Social change ,cost-benefit, COVID-19, economic recovery, re-engineering work, social inequity ,Public policy ,Public relations ,Political science ,Pandemic ,Economic recovery ,Conversation ,Cost benefit ,business ,media_common - Abstract
The COVID-19 pandemic is causing economic and social change. Moderated by David Banks, the Director of the Statistical and Applied Mathematical Sciences Institute (SAMSI), six eminent scientists who study different aspects of social change and public policy came together to discuss the impacts of the COVID-19 pandemic on the U.S. and the world. The discussion took a range of quantitative perspectives on how to respond to the crisis and to forecast what challenges lie ahead. Specific topics include the role of data science, strategies for beginning to reopen the economy, the international impact of the disease, and its effect upon universities.
- Published
- 2020
38. Estimating uncertainty in respondent-driven sampling using a tree bootstrap method
- Author
-
Aaron J. Baraff, Tyler H. McCormick, and Adrian E. Raftery
- Subjects
Male ,Colorado ,Adolescent ,Computer science ,Sexual Behavior ,High variability ,Human immunodeficiency virus (HIV) ,Social Sciences ,HIV Infections ,medicine.disease_cause ,01 natural sciences ,010104 statistics & probability ,03 medical and health sciences ,Risk-Taking ,0302 clinical medicine ,Surveys and Questionnaires ,Resampling ,Statistics ,Econometrics ,medicine ,Humans ,Computer Simulation ,Longitudinal Studies ,030212 general & internal medicine ,0101 mathematics ,Heterosexuality ,Substance Abuse, Intravenous ,Probability ,Models, Statistical ,Schools ,Sex Workers ,Multidisciplinary ,Patient Selection ,Uncertainty ,Social Support ,Sampling (statistics) ,United States ,Confidence interval ,Tree (data structure) ,Snowball sampling ,Adolescent Behavior ,Respondent ,Female ,Centers for Disease Control and Prevention, U.S ,Ukraine ,Algorithms - Abstract
Respondent-driven sampling (RDS) is a network-based form of chain-referral sampling used to estimate attributes of populations that are difficult to access using standard survey tools. Although it has grown quickly in popularity since its introduction, the statistical properties of RDS estimates remain elusive. In particular, the sampling variability of these estimates has been shown to be much higher than previously acknowledged, and even methods designed to account for RDS result in misleadingly narrow confidence intervals. In this paper, we introduce a tree bootstrap method for estimating uncertainty in RDS estimates based on resampling recruitment trees. We use simulations from known social networks to show that the tree bootstrap method not only outperforms existing methods but also captures the high variability of RDS, even in extreme cases with high design effects. We also apply the method to data from injecting drug users in Ukraine. Unlike other methods, the tree bootstrap depends only on the structure of the sampled recruitment trees, not on the attributes being measured on the respondents, so correlations between attributes can be estimated as well as variability. Our results suggest that it is possible to accurately assess the high level of uncertainty inherent in RDS.
- Published
- 2016
- Full Text
- View/download PDF
39. Model-based Clustering: Basic Ideas
- Author
-
Charles Bouveyron, Adrian E. Raftery, T. Brendan Murphy, and Gilles Celeux
- Subjects
Model based clustering ,Computer science ,Data mining ,computer.software_genre ,computer - Published
- 2019
- Full Text
- View/download PDF
40. Non-Gaussian Model-based Clustering
- Author
-
Charles Bouveyron, T. Brendan Murphy, Adrian E. Raftery, and Gilles Celeux
- Subjects
symbols.namesake ,business.industry ,Computer science ,symbols ,Pattern recognition ,Artificial intelligence ,business ,Cluster analysis ,Gaussian network model - Published
- 2019
- Full Text
- View/download PDF
41. Semi-supervised Clustering and Classification
- Author
-
Gilles Celeux, T. Brendan Murphy, Charles Bouveyron, and Adrian E. Raftery
- Subjects
Computer science ,business.industry ,Pattern recognition ,Artificial intelligence ,business ,Semi supervised clustering - Published
- 2019
- Full Text
- View/download PDF
42. High-dimensional Data
- Author
-
Gilles Celeux, Adrian E. Raftery, T. Brendan Murphy, and Charles Bouveyron
- Subjects
Clustering high-dimensional data ,Computer science ,Computational science - Published
- 2019
- Full Text
- View/download PDF
43. Discrete Data Clustering
- Author
-
Adrian E. Raftery, Gilles Celeux, Charles Bouveyron, and T. Brendan Murphy
- Subjects
Computer science ,business.industry ,Pattern recognition ,Artificial intelligence ,Cluster analysis ,business - Published
- 2019
- Full Text
- View/download PDF
44. Dealing with Difficulties
- Author
-
Gilles Celeux, Adrian E. Raftery, T. Brendan Murphy, and Charles Bouveyron
- Subjects
Computer science - Published
- 2019
- Full Text
- View/download PDF
45. Adaptive Incremental Mixture Markov chain Monte Carlo
- Author
-
Nial Friel, Antonietta Mira, Florian Maire, and Adrian E. Raftery
- Subjects
FOS: Computer and information sciences ,Statistics and Probability ,Importance weight ,Computer science ,Local adaptation ,Bayesian inference ,Independence sampler ,Sample (statistics) ,Statistics - Computation ,01 natural sciences ,Article ,Methodology (stat.ME) ,Independence Sampler ,010104 statistics & probability ,symbols.namesake ,0502 economics and business ,Discrete Mathematics and Combinatorics ,0101 mathematics ,Statistics - Methodology ,Computation (stat.CO) ,050205 econometrics ,Adaptive MCMC ,Importance Weight ,05 social sciences ,Markov chain Monte Carlo ,ComputingMethodologies_PATTERNRECOGNITION ,65C05 65C40 60G10 93E35 ,symbols ,Probability distribution ,Statistics, Probability and Uncertainty ,Algorithm - Abstract
We propose adaptive incremental mixture Markov chain Monte Carlo (AIMM), a novel approach to sample from challenging probability distributions defined on a general state-space. While adaptive MCMC methods usually update a parametric proposal kernel with a global rule, AIMM locally adapts a semiparametric kernel. AIMM is based on an independent Metropolis–Hastings proposal distribution which takes the form of a finite mixture of Gaussian distributions. Central to this approach is the idea that the proposal distribution adapts to the target by locally adding a mixture component when the discrepancy between the proposal mixture and the target is deemed to be too large. As a result, the number of components in the mixture proposal is not fixed in advance. Theoretically, we prove that there exists a stochastic process that can be made arbitrarily close to AIMM and that converges to the correct target distribution. We also illustrate that it performs well in practice in a variety of challenging situations, including high-dimensional and multimodal target distributions. Finally, the methodology is successfully applied to two real data examples, including the Bayesian inference of a semiparametric regression model for the Boston Housing dataset. Supplementary materials for this article are available online.
- Published
- 2019
46. Integration of Multiple Data Sources for Gene Network Inference Using Genetic Perturbation Data
- Author
-
Xiao Liang, Adrian E. Raftery, William Chad Young, Ka Yee Yeung, and Ling-Hong Hung
- Subjects
Lung Neoplasms ,Skin Neoplasms ,Computer science ,Systems biology ,Bayesian probability ,Gene regulatory network ,Inference ,Perturbation (astronomy) ,Computational biology ,Biology ,computer.software_genre ,01 natural sciences ,010104 statistics & probability ,03 medical and health sciences ,0302 clinical medicine ,Cell Line, Tumor ,Genetics ,Humans ,Gene Regulatory Networks ,0101 mathematics ,Molecular Biology ,Gene ,Melanoma ,Research Articles ,030304 developmental biology ,Regulation of gene expression ,Biological data ,Gene knockdown ,0303 health sciences ,Supervised learning ,Bayes Theorem ,Genomics ,Gene Expression Regulation, Neoplastic ,Computational Mathematics ,Multiple data ,ComputingMethodologies_PATTERNRECOGNITION ,Gene Ontology ,Computational Theory and Mathematics ,A549 Cells ,030220 oncology & carcinogenesis ,Modeling and Simulation ,Data mining ,ComputingMethodologies_GENERAL ,Supervised Machine Learning ,Transcriptome ,computer ,Data integration - Abstract
BackgroundThe inference of gene regulatory networks is of great interest and has various applications. The recent advances in high-throughout biological data collection have facilitated the construction and understanding of gene regulatory networks in many model organisms. However, the inference of gene networks from large-scale human genomic data can be challenging. Generally, it is difficult to identify the correct regulators for each gene in the large search space, given that the high dimensional gene expression data only provides a small number of observations for each gene.ResultsWe present a Bayesian approach integrating external data sources with knockdown data from human cell lines to infer gene regulatory networks. In particular, we assemble multiple data sources including gene expression data, genome-wide binding data, gene ontology, known pathways and use a supervised learning framework to compute prior probabilities of regulatory relationships. We show that our integrated method improves the accuracy of inferred gene networks. We apply our method to two different human cell lines, which illustrates the general scope of our method.ConclusionsWe present a flexible and systematic framework for external data integration that improves the accuracy of human gene network inference while retaining efficiency. Integrating various data sources of biological information also provides a systematic way to build on knowledge from existing literature.
- Published
- 2019
47. Consistency for the tree bootstrap in respondent-driven sampling
- Author
-
Tyler H. McCormick, A K B Green, and Adrian E. Raftery
- Subjects
Statistics and Probability ,Tree bootstrap ,Respondent-driven sampling ,Applied Mathematics ,General Mathematics ,Human immunodeficiency virus (HIV) ,Sampling (statistics) ,Miscellanea ,medicine.disease_cause ,Agricultural and Biological Sciences (miscellaneous) ,Tree (data structure) ,Block bootstrap ,Consistency (statistics) ,Statistics ,Respondent ,medicine ,Population proportion ,Fraction (mathematics) ,Consistency ,Statistics, Probability and Uncertainty ,General Agricultural and Biological Sciences ,Mathematics - Abstract
Summary Respondent-driven sampling is an approach for estimating features of populations that are difficult to access using standard survey tools, e.g., the fraction of injection drug users who are HIV positive. Baraff et al. (2016) introduced an approach to estimating uncertainty in population proportion estimates from respondent-driven sampling using the tree bootstrap method. In this paper we establish the consistency of this tree bootstrap approach in the case of \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{upgreek} \usepackage{mathrsfs} \setlength{\oddsidemargin}{-69pt} \begin{document} }{}$m$\end{document}-trees.
- Published
- 2019
48. Model-Based Clustering and Classification for Data Science: With Applications in R
- Author
-
Charles Bouveyron, Gilles Celeux, T. Brendan Murphy, Adrian E. Raftery, Laboratoire Jean Alexandre Dieudonné (LJAD), Université Nice Sophia Antipolis (1965 - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA), Modèles et algorithmes pour l’intelligence artificielle (MAASAI), Inria Sophia Antipolis - Méditerranée (CRISAM), Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Université Nice Sophia Antipolis (1965 - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Laboratoire Jean Alexandre Dieudonné (LJAD), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Scalable and Pervasive softwARe and Knowledge Systems (Laboratoire I3S - SPARKS), Laboratoire d'Informatique, Signaux, et Systèmes de Sophia Antipolis (I3S), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Université Nice Sophia Antipolis (1965 - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Laboratoire d'Informatique, Signaux, et Systèmes de Sophia Antipolis (I3S), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Centre National de la Recherche Scientifique (CNRS), Institut National de Recherche en Informatique et en Automatique (Inria), Statistique mathématique et apprentissage (CELESTE), Inria Saclay - Ile de France, Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire de Mathématiques d'Orsay (LMO), Université Paris-Sud - Paris 11 (UP11)-Centre National de la Recherche Scientifique (CNRS)-Université Paris-Sud - Paris 11 (UP11)-Centre National de la Recherche Scientifique (CNRS), University College Dublin [Dublin] (UCD), Department of Statistics, University of Washington [Seattle], ANR-19-P3IA-0002,3IA@cote d'azur,3IA Côte d'Azur(2019), Laboratoire Jean Alexandre Dieudonné (JAD), Université Côte d'Azur (UCA)-Université Nice Sophia Antipolis (... - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS), Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Université Nice Sophia Antipolis (... - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Laboratoire Jean Alexandre Dieudonné (JAD), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Centre National de la Recherche Scientifique (CNRS)-Scalable and Pervasive softwARe and Knowledge Systems (Laboratoire I3S - SPARKS), Université Nice Sophia Antipolis (... - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Université Nice Sophia Antipolis (... - 2019) (UNS), Université Côte d'Azur (UCA)-Université Côte d'Azur (UCA)-Centre National de la Recherche Scientifique (CNRS), Université Côte d'Azur (UCA)-Université Côte d'Azur (UCA)-Laboratoire Jean Alexandre Dieudonné (JAD), Université Côte d'Azur (UCA)-Université Côte d'Azur (UCA)-Centre National de la Recherche Scientifique (CNRS)-Centre National de la Recherche Scientifique (CNRS)-Scalable and Pervasive softwARe and Knowledge Systems (Laboratoire I3S - SPARKS), Université Côte d'Azur (UCA)-Université Côte d'Azur (UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Nice Sophia Antipolis (... - 2019) (UNS), Université Côte d'Azur (UCA)-Université Côte d'Azur (UCA)-Centre National de la Recherche Scientifique (CNRS)-Laboratoire d'Informatique, Signaux, et Systèmes de Sophia Antipolis (I3S), Université Côte d'Azur (UCA)-Université Côte d'Azur (UCA)-Centre National de la Recherche Scientifique (CNRS)-Centre National de la Recherche Scientifique (CNRS), INRIA Rocquencourt, and Centre National de la Recherche Scientifique (CNRS)-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS)-Université Paris-Saclay
- Subjects
010104 statistics & probability ,[MATH.MATH-ST]Mathematics [math]/Statistics [math.ST] ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,02 engineering and technology ,0101 mathematics ,01 natural sciences ,ComputingMilieux_MISCELLANEOUS ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] - Abstract
International audience
- Published
- 2019
- Full Text
- View/download PDF
49. Estimation of emigration, return migration, and transit migration between all pairs of countries
- Author
-
Adrian E. Raftery and Jonathan J. Azose
- Subjects
pseudo-Bayes estimation ,South asia ,Social Sciences ,Global migration ,Global population ,0502 economics and business ,international migration ,050602 political science & public administration ,Humans ,bilateral migration flows ,Country of birth ,050207 economics ,Transit (satellite) ,Mexico ,Estimation ,Transients and Migrants ,Multidisciplinary ,05 social sciences ,Statistics ,Emigration and Immigration ,United States ,0506 political science ,Emigration ,Geography ,PNAS Plus ,8. Economic growth ,Physical Sciences ,Demographic economics - Abstract
Significance Despite the importance of international migration, estimates of between-country migration flows are still imprecise. Reliable record keeping of migration events is typically available only in the developed world, and the best existing methods to produce global migration flow estimates are burdened by strong assumptions. We produce estimates of migration flows between all pairs of countries at 5-year intervals, revealing patterns obscured by previous estimation methods. In particular, our estimates reveal large bidirectional movements in all global regions, with roughly one-quarter of migration events consisting of returns to an individual’s country of birth., We propose a method for estimating migration flows between all pairs of countries that allows for decomposition of migration into emigration, return, and transit components. Current state-of-the-art estimates of bilateral migration flows rely on the assumption that the number of global migrants is as small as possible. We relax this assumption, producing complete estimates of all between-country migration flows with genuine estimates of total global migration. We find that the total number of individuals migrating internationally has oscillated between 1.13 and 1.29% of the global population per 5-year period since 1990. Return migration and transit migration are big parts of total migration; roughly one of four migration events is a return to an individual’s country of birth. In the most recent time period, we estimate particularly large return migration flows from the United States to Central and South America and from the Persian Gulf to south Asia.
- Published
- 2018
50. Interlocking directorates in Irish companies using a latent space model for bipartite networks
- Author
-
Adrian E. Raftery, Riccardo Rastelli, Nial Friel, and Jason Wyse
- Subjects
Engineering ,Operations research ,Markov process ,02 engineering and technology ,Bayesian inference ,01 natural sciences ,Boom ,010104 statistics & probability ,symbols.namesake ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Econometrics ,Company director ,0101 mathematics ,Multidisciplinary ,business.industry ,Corporate governance ,Statistical model ,Models, Economic ,Physical Sciences ,Financial crisis ,symbols ,Celtic Tiger ,business.job_title ,business ,Ireland - Abstract
We analyze the temporal bipartite network of the leading Irish companies and their directors from 2003 to 2013, encompassing the end of the Celtic Tiger boom and the ensuing financial crisis in 2008. We focus on the evolution of company interlocks, whereby a company director simultaneously sits on two or more boards. We develop a statistical model for this dataset by embedding the positions of companies and directors in a latent space. The temporal evolution of the network is modeled through three levels of Markovian dependence: one on the model parameters, one on the companies' latent positions, and one on the edges themselves. The model is estimated using Bayesian inference. Our analysis reveals that the level of interlocking, as measured by a contraction of the latent space, increased before and during the crisis, reaching a peak in 2009, and has generally stabilized since then.
- Published
- 2016
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.