549 results on '"Christophe Chesneau"'
Search Results
2. Univariate and bivariate extensions of the truncated inverted arctan power distribution with applications
- Author
-
H.E. Semary, Christophe Chesneau, Maha A. Aldahlan, Ibrahim Elbatal, Mohammed Elgarhy, Mahmoud M. Abdelwahab, and Ehab M. Almetwally
- Subjects
Unit distributions ,Trigonometric distributions ,Bivariate distributions ,Copulas ,Engineering data analysis ,Engineering (General). Civil engineering (General) ,TA1-2040 - Abstract
The importance of (probability) distributions in engineering science is undeniable. They are extensively used in this sector to carry out statistical studies and draw conclusions. In this article, we construct a new trigonometric distribution with support [0,1] called the truncated inverted arctan power (TIAP) distribution and demonstrate its use with engineering data. The corresponding probability density function is particularly flexible in the sense that it has various decreased right-skewness and unimodal forms. The associated hazard rate function shows that the TIAP distribution may adapt data with monotonic, U-shaped, and N-shaped failure rates. To reflect the interdependence of two random variables or sets of data, a bivariate variant of the TIAP distribution is also elaborated. All these distributional tools are then applied from a statistical perspective. In particular, the involved parameters are estimated using a maximum likelihood technique. Applications to engineering data indicate that the univariate and bivariate extensions for the TIAP distribution have a better fit compared to various current distributions.
- Published
- 2024
- Full Text
- View/download PDF
3. New comprehensive class of estimators for population proportion using auxiliary attribute: Simulation and an application
- Author
-
H.E. Semary, Sohaib Ahmad, Ibrahim Elbatal, Christophe Chesneau, Mohammed Elgarhy, and Ehab M. Almetwally
- Subjects
Simulation study ,Proportion ,Auxiliary attribute ,Mean square error ,Percentage relative efficiency ,Engineering (General). Civil engineering (General) ,TA1-2040 - Abstract
In this article, we present a comprehensive class of estimators designed for population proportion estimation by leveraging auxiliary attributes within the framework of simple random sampling. The proposed class encompasses a diverse range of estimators, each of which undergoes a thorough examination. We provide numerical expressions for both bias and mean squared error, employing a first-order approximation. The significance of the introduced class of estimators is underscored through a detailed analysis of numerical results. These findings demonstrate the marked superiority of the suggested estimators over their existing counterparts in terms of mean squared error and percentage relative efficiency, as observed in both actual and simulated data scenarios. Consequently, we advocate for the adoption of the proposed class of estimators, asserting its potential to yield improved outcomes when estimating population proportions through the utilization of simple random sampling techniques.
- Published
- 2024
- Full Text
- View/download PDF
4. Examining the determinants of student academic performance in Somaliland: estimating unobserved effects at student and school levels using multi-level logistic regression
- Author
-
Mohamoud Jama Ali, Christophe Chesneau, and Abdisalam Hassan Muse
- Subjects
Academic performance ,multi-level logistic regression ,ANOVA ,t-test ,Somaliland ,Adult Education and Lifelong Learning ,Education (General) ,L7-991 - Abstract
This study investigates the factors influencing student academic performance in Somaliland, including school type, gender, region, and residence type. It utilizes a dataset of 14,342 students and employs a multi-level logistic regression model to analyze how these variables affect students’ performance in national secondary exams. The preliminary analysis reveals interesting trends, such as girls generally outperforming boys in most regions and rural students scoring higher than their urban counterparts. Private school students also achieve better grades than those in public schools. The multi-level logistic regression analysis uncovers unobserved heterogeneity at the school level, suggesting the presence of hidden factors influencing student outcomes. The regression analysis suggests that school type, residence type, and region significantly impact academic performance at the school level, with urban students in public schools having a higher likelihood of excelling. Surprisingly, gender does not appear to be a significant factor in student performance. These findings highlight the complex interplay of factors influencing education outcomes in Somaliland and emphasize the need for targeted interventions at the school level to enhance educational quality. However, it’s important to acknowledge the study’s limitations and the necessity for further research to uncover additional hidden variables affecting student outcomes.
- Published
- 2024
- Full Text
- View/download PDF
5. Design and implementation of EventsKG for situational monitoring and security intelligence in India: An open-source intelligence gathering approach
- Author
-
Hashmy Hassan, Sudheep Elayidom, M.R. Irshad, and Christophe Chesneau
- Subjects
Knowledge graphs ,Ontology ,Domain specific KGs ,EventsKG ,Open-source intelligence gathering ,Cybernetics ,Q300-390 ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
This paper presents a method to construct and implement an Events Knowledge Graph (EventsKG) for security-related open-source intelligence gathering, focusing on event exploration for situation monitoring in India. The EventsKG is designed to process news articles, extract events of national security significance, and represent them in a consistent and intuitive manner. This method utilizes state-of-the-art natural language understanding techniques and the capabilities of graph databases to extract and organize events. A domain-specific ontology is created for effective storage and retrieval. In addition, we provide a user-friendly dashboard for querying and a complete visualization of events across India. The effectiveness of the EventsKG is assessed through a human evaluation of the information retrieval quality. Our approach contributes to rapid data availability and decision-making through a comprehensive understanding of events, including local events, from every part of India in just a few clicks. The system is evaluated against a manually annotated dataset and by involving human evaluators through a feedback survey, and it has shown good retrieval accuracy. The EventsKG can also be used for other applications such as threat intelligence, incident response, and situational awareness.
- Published
- 2024
- Full Text
- View/download PDF
6. Power unit inverse Lindley distribution with different measures of uncertainty, estimation and applications
- Author
-
Ahmed M. Gemeay, Najwan Alsadat, Christophe Chesneau, and Mohammed Elgarhy
- Subjects
shannon entropy ,rényi entropy ,exponential entropy ,havrda and charvat entropy ,unit inverse lindley distribution ,extropy ,weighted extropy ,maximum product spacing ,minimum spacing linex distance ,Mathematics ,QA1-939 - Abstract
This paper introduced and investigated the power unit inverse Lindley distribution (PUILD), a novel two-parameter generalization of the famous unit inverse Lindley distribution. Among its notable functional properties, the corresponding probability density function can be unimodal, decreasing, increasing, or right-skewed. In addition, the hazard rate function can be increasing, U-shaped, or N-shaped. The PUILD thus takes advantage of these characteristics to gain flexibility in the analysis of unit data compared to the former unit inverse Lindley distribution, among others. From a theoretical point of view, many key measures were determined under closed-form expressions, including mode, quantiles, median, Bowley's skewness, Moor's kurtosis, coefficient of variation, index of dispersion, moments of various types, and Lorenz and Bonferroni curves. Some important measures of uncertainty were also calculated, mainly through the incomplete gamma function. In the statistical part, the estimation of the parameters involved was studied using fifteen different methods, including the maximum likelihood method. The invariant property of this approach was then used to efficiently estimate different uncertainty measures. Some simulation results were presented to support this claim. The significance of the PUILD underlying model compared to several current statistical models, including the unit inverse Lindley, exponentiated Topp-Leone, Kumaraswamy, and beta and transformed gamma models, was illustrated by two applications using real datasets.
- Published
- 2024
- Full Text
- View/download PDF
7. Machine learning study using 2020 SDHS data to determine poverty determinants in Somalia
- Author
-
Abdirizak A. Hassan, Abdisalam Hassan Muse, and Christophe Chesneau
- Subjects
Machine learning ,Somalia ,Random forest ,Model precision ,Classical regression ,Sustainability ,Medicine ,Science - Abstract
Abstract Extensive research has been conducted on poverty in developing countries using conventional regression analysis, which has limited prediction capability. This study aims to address this gap by applying advanced machine learning (ML) methods to predict poverty in Somalia. Utilizing data from the first-ever 2020 Somalia Demographic and Health Survey (SDHS), a cross-sectional study design is considered. ML methods, including random forest (RF), decision tree (DT), support vector machine (SVM), and logistic regression, are tested and applied using R software version 4.1.2, while conventional methods are analyzed using STATA version 17. Evaluation metrics, such as confusion matrix, accuracy, precision, sensitivity, specificity, recall, F1 score, and area under the receiver operating characteristic (AUROC), are employed to assess the performance of predictive models. The prevalence of poverty in Somalia is notable, with approximately seven out of ten Somalis living in poverty, making it one of the highest rates in the region. Among nomadic pastoralists, agro-pastoralists, and internally displaced persons (IDPs), the poverty average stands at 69%, while urban areas have a lower poverty rate of 60%. The accuracy of prediction ranged between 67.21% and 98.36% for the advanced ML methods, with the RF model demonstrating the best performance. The results reveal geographical region, household size, respondent age group, husband employment status, age of household head, and place of residence as the top six predictors of poverty in Somalia. The findings highlight the potential of ML methods to predict poverty and uncover hidden information that traditional statistical methods cannot detect, with the RF model identified as the best classifier for predicting poverty in Somalia.
- Published
- 2024
- Full Text
- View/download PDF
8. A novel weighted family of probability distributions with applications to world natural gas, oil, and gold reserves
- Author
-
Amal S. Hassan, Najwan Alsadat, Christophe Chesneau, and Ahmed W. Shawki
- Subjects
weighted distribution ,incomplete moments ,tsallis measure ,maximum likelihood estimation ,censored samples ,Biotechnology ,TP248.13-248.65 ,Mathematics ,QA1-939 - Abstract
Recent innovations have focused on the creation of new families that extend well-known distributions while providing a huge amount of practical flexibility for data modeling. Weighted distributions offer an effective approach for addressing model building and data interpretation problems. The main objective of this work is to provide a novel family based on a weighted generator called the length-biased truncated Lomax-generated (LBTLo-G) family. Discussions are held about the characteristics of the LBTLo-G family, including expressions for the probability density function, moments, and incomplete moments. In addition, different measures of uncertainty are determined. We provide four new sub-distributions and investigated their functionalities. Subsequently, a statistical analysis is given. The LBTLo-G family's parameter estimation is carried out using the maximum likelihood technique on the basis of full and censored samples. Simulation research is conducted to determine the parameters of the LBTLo Weibull (LBTLoW) distribution. Four genuine data sets are considered to illustrate the fitting behavior of the LBTLoW distribution. In each case, the application outcomes demonstrate that the LBTLoW distribution can, in fact, fit the data more accurately than other rival distributions.
- Published
- 2023
- Full Text
- View/download PDF
9. Estimation and Prediction Under Different Schemes for a Flexible Symmetric Distribution With Applications
- Author
-
Christophe Chesneau, Reza Pakyari, Akram Kohansal, and Hassan S. Bakouch
- Subjects
Mathematics ,QA1-939 - Abstract
This paper introduces a new probability distribution called the mixture symmetric gamma (MSG) distribution, which is defined as a mixture of two symmetric gamma distributions. Its statistical properties and applications are explored. We first examine its mathematical properties, including the possible shapes of the corresponding probability density function, as well as the moments, and the moment-generating function. We then look at parameter estimation using various frequentist and Bayesian methods, such as moment estimation, maximum likelihood method, least-squares method, and Bayesian approaches. In addition, the prediction of future observations under the MSG model is extensively covered, considering both frequentist and Bayesian perspectives, including median prediction, best unbiased prediction, and Bayesian prediction. A comprehensive simulation study is conducted to evaluate the performance of the proposed estimation and prediction techniques. Finally, the practical applicability of the MSG model is demonstrated through the analysis of four real-world datasets. It is shown to outperform several well-known competing models in terms of goodness-of-fit. The results highlight the inherent simplicity, efficiency, robustness, and intuitive interpretability of the MSG distribution, making it a compelling choice for modeling data with a symmetric pattern, with potential applications in diverse domains.
- Published
- 2024
- Full Text
- View/download PDF
10. A New Zero–Inflated Regression Model with Applications to Australian Health Survey and Biochemistry Graduate Students Data
- Author
-
Caner Tanış, Mahmoud M. Mansour, Enayat M. Abd Elrazik, Christophe Chesneau, Hazem Al-Mofleh, and Ahmed Z. Afify
- Subjects
Mathematics ,QA1-939 - Abstract
In this study, we propose a new zero-inflated regression model as an alternative to zero-inflated regression models, such as the zero-inflated Poisson, zero-inflated negative binomial, zero-inflated hurdle-Poisson, and zero-inflated hurdle negative binomial models. In this regard, we take benefit of the flexibility of the Poisson–Bilal distribution and some of its notable properties. More concretely, it is employed as the baseline distribution to generate a new regression model called the zero-inflated Poisson-Bilal regression model. It is designed to be a good alternative for modeling overdispersed data quite effectively. This aspect is emphasized using two real-world data sets from the medicine and education fields. Furthermore, these data sets are analyzed to compare the goodness-of-fit of the suggested zero-inflated regression model with some of its direct competitors.
- Published
- 2024
- Full Text
- View/download PDF
11. Modern Approach in Pattern Recognition Using Circular Fermatean Fuzzy Similarity Measure for Decision Making with Practical Applications
- Author
-
Revathy Aruchsamy, Inthumathi Velusamy, Prasantha Bharathi Dhandapani, Suleman Nasiru, and Christophe Chesneau
- Subjects
Mathematics ,QA1-939 - Abstract
The circular Fermatean fuzzy (CFF) set is an advancement of the Fermatean fuzzy (FF) set and the interval-valued Fermatean fuzzy (IVFF) set which deals with uncertainty. The CFF set is represented as a circle of radius ranging from 0 to 2 with the center at the degree of association (DA) and degree of nonassociation (DNA). If multiple people are involved in making decisions, the CFF set, as an alternative to the FF and IVFF sets, can deal with ambiguity more effectively by encircling the decision values within a circle rather than taking an average. Using algorithms, a pattern can be observed computationally or visually. Machine learning algorithm utilizes pattern recognition as an instrument for identifying patterns and also similarity measure (SM) is a beneficial pattern recognition tool used to classify items, discover variations, and make future predictions for decision making. In this work, we introduce the CFF cosine and Dice similarity measures (CFFDMs and CFFSMs), and their properties are studied. Unlike traditional approaches of decision making, which emphasize a single number, the proposed CFFSMs observe the pattern over the circular region to help in dealing with uncertainty more effectively. We introduce an innovative decision-making method in the FF setting. Available bank loans and applicants’ eligibility levels are represented as CFF set using their FF criteria and are taken as loan patterns and customer eligibility patterns. The loan is allocated to the applicant by measuring the CFFCSM and CFFDSM between the two patterns. Also, laptops are suggested to the customers by measuring the similarity between specification pattern and requirement pattern. The correctness and consistency of the proposed models are ensured by comparison analysis and graphical simulations of the input and similarity CFFNs.
- Published
- 2024
- Full Text
- View/download PDF
12. Classical and Bayesian inferences on the stress-strength reliability R=P[Y
- Author
-
Amit Singh Nayal, Bhupendra Singh, Abhishek Tyagi, and Christophe Chesneau
- Subjects
bayes estimation ,geometric distribution ,gibbs sampler ,lindley's approximation ,maximum likelihood estimation method ,stress-strength reliability ,Mathematics ,QA1-939 - Abstract
The subject matter described herein includes the analysis of the stress-strength reliability of the system, in which the discrete strength of the system is impacted by two random discrete stresses. The reliability function of such systems is denoted by $ R = P[Y < X < Z] $, where $ X $ is the strength of the system and $ Y $ and $ Z $ are the stresses. We look at how $ X $, $ Y $ and $ Z $ fit into a well-known discrete distribution known as the geometric distribution. The stress-strength reliability of this form is not widely studied in the current literature, and research in this area has only considered the scenario when the strength and stress variables follow a continuous distribution, although it is essentially nil in the case of discrete stress and strength. There are numerous applications wherein a system is exposed to external stress, and its functionality depends on whether its intrinsic physical strength falls within specific stress limits. Furthermore, the continuous measurement of stress and strength variables presents inherent difficulties and inconveniences in such scenarios. For the suggested distribution, we obtain the maximum likelihood estimate of the variable $ R $, as well as its asymptotic distribution and confidence interval. Additionally, in the classical setup, we find the boot-p and boot-t confidence intervals for $ R $. In the Bayesian setup, we utilize the widely recognized Markov Chain Monte Carlo technique and the Lindley approximation method to find the Bayes estimate of $ R $ under the squared error loss function. A Monte Carlo simulation study and real data analysis are demonstrated to show the applicability of the suggested model in the real world.
- Published
- 2023
- Full Text
- View/download PDF
13. A Novel Zero-Truncated Katz Distribution by the Lagrange Expansion of the Second Kind with Associated Inferences
- Author
-
Damodaran Santhamani Shibu, Christophe Chesneau, Mohanan Monisha, Radhakumari Maya, and Muhammed Rasheed Irshad
- Subjects
Lagrange expansion of the first kind ,Lagrange expansion of the second kind ,zero-truncated Katz distribution ,dispersion ,maximum likelihood estimation ,simulation ,Electronic computers. Computer science ,QA75.5-76.95 ,Probabilities. Mathematical statistics ,QA273-280 - Abstract
In this article, the Lagrange expansion of the second kind is used to generate a novel zero-truncated Katz distribution; we refer to it as the Lagrangian zero-truncated Katz distribution (LZTKD). Notably, the zero-truncated Katz distribution is a special case of this distribution. Along with the closed form expression of all its statistical characteristics, the LZTKD is proven to provide an adequate model for both underdispersed and overdispersed zero-truncated count datasets. Specifically, we show that the associated hazard rate function has increasing, decreasing, bathtub, or upside-down bathtub shapes. Moreover, we demonstrate that the LZTKD belongs to the Lagrangian distribution of the first kind. Then, applications of the LZTKD in statistical scenarios are explored. The unknown parameters are estimated using the well-reputed method of the maximum likelihood. In addition, the generalized likelihood ratio test procedure is applied to test the significance of the additional parameter. In order to evaluate the performance of the maximum likelihood estimates, simulation studies are also conducted. The use of real-life datasets further highlights the relevance and applicability of the proposed model.
- Published
- 2023
- Full Text
- View/download PDF
14. Generalized Unit Half-Logistic Geometric Distribution: Properties and Regression with Applications to Insurance
- Author
-
Suleman Nasiru, Christophe Chesneau, Abdul Ghaniyyu Abubakari, and Irene Dekomwine Angbing
- Subjects
unit distribution ,statistical algorithms ,power transformation ,risk survey data ,quantile regression ,Bayesian estimation ,Electronic computers. Computer science ,QA75.5-76.95 ,Probabilities. Mathematical statistics ,QA273-280 - Abstract
The use of distributions to model and quantify risk is essential in risk assessment and management. In this study, the generalized unit half-logistic geometric (GUHLG) distribution is developed to model bounded insurance data on the unit interval. The corresponding probability density function plots indicate that the related distribution can handle data that exhibit left-skewed, right-skewed, symmetric, reversed-J, and bathtub shapes. The hazard rate function also suggests that the distribution can be applied to analyze data with bathtubs, N-shapes, and increasing failure rates. Subsequently, the inferential aspects of the proposed model are investigated. In particular, Monte Carlo simulation exercises are carried out to examine the performance of the estimation method by using an algorithm to generate random observations from the quantile function. The results of the simulation suggest that the considered estimation method is efficient. The univariate application of the distribution and the multivariate application of the associated regression using risk survey data reveal that the model provides a better fit than the other existing distributions and regression models. Under the multivariate application, we estimate the parameters of the regression model using both maximum likelihood and Bayesian estimations. The estimates of the parameters for the two methods are very close. Diagnostic plots of the Bayesian method using the trace, ergodic, and autocorrelation plots reveal that the chains converge to a stationary distribution.
- Published
- 2023
- Full Text
- View/download PDF
15. Revisit of an Improved Wilker Type Inequality
- Author
-
Rupali Shinde, Christophe Chesneau, Nitin Darkunde, Sanjay Ghodechor, and Aditya Lagad
- Subjects
Mathematics ,QA1-939 - Abstract
In this article, we revisit an improved Wilker type inequality established in 2020. We fill some gaps in the existing proof and propose an alternative proof using the same mathematical ingredients. All the details are given for checking purposes.
- Published
- 2023
- Full Text
- View/download PDF
16. On improved fitting using a new probability distribution and artificial neural network: Application
- Author
-
Sanaa Al-Marzouki, Afaf Alrashidi, Christophe Chesneau, Mohammed Elgarhy, Rana H. Khashab, and Suleman Nasiru
- Subjects
Physics ,QC1-999 - Abstract
Statistical modeling and forecasting are crucial to understanding the depth of information in data from all sources. For precision purposes, researchers are always in search of ways to improve the quality of modeling and forecasting, whatever the complexity of the situation. To this end, new (probability) distributions and suitable forecasting methods are demanded. The first part of this paper contributes to this direction. Indeed, we introduce a modified version of the flexible Weibull distribution, called the modified flexible Weibull distribution. It is constructed by mixing the flexible Weibull distribution with the exponential T-X scheme. This strategy is winning; the new distribution has a larger panel of functionalities in comparison to those of the classical Weibull distribution, among other things. To check the quality of the fitting of the modified flexible Weibull distribution, two different datasets are analyzed. After analyzing these datasets, it is observed that the modified flexible Weibull distribution has improved fitting power compared to other similar distributions. Apart from this, the conventional time series model, namely, the autoregressive integrated moving average (ARIMA) model, and the modern artificial neural network (ANN) model are considered for forecasting results. Utilizing the two datasets discussed earlier, it was discovered that the ANN model is more effective than the traditional ARIMA model.
- Published
- 2023
- Full Text
- View/download PDF
17. Theoretical Advancements on a Few New Dependence Models Based on Copulas with an Original Ratio Form
- Author
-
Christophe Chesneau
- Subjects
copulas ,dependence model ,trigonometric functions ,inequalities ,correlation ,Engineering design ,TA174 - Abstract
Copulas are well-known tools for describing the relationship between two or more quantitative variables. They have recently received a lot of attention, owing to the variable dependence complexity that appears in heterogeneous modern problems. In this paper, we offer five new copulas based on a common original ratio form. All of them are defined with a single tuning parameter, and all reduce to the independence copula when this parameter is equal to zero. Wide admissible domains for this parameter are established, and the mathematical developments primarily rely on non-trivial limits, two-dimensional differentiations, suitable factorizations, and mathematical inequalities. The corresponding functions and characteristics of the proposed copulas are looked at in some important details. In particular, as common features, it is shown that they are diagonally symmetric, but not Archimedean, not radially symmetric, and without tail dependence. The theory is illustrated with numerical tables and graphics. A final part discusses the multi-dimensional variation of our original ratio form. The contributions are primarily theoretical, but they provide the framework for cutting-edge dependence models that have potential applications across a wide range of fields. Some established two-dimensional inequalities may be of interest beyond the purposes of this paper.
- Published
- 2023
- Full Text
- View/download PDF
18. A Collection of New Trigonometric- and Hyperbolic-FGM-Type Copulas
- Author
-
Christophe Chesneau
- Subjects
copulas ,dependence models ,trigonometric functions ,hyperbolic functions ,inequalities ,correlation ,Mathematics ,QA1-939 - Abstract
Copula analysis was created to explain the dependence of two or more quantitative variables. Due to the need for in-depth data analysis involving complex variable relationships, there is always a need for new copula models with original features. As a modern example, for the analysis of circular or periodic data types, trigonometric copulas are particularly attractive and recommended. This is, however, an underexploited topic. In this article, we propose a new collection of eight trigonometric and hyperbolic copulas, four based on the sine function and the others on the tangent function, all derived from the construction of the famous Farlie–Gumbel–Morgenstern copula. In addition to their original trigonometric and hyperbolic functionalities, the proposed copulas have the feature of depending on three parameters with complementary roles: one is a dependence parameter; one is a shape parameter; and the last can be viewed as an angle parameter. In our main findings, for each of the eight copulas, we determine a wide range of admissible values for these parameters. Subsequently, the capabilities, features, and functions of the new copulas are thoroughly examined. The shapes of the main functions of some copulas are illustrated graphically. Theoretically, symmetry in general, stochastic dominance, quadrant dependence, tail dependence, Archimedean nature, correlation measures, and inference on the parameters are investigated. Some copula shapes are illustrated with the help of figures. On the other hand, some two-dimensional inequalities are established and may be of separate interest.
- Published
- 2023
- Full Text
- View/download PDF
19. The Harris Extended Bilal Distribution with Applications in Hydrology and Quality Control
- Author
-
Radhakumari Maya, Muhammed Rasheed Irshad, Muhammed Ahammed, and Christophe Chesneau
- Subjects
Harris extended distributions ,bilal distribution ,hazard rate function ,quantile function ,maximum likelihood estimation ,acceptance sampling plans ,Mathematics ,QA1-939 - Abstract
In this research work, a new three-parameter lifetime distribution is introduced and studied. It is called the Harris extended Bilal distribution due to its construction from a mixture of the famous Bilal and Harris distributions, resulting from a branching process. The basic properties, such as the moment generating function, moments, quantile function, and Rényi entropy, are discussed. We show that the hazard rate function has ideal features for modeling increasing, upside-down bathtub, and roller-coaster data sets. In a second part, the Harris extended Bilal model is investigated from a statistical viewpoint. The maximum likelihood estimation is used to estimate the parameters, and a simulation study is carried out. The flexibility of the proposed model in a hydrological data analysis scenario is demonstrated using two practical data sets and compared with important competing models. After that, we establish an acceptance sampling plan that takes advantage of all of the features of the Harris extended Bilal model. The operating characteristic values, the minimum sample size that corresponds to the maximum possible defects, and the minimum ratios of lifetime associated with the producer’s risk are discussed.
- Published
- 2023
- Full Text
- View/download PDF
20. Farlie–Gumbel–Morgenstern Bivariate Moment Exponential Distribution and Its Inferences Based on Concomitants of Order Statistics
- Author
-
Sasikumar Padmini Arun, Christophe Chesneau, Radhakumari Maya, and Muhammed Rasheed Irshad
- Subjects
concomitants of order statistics ,moment exponential distribution ,inference ,Statistics ,HA1-4737 - Abstract
In this research, we design the Farlie–Gumbel–Morgenstern bivariate moment exponential distribution, a bivariate analogue of the moment exponential distribution, using the Farlie–Gumbel–Morgenstern approach. With the analysis of real-life data, the competitiveness of the Farlie–Gumbel–Morgenstern bivariate moment exponential distribution in comparison with the other Farlie–Gumbel–Morgenstern distributions is discussed. Based on the Farlie–Gumbel–Morgenstern bivariate moment exponential distribution, we develop the distribution theory of concomitants of order statistics and derive the best linear unbiased estimator of the parameter associated with the variable of primary interest (study variable). Evaluations are also conducted regarding the efficiency comparison of the best linear unbiased estimator relative to the respective unbiased estimator. Additionally, empirical illustrations of the best linear unbiased estimator with respect to the unbiased estimator are performed.
- Published
- 2023
- Full Text
- View/download PDF
21. A Novel Flexible Class of Intervened Poisson Distribution by Lagrangian Approach
- Author
-
Muhammed Rasheed Irshad, Mohanan Monisha, Christophe Chesneau, Radhakumari Maya, and Damodaran Santhamani Shibu
- Subjects
Lagrange expansion ,intervened Poisson distribution ,Lagrangian intervened Poisson distribution ,regression ,inverse transformation method ,Statistics ,HA1-4737 - Abstract
The zero-truncated Poisson distribution (ZTPD) generates a statistical model that could be appropriate when observations begin once at least one event occurs. The intervened Poisson distribution (IPD) is a substitute for the ZTPD, in which some intervention processes may change the mean of the rare events. These two zero-truncated distributions exhibit underdispersion (i.e., their variance is less than their mean). In this research, we offer an alternative solution for dealing with intervention problems by proposing a generalization of the IPD by a Lagrangian approach called the Lagrangian intervened Poisson distribution (LIPD), which in fact generalizes both the ZTPD and the IPD. As a notable feature, it has the ability to analyze both overdispersed and underdispersed datasets. In addition, the LIPD has a closed-form expression of all of its statistical characteristics, as well as an increasing, decreasing, bathtub-shaped, and upside-down bathtub-shaped hazard rate function. A consequent part is devoted to its statistical application. The maximum likelihood estimation method is considered, and the effectiveness of the estimates is demonstrated through a simulated study. To evaluate the significance of the new parameter in the LIPD, a generalized likelihood ratio test is performed. Subsequently, we present a new count regression model that is suitable for both overdispersed and underdispersed datasets using the mean-parametrized form of the LIPD. Additionally, the LIPD’s relevance and application are shown using real-world datasets.
- Published
- 2023
- Full Text
- View/download PDF
22. Theoretical Contributions to Three Generalized Versions of the Celebioglu–Cuadras Copula
- Author
-
Christophe Chesneau
- Subjects
two-dimensional copulas ,exponential function ,power functions ,correlation ,Electronic computers. Computer science ,QA75.5-76.95 ,Probabilities. Mathematical statistics ,QA273-280 - Abstract
Copulas are probabilistic functions that are being used more and more frequently to describe, examine, and model the interdependence of continuous random variables. Among the numerous proposed copulas, renewed interest has recently been shown in the so-called Celebioglu–Cuadras copula. It is mainly because of its simplicity, exploitable dependence properties, and potential for applicability. In this article, we contribute to the development of this copula by proposing three generalized versions of it, each involving three tuning parameters. The main results are theoretical: they consist of determining wide and manageable intervals of admissible values for the involved parameters. The proofs are mainly based on limit, differentiation, and factorization techniques as well as mathematical inequalities. Some of the configuration parameters are new in the literature, and original phenomena are revealed. Subsequently, the basic properties of the proposed copulas are studied, such as symmetry, quadrant dependence, various expansions, concordance ordering, tail dependences, medial correlation, and Spearman correlation. Detailed examples, numerical tables, and graphics are used to support the theory.
- Published
- 2023
- Full Text
- View/download PDF
23. Simulation analysis, properties and applications on a new Burr XII model based on the Bell-X functionalities
- Author
-
Ayed. R. A. Alanzi, Muhammad Imran, M. H. Tahir, Christophe Chesneau, Farrukh Jamal, Saima Shakoor, and Waqas Sami
- Subjects
burr xii distribution ,entropy measures ,estimation ,gasp ,quality parameter ,t-x family ,Mathematics ,QA1-939 - Abstract
In this article, we make mathematical and practical contributions to the Bell-X family of absolutely continuous distributions. As a main member of this family, a special distribution extending the modeling perspectives of the famous Burr XII (BXII) distribution is discussed in detail. It is called the Bell-Burr XII (BBXII) distribution. It stands apart from the other extended BXII distributions because of its flexibility in terms of functional shapes. On the theoretical side, a linear representation of the probability density function and the ordinary and incomplete moments are among the key properties studied in depth. Some commonly used entropy measures, namely Rényi, Havrda and Charvat, Arimoto, and Tsallis entropy, are derived. On the practical (inferential) side, the associated parameters are estimated using seven different frequentist estimation methods, namely the methods of maximum likelihood estimation, percentile estimation, least squares estimation, weighted least squares estimation, Cramér von-Mises estimation, Anderson-Darling estimation, and right-tail Anderson-Darling estimation. A simulation study utilizing all these methods is offered to highlight their effectiveness. Subsequently, the BBXII model is successfully used in comparisons with other comparable models to analyze data on patients with acute bone cancer and arthritis pain. A group acceptance sampling plan for truncated life tests is also proposed when an item's lifetime follows a BBXII distribution. Convincing results are obtained.
- Published
- 2023
- Full Text
- View/download PDF
24. A Novel Generalization of Zero-Truncated Binomial Distribution by Lagrangian Approach with Applications for the COVID-19 Pandemic
- Author
-
Muhammed Rasheed Irshad, Christophe Chesneau, Damodaran Santhamani Shibu, Mohanan Monisha, and Radhakumari Maya
- Subjects
Lagrangian zero-truncated binomial distribution ,index of dispersion ,maximum likelihood method ,generalized likelihood ratio test ,COVID-19 ,simulation ,Statistics ,HA1-4737 - Abstract
The importance of Lagrangian distributions and their applicability in real-world events have been highlighted in several studies. In light of this, we create a new zero-truncated Lagrangian distribution. It is presented as a generalization of the zero-truncated binomial distribution (ZTBD) and hence named the Lagrangian zero-truncated binomial distribution (LZTBD). The moments, probability generating function, factorial moments, as well as skewness and kurtosis measures of the LZTBD are discussed. We also show that the new model’s finite mixture is identifiable. The unknown parameters of the LZTBD are estimated using the maximum likelihood method. A broad simulation study is executed as an evaluation of the well-established performance of the maximum likelihood estimates. The likelihood ratio test is used to assess the effectiveness of the third parameter in the new model. Six COVID-19 datasets are used to demonstrate the LZTBD’s applicability, and we conclude that the LZTBD is very competitive on the fitting objective.
- Published
- 2022
- Full Text
- View/download PDF
25. Bayesian and non-Bayesian estimations of truncated inverse power Lindley distribution under progressively type-II censored data with applications
- Author
-
Mohammed Elgarhy, Aned Al Mutairi, Amal S. Hassan, Christophe Chesneau, and Alaa H. Abdel-Hamid
- Subjects
Physics ,QC1-999 - Abstract
In this article, we introduce and study the truncated inverse power Lindley distribution. The aim is to transpose the remarkable flexibility of the two-parameter inverse power Lindley distribution to the interval [0,1]. The corresponding probability density function has the potential to be unimodal, decreasing, right-skewed, and heavy-tailed. On the other hand, the hazard rate function can be increasing, N-shaped, or U-shaped. These shapes’ versatility enables accurate representation and analysis of proportional or percentage data across a wide range of applications, such as survival analysis, reliability, and uncertainty modeling. Several statistical features, such as the mode, quantiles, Bowley’s skewness, Moor’s kurtosis, MacGillivray’s skewness, moments, inverse moments, incomplete moments, and probability-weighted moments, are computed. In practice, for the estimation of the model parameters from truncated data under the progressively type-II censoring scheme, the maximum likelihood, maximum product spacing, and Bayesian approaches are used. The Tierney–Kadane approximation and Markov chain Monte Carlo techniques are employed to produce the Bayesian estimates under the squared error loss function. We present some simulation results to evaluate these approaches. Four applications based on real-world datasets—one of them is on times of infection, the second is on failure times, and the other two are on the rate of inflation in Asia and Africa—explain the significance of the new truncated model in comparison to some reputed comparable models, such as the inverse power Lindley, Kumaraswamy, truncated power Lomax, beta, truncated Weibull, unit-Weibull, Kumaraswamy Kumaraswamy, and exponentiated Kumaraswamy models.
- Published
- 2023
- Full Text
- View/download PDF
26. Bayesian inference using MCMC algorithm of sine truncated Lomax distribution with application
- Author
-
Mohammed. Elgarhy, Najwan Alsadat, Amal S. Hassan, and Christophe Chesneau
- Subjects
Physics ,QC1-999 - Abstract
This study makes a significant contribution to the creation of a versatile trigonometric extension of the well-known truncated Lomax distribution. Specifically, we construct a novel one-parameter distribution known as the sine truncated Lomax (STLo) distribution using characteristics from the sine generalized family of distributions. Quantiles, moments, stress–strength reliability, some information measures, residual moments, and reversed residual moments are a few of the crucial elements and characteristics we explored in our research. The flexibility of the STLo distribution in terms of the forms of the hazard rate and probability density functions illustrates how effectively it is able to match many types of data. Maximum likelihood and Bayesian estimation techniques are used to estimate the model parameter. The squared error loss function is employed in the Bayesian approach. To evaluate how various estimates behave, a Monte Carlo simulation study is carried out with the aid of a useful algorithm. Additionally, the STLo distribution has a good fit, making it a viable option when compared to certain other competing models using specific criteria to describe the given dataset.
- Published
- 2023
- Full Text
- View/download PDF
27. On the Type-I Half-logistic Distribution and Related Contributions: A Review
- Author
-
Mustapha Muhammad, M.H. Tahir, Lixia Liu, Farrukh Jamal, Christophe Chesneau, and Badamasi Abba
- Subjects
Probabilities. Mathematical statistics ,QA273-280 ,Statistics ,HA1-4737 - Abstract
The half-logistic (HL) distribution is a widely considered statistical model for studying lifetime phenomena arising in science, engineering, finance, and biomedical sciences. One of its weaknesses is that it has a decreasing probability density function and an increasing hazard rate function only. Due to that, researchers have been modifying the HL distribution to have more functional ability. This article provides an extensive overview of the HL distribution and its generalization (or extensions). The recent advancements regarding the HL distribution have led to numerous results in modern theory and statistical computing techniques across science and engineering. This work extended the body of literature in a summarized way to clarify some of the states of knowledge, potentials, and important roles played by the HL distribution and related models in probability theory and statistical studies in various areas and applications. In particular, at least sixty-seven flexible extensions of the HL distribution have been proposed in the past few years. We give a brief introduction to these distributions, emphasizing model parameters, properties derived, and the estimation method. Conclusively, there is no doubt that this summary could create a consensus between various related results in both theory and applications of the HL-related models to develop an interest in future studies.
- Published
- 2023
- Full Text
- View/download PDF
28. Different estimation methods for the generalized unit half-logistic geometric distribution: Using ranked set sampling
- Author
-
Najwan Alsadat, Amal S. Hassan, Ahmed M. Gemeay, Christophe Chesneau, and Mohammed Elgarhy
- Subjects
Physics ,QC1-999 - Abstract
The generalized unit half-logistic geometric distribution (GUHLGD) is a modern two-parameter unit distribution with attractive shape flexibility for the corresponding probability density and hazard rate functions. Due to its versatility, it may be used to model a variety of current bounded real-world datasets. On the other hand, an effective sampling strategy for both parametric and non-parametric inferences is the ranked set sampling (RSS) method. This article focuses on estimating the parameters of the GUHLGD based on the RSS method as well as the simple random sampling (SRS) method. Eleven traditional estimation methods are taken into consideration, including the percentile, Cramér–von Mises, maximum likelihood, Anderson–Darling, right-tailed Anderson–Darling, left-tailed Anderson–Darling, least squares, weighted least squares, minimum spacing absolute-log distance, maximum product of spacing, and minimum spacing absolute distance methods. A Monte Carlo simulation is employed to compare the performance of the resultant estimates based on some accuracy measures. We draw the conclusion that, for both sampling procedures, the maximum likelihood estimation methodology is the best option among the rest based on the partial and total ranking measures. The estimates based on the RSS method are more efficient than the others based on the SRS method. Results from actual data further support the advantage of the RSS design over the SRS design.
- Published
- 2023
- Full Text
- View/download PDF
29. Risk factors identification of COVID‐19 patients with chronic obstructive pulmonary disease: A retrospective study in Punjab‐Pakistan
- Author
-
Muhammad Muneeb Hassan, M. H. Tahir, Muhammad Ameeq, Farrukh Jamal, John T. Mendy, and Christophe Chesneau
- Subjects
COPD ,COVID‐19 ,Cox proportional hazard model and negative binomial distribution ,SARS‐CoV‐2 ,Immunologic diseases. Allergy ,RC581-607 - Abstract
Abstract Background Accessibility to the immense collection of studies on noncommunicable diseases related to coronavirus disease of 2019 (COVID‐19) and severe acute respiratory syndrome coronavirus 2 (SARS‐CoV‐2) is an immediate focus of researchers. However, there is a scarcity of information about chronic obstructed pulmonary disease (COPD), which is associated with a high rate of infection in COVID‐19 patients. Moreover, by combining the effects of the SARS‐CoV‐2 on COPD patients, we may be able to overcome formidable obstacles factors, and diagnosis influencers. Materials and Methods A retrospective study of 280 patients was conducted at DHQ Hospital Muzaffargarh in Punjab, Pakistan. Negative binomial regression describes the risk of fixed successive variables. The association is described by the Cox proportional hazard model and the model coefficient is determined through log‐likelihood observation. Patients with COPD had their survival and mortality plotted on Kaplan–Meier curves. Results The increased risk of death in COPD patients was due to the effects of variables such as cough, lower respiratory tract infection (LRTI), tuberculosis (TB), and body‐aches being 1.369, 0.693, 0.170, and 0.217 times higher at (95% confidence interval [CI]: 0.747–1.992), (95% CI: 0.231–1.156), (95% CI: 0.008–0.332), and (95% CI: −0.07 to 0.440) while it decreased 0.396 in normal condition. Conclusion We found that the symptoms of COPD (cough, LRTI, TB, and bodyaches) are statistically significant in patients who were most infected by SARS‐CoV‐2.
- Published
- 2023
- Full Text
- View/download PDF
30. Poisson Extended Exponential Distribution with Associated INAR(1) Process and Applications
- Author
-
Radhakumari Maya, Christophe Chesneau, Anuresha Krishna, and Muhammed Rasheed Irshad
- Subjects
extended exponential distribution ,overdispersion ,simulation ,regression model ,INAR(1) process ,Statistics ,HA1-4737 - Abstract
The significance of count data modeling and its applications to real-world phenomena have been highlighted in several research studies. The present study focuses on a two-parameter discrete distribution that can be obtained by compounding the Poisson and extended exponential distributions. It has tractable and explicit forms for its statistical properties. The maximum likelihood estimation method is used to estimate the unknown parameters. An extensive simulation study was also performed. In this paper, the significance of the proposed distribution is demonstrated in a count regression model and in a first-order integer-valued autoregressive process, referred to as the INAR(1) process. In addition to this, the empirical importance of the proposed model is proved through three real-data applications, and the empirical findings indicate that the proposed INAR(1) model provides better results than other competitive models for time series of counts that display overdispersion.
- Published
- 2022
- Full Text
- View/download PDF
31. Wheat Yield Prediction in India Using Principal Component Analysis-Multivariate Adaptive Regression Splines (PCA-MARS)
- Author
-
B. M. Nayana, Kolla Rohit Kumar, and Christophe Chesneau
- Subjects
MARS ,principal component analysis ,regression ,wheat prediction ,Agriculture (General) ,S1-972 ,Engineering (General). Civil engineering (General) ,TA1-2040 - Abstract
Crop yield forecasting is becoming more essential in the current scenario when food security must be assured, despite the problems posed by an increasingly globalized community and other environmental challenges such as climate change and natural disasters. Several factors influence crop yield prediction, which has complex non-linear relationships. Hence, to study these relationships, machine learning methodologies have been increasingly adopted from conventional statistical methods. With wheat being a primary and staple food crop in the Indian community, ensuring the country’s food security is crucial. In this paper, we study the prediction of wheat yield for India overall and the top wheat-producing states with a comparison. To accomplish this, we use Multivariate Adaptive Regression Splines (MARS) after extracting the main features by Principal Component Analysis (PCA) considering the parameters such as area under cultivation and production for the years 1962–2018. The performance is evaluated by error analyses such as RMSE, MAE, and R2. The best-fitted MARS model is chosen using cross-validation and user-defined parameter optimization. We find that the MARS model is well suited to India as a whole and other top wheat-producing states. A comparative result is obtained on yield prediction between India overall and other states, wherein the state of Rajasthan has a better model than other major wheat-producing states. This research will emphasize the importance of improved government decision-making as well as increased knowledge and robust forecasting among Indian farmers in various states.
- Published
- 2022
- Full Text
- View/download PDF
32. Theoretical Study of Some Angle Parameter Trigonometric Copulas
- Author
-
Christophe Chesneau
- Subjects
copula ,two-dimensional modeling ,trigonometric function ,multivariate distributions ,dependence ,Engineering design ,TA174 - Abstract
Copulas are important probabilistic tools to model and interpret the correlations of measures involved in real or experimental phenomena. The versatility of these phenomena implies the need for diverse copulas. In this article, we describe and investigate theoretically new two-dimensional copulas based on trigonometric functions modulated by a tuning angle parameter. The independence copula is, thus, extended in an original manner. Conceptually, the proposed trigonometric copulas are ideal for modeling correlations into periodic, circular, or seasonal phenomena. We examine their qualities, such as various symmetry properties, quadrant dependence properties, possible Archimedean nature, copula ordering, tail dependences, diverse correlations (medial, Spearman, and Kendall), and two-dimensional distribution generation. The proposed copulas are fleshed out in terms of data generation and inference. The theoretical findings are supplemented by some graphical and numerical work. The main results are proved using two-dimensional inequality techniques that can be used for other copula purposes.
- Published
- 2022
- Full Text
- View/download PDF
33. Stochastic Transportation Problem with Multichoice Random Parameter
- Author
-
Talari Ganesh, K. K. Paidipati, and Christophe Chesneau
- Subjects
Mathematics ,QA1-939 ,Applied mathematics. Quantitative methods ,T57-57.97 - Abstract
This paper deals with the situation of multiple random choices along with multiple objective functions of the transportation problem. Due to the uncertainty in the environment, the choices of the cost coefficients are considered multichoice random parameters. The other parameters (supply and demand) are replaced by random variables with Gaussian distributions, and each multichoice parameter alternative is treated as a random variable. In this paper, the Newton divided difference interpolation technique is used to convert the multichoice parameter into a single choice in the objective function. Then, the chance-constrained method is applied to transform the probabilistic constraints into deterministic constraints. Due to the consideration of multichoices in the objective function, the expectation minimization model is used to get the deterministic form. Moreover, the fuzzy programming approach with the membership function is utilized to convert the multiobjective function into a single-objective function. A case study is also illustrated for a better understanding of the methodology.
- Published
- 2023
- Full Text
- View/download PDF
34. A New Discrete Distribution: Properties, Characterizations, Modeling Real Count Data, Bayesian and Non-Bayesian Estimations
- Author
-
Haitham Mosad Yousof, Christophe Chesneau, Gholamhossein Hamedani, and Mohamed Ibrahim
- Subjects
discretization ,characterizations ,discrete burr-hatke distribution ,bayesian estimation ,metropolis hastings ,markov chain monte carlo ,maximum likelihood ,cramer-von-mises. ,Statistics ,HA1-4737 - Abstract
In this work, a new discrete distribution which includes the discrete Burr-Hatke distribution is defined and studied. Relevant statistical properties are derived. The probability mass function of the new distribution can be "right skewed" with different shapes, bimodal and "uniformed". Also, the corresponding hazard rate function can be "monotonically decreasing", "upside down", "monotonically increasing", "upside down increasing", and "upside down-constant-increasing". A numerical analysis for the mean, variance, skewness, kurtosis and the index of dispersion is presented. The new distribution could be useful in the modeling of "under-dispersed" or "overdispersed" count data. Certain characterizations of the new distribution are presented. These characterizations are based on the conditional expectation of a certain function of the random variable and in terms of the hazard rate function. Bayesian and non-Bayesian estimation methods are considered. Numerical simulations for comparing Bayesian and non-Bayesian estimation methods are performed. The new model is applied for modeling carious teeth data and counts of cysts of kidneys data.
- Published
- 2021
- Full Text
- View/download PDF
35. A Flexible Dispersed Count Model Based on Bernoulli Poisson–Lindley Convolution and Its Regression Model
- Author
-
Hassan S. Bakouch, Christophe Chesneau, Radhakumari Maya, Muhammed Rasheed Irshad, Sreedeviamma Aswathy, and Najla Qarmalah
- Subjects
discrete statistical model ,dispersion index ,hazard rate function ,parameter estimation ,simulation ,regression ,Mathematics ,QA1-939 - Abstract
Count data are encountered in real-life dealings. More understanding of such data and the extraction of important information about the data require some statistical analysis or modeling. One innovative technique to increase the modeling flexibility of well-known distributions is to use the convolution of random variables. This study examines the distribution that results from adding two independent random variables, one with the Bernoulli distribution and the other with the Poisson–Lindley distribution. The considered distribution is named as the two-parameter Bernoulli–Poisson–Lindley distribution. Many of its statistical properties are investigated, such as moments, survival and hazard rate functions, mode, dispersion behavior, mean deviation about the mean, and parameter inference based on the maximum likelihood method. To evaluate the effectiveness of the bias and mean square error of the produced estimates, a simulation exercise is carried out. Then, applications to two practical data sets are given. Finally, we construct a flexible count data regression model based on the proposed distribution with two practical examples.
- Published
- 2023
- Full Text
- View/download PDF
36. Sampling Plan for the Kavya–Manoharan Generalized Inverted Kumaraswamy Distribution with Statistical Inference and Applications
- Author
-
Najwan Alsadat, Amal S. Hassan, Mohammed Elgarhy, Christophe Chesneau, and Ahmed R. El-Saeed
- Subjects
Kavya–Manoharan generated family ,generalized inverse Kumaraswamy distribution ,entropy ,maximum product of spacing ,Bayesian estimation ,Mathematics ,QA1-939 - Abstract
In this article, we introduce the Kavya–Manoharan generalized inverse Kumaraswamy (KM-GIKw) distribution, which can be presented as an improved version of the generalized inverse Kumaraswamy distribution with three parameters. It contains numerous referenced lifetime distributions of the literature and a large panel of new ones. Among the essential features and attributes covered in our research are quantiles, moments, and information measures. In particular, various entropy measures (Rényi, Tsallis, etc.) are derived and discussed numerically. The adaptability of the KM-GIKw distribution in terms of the shapes of the probability density and hazard rate functions demonstrates how well it is able to fit different types of data. Based on it, an acceptance sampling plan is created when the life test is truncated at a predefined time. More precisely, the truncation time is intended to represent the median of the KM-GIKw distribution with preset factors. In a separate part, the focus is put on the inference of the KM-GIKw distribution. The related parameters are estimated using the Bayesian, maximum likelihood, and maximum product of spacings methods. For the Bayesian method, both symmetric and asymmetric loss functions are employed. To examine the behaviors of various estimates based on criterion measurements, a Monte Carlo simulation research is carried out. Finally, with the aim of demonstrating the applicability of our findings, three real datasets are used. The results show that the KM-GIKw distribution offers superior fits when compared to other well-known distributions.
- Published
- 2023
- Full Text
- View/download PDF
37. A New Asymmetric Modified Topp–Leone Distribution: Classical and Bayesian Estimations under Progressive Type-II Censored Data with Applications
- Author
-
Mohammed Elgarhy, Najwan Alsadat, Amal S. Hassan, Christophe Chesneau, and Alaa H. Abdel-Hamid
- Subjects
maximum likelihood and Bayesian estimations ,inverse Topp–Leone distribution ,maximum product spacing estimation ,Markov chain Monte Carlo ,entropy and extropy ,progressive type-II censoring ,Mathematics ,QA1-939 - Abstract
In this article, a new modified asymmetric Topp–Leone distribution is created and developed from a theoretical and inferential point of view. It has the feature of extending the remarkable flexibility of a special one-shape-parameter lifetime distribution, known as the inverse Topp–Leone distribution, to the bounded interval [0, 1]. The probability density function of the proposed truncated distribution has the potential to be unimodal and right-skewed, with different levels of asymmetry. On the other hand, its hazard rate function can be increasingly shaped. Some important statistical properties are examined, including several different measures. In practice, the estimation of the model parameters under progressive type-II censoring is considered. To achieve this aim, the maximum likelihood, maximum product of spacings, and Bayesian approaches are used. The Markov chain Monte Carlo approach is employed to produce the Bayesian estimates under the squared error and linear exponential loss functions. Some simulation studies to evaluate these approaches are discussed. Two applications based on real-world datasets—one on the times of infection, and the second dataset is on trading economics credit rating—are considered. Thanks to its flexible asymmetric features, the new model is preferable to some known comparable models.
- Published
- 2023
- Full Text
- View/download PDF
38. A Transmuted Modified Power-Generated Family of Distributions with Practice on Submodels in Insurance and Reliability
- Author
-
Sidra Naz, Laila A. Al-Essa, Hassan S. Bakouch, and Christophe Chesneau
- Subjects
statistical model ,transmuted family ,modified power distribution ,estimation ,actuarial measures ,simulation ,Mathematics ,QA1-939 - Abstract
In this article, we propose a new transmuted modified power-generalized family of distributions constructed from the transmuted-generated and modified power-generated families. The proposed approach is flexible and provides a tradeoff between the two baseline families. For a prime study, we identify the main characteristics of the new transmuted modified power family, such as the asymptotic results, quantile function, series representation, and the various kinds of moment measures. By using the exponential distribution as the baseline, a new three-parameter lifetime distribution is constructed. The associated probability functions (density and hazard rate) are flexible and have a variety of asymmetric shapes, which make them attractive for statistical purposes. In particular, for the related probability density function, reversed-J, unimodal, and right-skewed shapes are observed. Measures relating to risk theory are also computed, such as the value at risk and the expected shortfall. By using both simulation analysis and the maximum likelihood approach, the estimation of the model parameters is evaluated. The effectiveness of the proposed model is demonstrated by two real-world cases (one in insurance and the other in reliability), and we show that it yields better fits when compared to other extended models connected to the exponential model.
- Published
- 2023
- Full Text
- View/download PDF
39. Similarity-Based Predictive Models: Sensitivity Analysis and a Biological Application with Multi-Attributes
- Author
-
Jeniffer D. Sanchez, Leandro C. Rêgo, Raydonal Ospina, Víctor Leiva, Christophe Chesneau, and Cecilia Castro
- Subjects
biological data ,coefficient of variation ,data science ,distance measures ,estimation methods ,Monte Carlo simulation ,Biology (General) ,QH301-705.5 - Abstract
Predictive models based on empirical similarity are instrumental in biology and data science, where the premise is to measure the likeness of one observation with others in the same dataset. Biological datasets often encompass data that can be categorized. When using empirical similarity-based predictive models, two strategies for handling categorical covariates exist. The first strategy retains categorical covariates in their original form, applying distance measures and allocating weights to each covariate. In contrast, the second strategy creates binary variables, representing each variable level independently, and computes similarity measures solely through the Euclidean distance. This study performs a sensitivity analysis of these two strategies using computational simulations, and applies the results to a biological context. We use a linear regression model as a reference point, and consider two methods for estimating the model parameters, alongside exponential and fractional inverse similarity functions. The sensitivity is evaluated by determining the coefficient of variation of the parameter estimators across the three models as a measure of relative variability. Our results suggest that the first strategy excels over the second one in effectively dealing with categorical variables, and offers greater parsimony due to the use of fewer parameters.
- Published
- 2023
- Full Text
- View/download PDF
40. Wavelet Estimation of Regression Derivatives for Biased and Negatively Associated Data
- Author
-
Junke Kou and Christophe Chesneau
- Subjects
Regression derivatives estimation ,negatively associated ,Lp risk ,wavelets ,Statistics ,HA1-4737 ,Probabilities. Mathematical statistics ,QA273-280 - Abstract
This paper considers the estimation of the derivatives of a regression function based on biased data. The main feature of the study is to explore the case where the data comes from a negatively associated process. In this context, two different wavelet estimators are introduced: a linear wavelet estimator and a nonlinear wavelet estimator using the hard thresholding rule. Their theoretical performance is evaluated by determining sharp rates of convergence under Lp risk, assuming that the unknown function of interest belongs to a ball of Besov spaces Bsp,q (ℝ). The obtained results extend some existing works on biased data in the independent case to the negatively associated case.
- Published
- 2022
- Full Text
- View/download PDF
41. On a new modeling strategy: The logarithmically-exponential class of distributions
- Author
-
Abdulhakim A. Al-Babtain, Ibrahim Elbatal, Christophe Chesneau, and Mohammed Elgarhy
- Subjects
probability distribution theory ,lomax distribution ,information measures ,reliability measures ,parametric estimation ,simulation ,data fitting ,Mathematics ,QA1-939 - Abstract
In this paper, a promising modeling strategy for data fitting is derived from a new general class of univariate continuous distributions. This class is governed by an original logarithmically-exponential one-parameter transformation which has the ability to enhance some modeling capabilities of any parental distribution. In relation to the current literature, it appears to be a "limit case" of the well-established truncated generalized Fréchet generated class. In addition, it offers a natural alternative to the famous odd inverse exponential generated class. Some special distributions are presented, with particular interest in a novel heavy-tailed three-parameter distribution based on the Lomax distribution. Functional equivalences, modes analysis, stochastic ordering, functional expansions, moment measures, information measures and reliability measures are derived. From generic or real data, our modeling strategy is based on the new class combined with the maximum likelihood approach. We apply this strategy to the introduced modified Lomax model. The efficiency of the three parameter estimates is validated by a simulation study. Subsequently, two referenced real data sets are adjusted according to the rules of the art; the first one containing environmental data and the second one, financial data. In particular, we show that the proposed model is preferable to four concurrents also derived from the Lomax model, including the odd inverse exponential Lomax model.
- Published
- 2021
- Full Text
- View/download PDF
42. Prediction of Rice Cultivation in India—Support Vector Regression Approach with Various Kernels for Non-Linear Patterns
- Author
-
Kiran Kumar Paidipati, Christophe Chesneau, B. M. Nayana, Kolla Rohith Kumar, Kalpana Polisetty, and Chinnarao Kurangi
- Subjects
rice cultivation ,food security ,prediction ,support vector regression with kernels ,RMSE and MAE ,Agriculture (General) ,S1-972 ,Engineering (General). Civil engineering (General) ,TA1-2040 - Abstract
The prediction of rice yields plays a major role in reducing food security problems in India and also suggests that government agencies manage the over or under situations of production. Advanced machine learning techniques are playing a vital role in the accurate prediction of rice yields in dealing with nonlinear complex situations instead of traditional statistical methods. In the present study, the researchers made an attempt to predict the rice yield through support vector regression (SVR) models with various kernels (linear, polynomial, and radial basis function) for India overall and the top five rice producing states by considering influence parameters, such as the area under cultivation and production, as independent variables for the years 1962–2018. The best-fitted models were chosen based on the cross-validation and hyperparameter optimization of various kernel parameters. The root-mean-square error (RMSE) and mean absolute error (MAE) were calculated for the training and testing datasets. The results revealed that SVR with various kernels fitted to India overall, as well as the major rice producing states, would explore the nonlinear patterns to understand the precise situations of yield prediction. This study will be helpful for farmers as well as the central and state governments for estimating rice yield in advance with optimal resources.
- Published
- 2021
- Full Text
- View/download PDF
43. Tan-G class of trigonometric distributions and its applications
- Author
-
Luciano Souza, Wilson Rosa de O. Júnior, Cícero Carlos R. de Brito, Christophe Chesneau, Renan L. Fernandes, and Tiago A. E. Ferreira
- Subjects
trigonometric class of distribution ,tangent function ,burr xii distribution ,maximum likelihood estimation ,entropy ,Mathematics ,QA1-939 - Abstract
In this paper, we introduce a new general class of trigonometric distributions based on the tangent function, called the Tan-G class. A mathematical procedure of the Tan-G class is carried out, including expansions for the probability density function, moments, central moments and Rényi entropy. The estimates are acquired in a non-closed form by the maximum likelihood estimation method. Then, an emphasis is put on a particular member of this class defined with the Burr XII distribution as baseline, called the Tan-BXII distribution. The inferential properties of the Tan-BXII model are investigated. Finally, the Tan-BXII model is applied to a practical data set, illustrating the interest of the Tan-G class for the practitioner.
- Published
- 2021
- Full Text
- View/download PDF
44. Modeling Asymmetric Volatility: A News Impact Curve Approach
- Author
-
Debopam Rakshit, Ranjit Kumar Paul, Md Yeasin, Walid Emam, Yusra Tashkandy, and Christophe Chesneau
- Subjects
GARCH ,News Impact Curve ,volatility ,Mathematics ,QA1-939 - Abstract
Seasonal production, weather abnormalities, and high perishability introduce a high degree of volatility to potato prices. Price volatility is said to be asymmetric when positive and negative shocks of the same magnitude affect it in a dissimilar way. GARCH is a symmetric model, and it cannot capture asymmetric price volatility. EGARCH, APARCH, and GJR-GARCH models are popularly used to capture asymmetric price volatility. In this paper, an attempt is made to model the price volatility of the weekly wholesale modal price of potatoes for the Agra, Ahmedabad, Bengaluru, Delhi, Kolkata, and Mumbai markets using the above-mentioned models. The News Impact Curves (NICs) are derived from the fitted models, which confirmed the presence of asymmetry in the price volatility. To this end, NICs are used to describe the degree of asymmetry in volatility present in different markets.
- Published
- 2023
- Full Text
- View/download PDF
45. Bayesian Subset Selection of Seasonal Autoregressive Models
- Author
-
Ayman A. Amin, Walid Emam, Yusra Tashkandy, and Christophe Chesneau
- Subjects
SAR models ,SSVS procedure ,posterior analysis ,mixture–normal ,Mathematics ,QA1-939 - Abstract
Seasonal autoregressive (SAR) models have many applications in different fields, such as economics and finance. It is well known in the literature that these models are nonlinear in their coefficients and that their Bayesian analysis is complicated. Accordingly, choosing the best subset of these models is a challenging task. Therefore, in this paper, we tackled this problem by introducing a Bayesian method for selecting the most promising subset of the SAR models. In particular, we introduced latent variables for the SAR model lags, assumed model errors to be normally distributed, and adopted and modified the stochastic search variable selection (SSVS) procedure for the SAR models. Thus, we derived full conditional posterior distributions of the SAR model parameters in the closed form, and we then introduced the Gibbs sampler, along with SSVS, to present an efficient algorithm for the Bayesian subset selection of the SAR models. In this work, we employed mixture–normal, inverse gamma, and Bernoulli priors for the SAR model coefficients, variance, and latent variables, respectively. Moreover, we introduced a simulation study and a real-world application to evaluate the accuracy of the proposed algorithm.
- Published
- 2023
- Full Text
- View/download PDF
46. Wavelets in Combination with Stochastic and Machine Learning Models to Predict Agricultural Prices
- Author
-
Sandip Garai, Ranjit Kumar Paul, Debopam Rakshit, Md Yeasin, Walid Emam, Yusra Tashkandy, and Christophe Chesneau
- Subjects
decomposition ,interrelations ,machine learning ,nonlinearity ,wavelet filters ,Mathematics ,QA1-939 - Abstract
Wavelet decomposition in signal processing has been widely used in the literature. The popularity of machine learning (ML) algorithms is increasing day by day in agriculture, from irrigation scheduling and yield prediction to price prediction. It is quite interesting to study wavelet-based stochastic and ML models to appropriately choose the most suitable wavelet filters to predict agricultural commodity prices. In the present study, some popular wavelet filters, such as Haar, Daubechies (D4), Coiflet (C6), best localized (BL14), and least asymmetric (LA8), were considered. Daily wholesale price data of onions from three major Indian markets, namely Bengaluru, Delhi, and Lasalgaon, were used to illustrate the potential of different wavelet filters. The performance of wavelet-based models was compared with that of benchmark models. It was observed that, in general, the wavelet-based combination models outperformed other models. Moreover, wavelet decomposition with the Haar filter followed by application of the random forest (RF) model gave better prediction accuracy than other combinations as well as other individual models.
- Published
- 2023
- Full Text
- View/download PDF
47. On Some Inequalities for the Generalized Euclidean Operator Radius
- Author
-
Mohammad W. Alomari, Gabriel Bercu, Christophe Chesneau, and Hala Alaqad
- Subjects
Euclidean operator radius ,numerical radius ,self-adjoint operator ,Mathematics ,QA1-939 - Abstract
In the literature, there are many criteria to generalize the concept of a numerical radius; one of the most recent and interesting generalizations is the so-called generalized Euclidean operator radius, which reads: ωpT1,⋯,Tn:=supx=1∑i=1nTix,xp1/p,p≥1, for all Hilbert space operators T1,⋯,Tn. Simply put, it is the numerical radius of multivariable operators. This study establishes a number of new inequalities, extensions, and generalizations for this type of numerical radius. More precisely, by utilizing the mixed Schwarz inequality and the extension of Furuta’s inequality, some new refinement inequalities are obtained for the numerical radius of multivariable Hilbert space operators. In the case of n=1, the resulting inequalities could be considered extensions and generalizations of the classical numerical radius.
- Published
- 2023
- Full Text
- View/download PDF
48. Introduction to the Special Issue in Symmetry Titled 'Symmetry in Statistics and Data Science'
- Author
-
Christophe Chesneau
- Subjects
n/a ,Mathematics ,QA1-939 - Abstract
In order to introduce this Special Issue, some motivational facts are given [...]
- Published
- 2023
- Full Text
- View/download PDF
49. An Efficient Stress–Strength Reliability Estimate of the Unit Gompertz Distribution Using Ranked Set Sampling
- Author
-
Najwan Alsadat, Amal S. Hassan, Mohammed Elgarhy, Christophe Chesneau, and Rokaya Elmorsy Mohamed
- Subjects
Unit-Gompertz distribution ,percentiles method ,right tail Anderson–Darling ,maximum product spacing ,maximum likelihood ,ranked set sampling ,Mathematics ,QA1-939 - Abstract
In this paper, the estimation of the stress–strength reliability is taken into account when the stress and strength variables have unit Gompertz distributions with a similar scale parameter. The consideration of the unit Gompertz distribution in this context is because of its intriguing symmetric and asymmetric properties that can accommodate various histogram proportional-type data shapes. As the main contribution, the reliability estimate is determined via seven frequentist techniques using the ranked set sampling (RSS) and simple random sampling (SRS). The proposed methods are the maximum likelihood, least squares, weighted least squares, maximum product spacing, Cramér–von Mises, Anderson–Darling, and right tail Anderson–Darling methods. We perform a simulation work to evaluate the effectiveness of the recommended RSS-based estimates by using accuracy metrics. We draw the conclusion that the reliability estimates in the maximum product spacing approach have the lowest value compared to other approaches. In addition, we note that the RSS-based estimates are superior to those obtained by a comparable SRS approach. Additional results are obtained using two genuine data sets that reflect the survival periods of head and neck cancer patients.
- Published
- 2023
- Full Text
- View/download PDF
50. Inverse Unit Teissier Distribution: Theory and Practical Examples
- Author
-
Najwan Alsadat, Mohammed Elgarhy, Kadir Karakaya, Ahmed M. Gemeay, Christophe Chesneau, and M. M. Abd El-Raouf
- Subjects
Teissier distribution ,estimation ,simulation ,inverse distribution ,data fitting ,Mathematics ,QA1-939 - Abstract
In this paper, we emphasize a new one-parameter distribution with support as [1,+∞). It is constructed from the inverse method applied to an understudied one-parameter unit distribution, the unit Teissier distribution. Some properties are investigated, such as the mode, quantiles, stochastic dominance, heavy-tailed nature, moments, etc. Among the strengths of the distribution are the following: (i) the closed-form expressions and flexibility of the main functions, and in particular, the probability density function is unimodal and the hazard rate function is increasing or unimodal; (ii) the manageability of the moments; and, more importantly, (iii) it provides a real alternative to the famous Pareto distribution, also with support as [1,+∞). Indeed, the proposed distribution has different functionalities but also benefits from the heavy-right-tailed nature, which is demanded in many applied fields (finance, the actuarial field, quality control, medicine, etc.). Furthermore, it can be used quite efficiently in a statistical setting. To support this claim, the maximum likelihood, Anderson–Darling, right-tailed Anderson–Darling, left-tailed Anderson–Darling, Cramér–Von Mises, least squares, weighted least-squares, maximum product of spacing, minimum spacing absolute distance, and minimum spacing absolute-log distance estimation methods are examined to estimate the unknown unique parameter. A Monte Carlo simulation is used to compare the performance of the obtained estimates. Additionally, the Bayesian estimation method using an informative gamma prior distribution under the squared error loss function is discussed. Data on the COVID mortality rate and the timing of pain relief after receiving an analgesic are considered to illustrate the applicability of the proposed distribution. Favorable results are highlighted, supporting the importance of the findings.
- Published
- 2023
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.