950 results on '"62P30"'
Search Results
2. A Spatial-statistical model to analyse historical rutting data
- Author
-
Jourdain, Natoya O. A. S., Steinsland, Ingelin, Birkhez-Shami, Mamoona, Vedvik, Emil, Olsen, William, Gryteselv, Dagfin, Siebert, Doreen, and Klein-Paste, Alex
- Subjects
Statistics - Applications ,Statistics - Computation ,Statistics - Methodology ,62P30 - Abstract
Pavement rutting poses a significant challenge in flexible pavements, necessitating costly asphalt resurfacing. To address this issue comprehensively, we propose an advanced Bayesian hierarchical framework of latent Gaussian models with spatial components. Our model provides a thorough diagnostic analysis, pinpointing areas exhibiting unexpectedly high rutting rates. Incorporating spatial and random components, and important explanatory variables like annual average daily traffic (traffic intensity), asphalt type, rut depth and lane width, our proposed models account for and estimate the influence of these variables on rutting. This approach not only quantifies uncertainties and discerns locations at the highest risk of requiring maintenance, but also uncover spatial dependencies in rutting (millimetre/year). We apply our models to a data set spanning eleven years (2010-2020). Our findings emphasise the systematic unexplained spatial rutting effect, where some of the rutting variability is accounted for by spatial components, asphalt type, in conjunction with traffic intensity, is also found to be the primary driver of rutting. Furthermore, the spatial dependencies uncovered reveal road sections experiencing more than 1 millimeter of rutting beyond annual expectations. This leads to a halving of the expected pavement lifespan in these areas. Our study offers valuable insights, presenting maps indicating expected rutting, and identifying locations with accelerated rutting rates, resulting in a reduction in pavement life expectancy of at least 10 years., Comment: 29 pages, 5 figures and 3 tables in main text, 2 tables and 4 figures in appendix 4 figures; and 3 tables in supplementary
- Published
- 2024
3. Process incapability index for autocorrelated data in the presence of measurement errors.
- Author
-
Bera, Kuntal and Anis, M. Z.
- Subjects
- *
MEASUREMENT errors , *MANUFACTURING processes , *PRODUCT quality , *MEASURING instruments - Abstract
Process incapability index is a useful tool for measuring the performance of a process to maintain the quality of products. In this article, we discuss some statistical properties of the estimator of process incapability index when process observations are autocorrelated and contaminated by measurement errors. In many industry, it has been proven that observed quality characteristics for consecutive items are autocorrelated. Moreover, some amount of measurement errors is always present in the observed data. Therefore, it is important to analyze the statistical properties of the estimator under the combined effect of autocorrelation and measurement error for commenting approximately on the ability of a manufacturing process. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. On the Averaging and Closure of Fiber Orientation Tensors in Virtual Process Chains.
- Author
-
Krauß, Constantin, Bauer, Julian Karl, Mitsch, Johannes, Böhlke, Thomas, and Kärger, Luise
- Subjects
FIBER orientation ,FLOW simulations ,FIBER-reinforced plastics ,EQUATIONS - Abstract
Fiber orientation tensors (FOT) are widely used to approximate statistical orientation distributions of fibers within fiber-reinforced polymers. The design process of components made of such fiber-reinforced composites is usually accompanied by a virtual process chain. In this virtual process chain, process-induced FOT are computed in a flow simulation and transferred to the structural simulation. Within the structural simulation, effective macroscopic properties are identified based on the averaged information contained in the FOT. Solving the field equations in flow simulations as well as homogenization of effective stiffnesses necessitates the application of a closure scheme, computing higher-order statistical moments based on assumptions. Additionally, non-congruent spatial discretizations require an intermediate mapping operation. This mapping operation is required, if the discretization, i.e., mesh, of the flow simulation differs from the discretization of the structural simulation. The main objective of this work is to give an answer to the question: Does the sequence of closure and mapping influence the achieved results? It will turn out, that the order influences the result, raising the consecutive question: Which order is beneficial? Both questions are addressed by deriving a quantification of the closure-related uncertainty. The two possible sequences, mapping followed by closure and closure followed by mapping, yield strongly different results, with the magnitude of the deviation even exceeding the magnitude of a reference result. Graphical consideration reveals that for both transversely isotropic and planar FOT-input, invalid results occur if the mapping takes place prior to closure. This issue is retrieved by orientation averaging stiffness tensors. As a by-product, we explicitly define for the first time the admissible parameter space of orthotropic fourth-order fiber orientation tensors and define a distance measure in this parameter space. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Models, metrics, and their formulas for typical electric power system resilience events
- Author
-
Dobson, Ian
- Subjects
Electrical Engineering and Systems Science - Systems and Control ,Mathematics - Probability ,Statistics - Applications ,62P30 - Abstract
Poisson process models are defined in terms of their rates for outage and restore processes in power system resilience events. These outage and restore processes easily yield the performance curves that track the evolution of resilience events, and the area, nadir, and duration of the performance curves are standard resilience metrics. This letter analyzes typical resilience events by analyzing the area, nadir, and duration of mean performance curves. Explicit and intuitive formulas for these metrics are derived in terms of the Poisson process model parameters, and these parameters can be estimated from utility data. This clarifies the calculation of metrics of typical resilience events, and shows what they depend on. The metric formulas are derived with lognormal, exponential, or constant rates of restoration. The method is illustrated with a typical North American transmission event. Similarly nice formulas are obtained for the area metric for empirical power system data.
- Published
- 2023
- Full Text
- View/download PDF
6. How Generative AI models such as ChatGPT can be (Mis)Used in SPC Practice, Education, and Research? An Exploratory Study
- Author
-
Megahed, Fadel M., Chen, Ying-Ju, Ferris, Joshua A., Knoth, Sven, and Jones-Farmer, L. Allison
- Subjects
Computer Science - Machine Learning ,62P30 ,G.3 ,G.4 ,J.2 ,J.6 - Abstract
Generative Artificial Intelligence (AI) models such as OpenAI's ChatGPT have the potential to revolutionize Statistical Process Control (SPC) practice, learning, and research. However, these tools are in the early stages of development and can be easily misused or misunderstood. In this paper, we give an overview of the development of Generative AI. Specifically, we explore ChatGPT's ability to provide code, explain basic concepts, and create knowledge related to SPC practice, learning, and research. By investigating responses to structured prompts, we highlight the benefits and limitations of the results. Our study indicates that the current version of ChatGPT performs well for structured tasks, such as translating code from one language to another and explaining well-known concepts but struggles with more nuanced tasks, such as explaining less widely known terms and creating code from scratch. We find that using new AI tools may help practitioners, educators, and researchers to be more efficient and productive. However, in their current stages of development, some results are misleading and wrong. Overall, the use of generative AI models in SPC must be properly validated and used in conjunction with other methods to ensure accurate results., Comment: 30 pages, 20 figures
- Published
- 2023
- Full Text
- View/download PDF
7. Online Continual Learning via the Knowledge Invariant and Spread-out Properties
- Author
-
Han, Ya-nan and Liu, Jian-wei
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,62P30 ,I.2.4 - Abstract
The goal of continual learning is to provide intelligent agents that are capable of learning continually a sequence of tasks using the knowledge obtained from previous tasks while performing well on prior tasks. However, a key challenge in this continual learning paradigm is catastrophic forgetting, namely adapting a model to new tasks often leads to severe performance degradation on prior tasks. Current memory-based approaches show their success in alleviating the catastrophic forgetting problem by replaying examples from past tasks when new tasks are learned. However, these methods are infeasible to transfer the structural knowledge from previous tasks i.e., similarities or dissimilarities between different instances. Furthermore, the learning bias between the current and prior tasks is also an urgent problem that should be solved. In this work, we propose a new method, named Online Continual Learning via the Knowledge Invariant and Spread-out Properties (OCLKISP), in which we constrain the evolution of the embedding features via Knowledge Invariant and Spread-out Properties (KISP). Thus, we can further transfer the inter-instance structural knowledge of previous tasks while alleviating the forgetting due to the learning bias. We empirically evaluate our proposed method on four popular benchmarks for continual learning: Split CIFAR 100, Split SVHN, Split CUB200 and Split Tiny-Image-Net. The experimental results show the efficacy of our proposed method compared to the state-of-the-art continual learning algorithms., Comment: 30 pages
- Published
- 2023
- Full Text
- View/download PDF
8. Safe machine learning.
- Author
-
Giudici, Paolo
- Abstract
The rapid development of artificial intelligence applications based on machine learning is creating not only many opportunities but also risks. The recent regulatory and political debate, at the international level, emphasizes the urgent need to develop appropriate statistical methods that can measure the safety and the risks of AI applications. In line with this emerging need, the aim of this work is to launch a call for discussion papers in the
Statistics journal on safe machine learning methods. To this aim, we define which metrics to develop and an example of a recent proposal, along with its advantages and disadvantages [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
9. Online RSSI selection strategy for indoor positioning in low-effort training scenarios.
- Author
-
Pinto, Braulio and Oliveira, Horacio
- Subjects
- *
INDOOR positioning systems , *WIRELESS sensor networks , *K-nearest neighbor classification , *LEAST squares - Abstract
Indoor positioning has been extensively studied for at least the past twenty years. In the list of the most common solutions, those based on the Received Strength Signal Indicator (RSSI) have gained importance due to the simplicity of RSSI as well as the fact that it is available in several wireless sensor networks. In this work, we propose SeALS (Selection Strategy of Access Points with Least Squares Estimation), a new RSSI-based indoor positioning system using Bluetooth Low-Energy (BLE) access points, whose accuracy is improved by a new selection strategy of collected RSSI combined with the Ordinary Least Squares (OLS) estimation method. The main advantage of the proposed solution is the fact that it requires less time in the training phase allied with better system accuracy if compared to traditional methods. The proposed system is validated in a large-scale, real-world scenario, and the obtained results for the positioning error are reduced by up to 13% concerning the pure OLS method, and by up to 30% concerning the widely deployed K-Nearest Neighbors technique. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Analyzing stress-strength reliability δ=P[U<V<W]: a Bayesian and frequentist perspective with Burr-XII distribution under progressive Type-II censoring.
- Author
-
Nayal, Amit Singh, Singh, Bhupendra, Tripathi, Vrijesh, and Tyagi, Abhishek
- Abstract
This research focuses on estimating the stress-strength reliability in a system characterized by the influence of two random stresses on its strength, employing both frequentist and Bayesian approaches. The reliability of such systems is represented by the function δ = P (U < V < W) , where V denotes the system's strength, and U and W represent the stresses. The analysis is performed under a progressive Type-II censoring scheme, considering the random variables U, V, and W as independent and following the Burr-XII distribution. In a frequentist setup, both the maximum likelihood estimator and the maximum product spacings estimator of δ have been obtained. In the Bayesian paradigm, the Bayes estimator of δ under the squared error loss function is derived utilizing the Markov chain Monte Carlo method, considering independent gamma priors for the unknown parameters. In addition, asymptotic confidence intervals and highest probability density credible intervals for δ are also formulated. An extensive simulation experiment is carried out to compare the performances of the different developed estimators. Finally, a real-life application is presented to demonstrate the practical applicability of the proposed theory. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Classifying one-parameter Fréchet populations on the basis of a non-linear fixed effects model.
- Author
-
Baratnia, Mohammad and Doostparast, Mahdi
- Subjects
- *
FIXED effects model , *HOMOGENEITY - Abstract
This article deals with the problem of classifying Fréchet populations by proposing a non-linear fixed effects model. Various properties of the proposed model are investigated in detail. Estimations of the effects are achieved in explicit expressions and stochastic properties of the estimators are discussed. Two procedures for testing the homogeneity of populations are proposed. Illustrative examples are also given. The proposed model is an alternative method for situations in which the existing parametric methods such as generalized linear models provide poor efficiency and factor effects are on the response non-linearly. The proposed model is also applicable in situations where the mean response does not exist. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. Distribution-free Phase II triple EWMA control chart for joint monitoring the process location and scale parameters.
- Author
-
Alevizakos, Vasileios, Chatterjee, Kashinath, and Koukouvinos, Christos
- Subjects
- *
QUALITY control charts , *MONTE Carlo method , *CONTINUOUS distributions , *MANUFACTURING processes , *MOVING average process - Abstract
Distribution-free or nonparametric control charts are used for monitoring the process parameters when there is a lack of knowledge about the underlying distribution. In this paper, we investigate a single distribution-free triple exponentially weighted moving average control chart based on the Lepage statistic (referred as TL chart) for simultaneously monitoring shifts in the unknown location and scale parameters of a univariate continuous distribution. The design and implementation of the proposed chart are discussed using time-varying and steady-state control limits for the zero-state case. The run-length distribution of the TL chart is evaluated by performing Monte Carlo simulations. The performance of the proposed chart is compared to those of the existing EWMA-Lepage (EL) and DEWMA-Lepage (DL) charts. It is observed that the TL chart with a time-varying control limit is superior to its competitors, especially for small to moderate shifts in the process parameters. We also provide a real example from a manufacturing process to illustrate the application of the proposed chart. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. An ADMM based method for underdetermined box-constrained integer least squares problems.
- Author
-
Chang, Xiao-Wen and Ma, Tianchi
- Subjects
- *
LEAST squares , *COMMUNICATION barriers - Abstract
To solve underdetermined box-constrained integer least squares (UBILS) problems, we propose an integer-constrained alternating direction method of multipliers (IADMM), which can be much more accurate than the ADMM method. To guarantee to find the optimal solution, then we incorporate IADMM to DTS, a tree search method, to make the latter more efficient. Numerical tests show that the combined method IADMM-DTS can be much faster than the original DTS method. Finally, we apply the combined method to a practical communication problem. Numerical results indicate that IADMM-DTS typically performs better than the commercial solvers CPLEX and MOSEK in terms of both efficiency and accuracy, and it can be used as an alternative to the commercial solver Gurobi for UBILS problems. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. Max-EWMA chart for time and magnitude monitoring using generalized exponential distribution.
- Author
-
Talib, Ayesha, Ali, Sajid, and Shah, Ismail
- Subjects
- *
DISTRIBUTION (Probability theory) , *MOVING average process , *STANDARD deviations , *QUANTILES - Abstract
Joint monitoring of time and magnitude is an important issue in many industrial and non-industrial fields and several proposals exist in the literature. The aim of this article is to propose a new maximum exponentially weighted moving average (Max-EWMA) chart assuming generalized exponential distribution for time as well as magnitude. As the test-statistic of Max-EWMA chart consists of two EWMA statistics, it is easy to identify the out-of-control signal raised by the chart. The performance of the chart is evaluated using different run length characteristics including, average run length (ARL), standard deviation of run length (SDRL), and quantiles of the run length distribution. Besides extensive simulation studies, a real data set is also used to show the practicality of the proposed chart. The results show that the proposed chart is efficient in detecting different size of small to moderate shifts. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Detecting shifts in Conway–Maxwell–Poisson profile with deviance residual-based CUSUM and EWMA charts under multicollinearity.
- Author
-
Mammadova, Ulduz and Özkale, M. Revan
- Abstract
Monitoring profiles with count responses is a common situation in industrial processes and for a count distributed process, the Conway–Maxwell–Poisson (COM-Poisson) regression model yields better outcomes for under- and overdispersed count variables. In this study, we propose CUSUM and EWMA charts based on the deviance residuals obtained from the COM-Poisson model, which are fitted by the PCR and r–k class estimators. We conducted a simulation study to evaluate the effect of additive and multiplicative types shifts in various shift sizes, the number of predictor, and several dispersion levels and to compare the performance of the proposed control charts with control charts in the literature in terms of average run length and standard deviation of run length. Moreover, a real data set is also analyzed to see the performance of the newly proposed control charts. The results show the superiority of the newly proposed control charts against some competitors, including CUSUM and EWMA control charts based on ML, PCR, and ridge deviance residuals in the presence of multicollinearity. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. A Seismologically Consistent Surface Rupture Length Model for Unbounded and Width-Limited Event
- Author
-
Lavrentiadis, Grigorios, Wang, Yongfei, Abrahamson, Norman A., Bozorgnia, Yousef, and Goulet, Christine
- Subjects
Physics - Geophysics ,Astrophysics - Earth and Planetary Astrophysics ,62P30 - Abstract
A new surface-rupture-length ($SRL$) relationship as a function of magnitude ($\mathbf{M}$), fault thickness, and fault dip angle is presented in this paper. The objective of this study is to model the change in scaling between unbounded and width-limited ruptures. This is achieved through the use of seismological-theory-based relationships for the average displacement scaling and the aid of dynamic fault rupture simulations to constrain the rupture width scaling. The empirical dataset used in the development of this relationship is composed of $123$ events ranging from $\mathbf{M}~5$ to $8.1$ and $SRL~1.1$ to $432~km$. The dynamic rupture simulations dataset includes $554$ events ranging from $\mathbf{M}~4.9$ to $8.2$ and $SRL~1$ to $655~km$. For the average displacement ($\bar{D}$) scaling, three simple models and two composite models were evaluated. The simple average displacement models were: a square root of the rupture area ($\sqrt{A}$), a down-dip width ($W$), and a rupture length ($L$) proportional model. The two composite models followed a $\sqrt{A}$ scaling for unbounded ruptures and transitioned to $W$ and $L$ scaling for width-limited events, respectively. The empirical data favors a $\bar{D} \sim \sqrt{A}$ scaling for both unbounded and width-limited ruptures. The proposed model exhibits better predictive performance compared to linear $\log(SLR)\sim\mathbf{M}$ type models, especially in the large magnitude range, which is dominated by width-limited events. A comparison with existing $SRL$ models shows consistent scaling at different magnitude ranges that is believed to be the result of the different magnitude ranges in the empirical dataset of the published relationships., Comment: 21 pages, 11 figures
- Published
- 2022
17. Two-stage group reliability acceptance sampling plan for Lindley distribution
- Author
-
Mondal, Anumita and Maiti, Sudhansu S.
- Published
- 2024
- Full Text
- View/download PDF
18. Odd Log-Logistic XGamma Model: Bayesian and Classical Estimation with Risk Analysis Utilizing Reinsurance Revenues Data
- Author
-
Ranjbar, Vahid, Alizadeh, Morad, Afshari, Mahmoud, and Yousof, Haitham M.
- Published
- 2024
- Full Text
- View/download PDF
19. Julius Kruopis: Pioneer of the applications of mathematical statistics in Lithuania
- Author
-
Bagdonavičius, Vilijandas, Čekanavičius, Vydas, Levulienė, Rūta, and Vaitkus, Pranas
- Published
- 2024
- Full Text
- View/download PDF
20. Inferences on stress-strength reliability R=P[Y<X<Z] for progressively type-II censored Weibull half logistic distribution
- Author
-
Nayal, Amit Singh, Tyagi, Abhishek, Rathi, Shubhi, and Singh, Bhupendra
- Published
- 2024
- Full Text
- View/download PDF
21. Using Sum-Product Networks to Assess Uncertainty in Deep Active Learning
- Author
-
Khosravani, Mohamadsadegh and Zilles, Sandra
- Subjects
Computer Science - Machine Learning ,Computer Science - Information Theory ,62P30 ,I.5.0 ,I.4.10 - Abstract
The success of deep active learning hinges on the choice of an effective acquisition function, which ranks not yet labeled data points according to their expected informativeness. Many acquisition functions are (partly) based on the uncertainty that the current model has about the class label of a point, yet there is no generally agreed upon strategy for computing such uncertainty. This paper proposes a new and very simple approach to computing uncertainty in deep active learning with a Convolutional Neural Network (CNN). The main idea is to use the feature representation extracted by the CNN as data for training a Sum-Product Network (SPN). Since SPNs are typically used for estimating the distribution of a dataset, they are well suited to the task of estimating class probabilities that can be used directly by standard acquisition functions such as max entropy and variational ratio. The effectiveness of our method is demonstrated in an experimental study on several standard benchmark datasets for image classification, where we compare it to various state-of-the-art methods for assessing uncertainty in deep active learning., Comment: 15 pages,9 figures, 4 tables
- Published
- 2022
22. Two-Sided EWMA Charts for Monitoring Double Bounded Processes
- Author
-
Lafatzi, Argyro and Rakitzis, Athanasios
- Subjects
Statistics - Applications ,Statistics - Methodology ,62P30 - Abstract
In this work, we study the performance of two-sided EWMA charts for monitoring double bounded processes using individual observations. Specifically, the term double bounded refers to observations in the interval (0, 1) and thus, these charts are suitable for monitoring rates, proportions and percentages. There are several models that can be used to describe this kind of data (and the respective processes, as well) such as the Beta distribution, the Simplex distribution and the Unit Gamma distribution. For each of these three models, we provide the statistical design and the performance of the proposed EWMA charts. Also, apart from providing the appropriate values for the design parameters of each chart, we investigate how much the performance of the EWMA schemes is affected by using the values of control limits which have not been calculated under the true model. Finally, an illustrative example is also presented., Comment: 39 pages, 4 figures, 18 tables
- Published
- 2022
23. Evaluation and optimal designing of a two-level skip-lot sampling plan for resubmitted lots.
- Author
-
Murugeswari, N., Jeyadurga, P., and Balamurali, S.
- Subjects
- *
MARKOV processes , *BINOMIAL distribution , *POISSON distribution , *CONSUMERS , *TOTAL quality management - Abstract
In the quality management system, skip-lot sampling is a prominent beneficial sampling plan concerning the economic aspect since it requires minimum expenses for inspection. This study attempts to propose an innovative sampling plan by assimilating the concept of a resampling scheme in a two-level skip-lot sampling plan of type 3, that is, SkSP-2L.3, and the resultant plan is designated as SkSP-2L.3-R with a single sampling plan as the reference plan. Using Markov chain formulation, the performance measures are derived. The designing methodology is provided to determine the plan parameters of the SkSP-2L.3-R plan to minimize the average sample number by using two points on the operating characteristic curve approach. Tables are constructed to select the plan parameters under binomial distribution and Poisson distribution by considering various combinations of producer and consumer quality levels along with respective producer and consumer risks. An illustrative example is provided to depict the execution of the proposed SkSP-2L.3-R plan. The comparative study substantiates that the proposed SkSP-2L.3-R plan with a single sampling plan as the reference plan has outstanding performance than some other existing plans under binomial and Poisson distributions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
24. MEWMA based control charts with runs rules for monitoring multivariate simple linear regression profiles in Phase II.
- Author
-
Ahmadi Karavigh, Mohammad Hassan and Amiri, Amirhossein
- Subjects
- *
QUALITY control charts , *STANDARD deviations , *AUTOMOBILE industry - Abstract
In this article, three control charts including MEWMA, MEWMA/ χ 2 and DMEWMA equipped with proposed runs rules schemes, are developed to improve the performance of existing control charts in monitoring multivariate simple linear profiles. Using five criteria of average run length, percentile of the run length, standard deviation of run length, relative mean index and relative standard deviation index the behavior of the suggested control schemes has been assessed. The results of our study demonstrate that the suggested schemes have a better performance compared to current charts for detecting shifts. Ultimately, using an actual calibration application case in the automotive industry, the suggested control charts applicability is depicted. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
25. Generalised score distribution: underdispersed continuation of the beta-binomial distribution.
- Author
-
Ćmiel, Bogdan, Nawała, Jakub, Janowski, Lucjan, and Rusek, Krzysztof
- Subjects
DISTRIBUTION (Probability theory) ,RANDOM variables ,MAXIMUM likelihood statistics ,BINOMIAL distribution ,LATENT variables ,GAUSSIAN distribution - Abstract
Consider a class of discrete probability distributions with a limited support. A typical example of such support is some variant of a Likert scale, with a response mapped to either the { 1 , 2 , ... , 5 } or { - 3 , - 2 , ... , 2 , 3 } set. Such type of data is common for Multimedia Quality Assessment but can also be found in many other research fields. For modelling such data a latent variable approach is usually used (e.g., Ordered Probit). In many cases it is convenient or even necessary to avoid latent variable approach (e.g., when dealing with too small sample size). To avoid it the proper class of discrete distributions is needed. The main idea of this paper is to propose a family of discrete probability distributions with only two parameters that play the same role as the parameters of the normal distribution. We call the new class the Generalised Score Distribution (GSD). The proposed GSD class covers the entire set of possible means and variances, for any fixed and finite support. Furthermore, the GSD class can be treated as an underdispersed continuation of a reparametrized beta-binomial distribution. The GSD class parameters are intuitive and can be easily estimated by the method of moments. We also offer a Maximum Likelihood Estimation (MLE) algorithm for the GSD class and evidence that the class properly describes response distributions coming from 24 Multimedia Quality Assessment experiments. At last, we show that the GSD class can be represented as a sum of dichotomous zero–one random variables, which points to an interesting interpretation of the class. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
26. Estimation of fixed-accuracy confidence interval of the stress–strength reliability for inverse Pareto distribution using two-stage sampling technique.
- Author
-
Joshi, Neeraj, Bapat, Sudeep R., and Sengupta, Raghu Nandan
- Subjects
- *
PARETO distribution , *SAMPLING (Process) , *ASYMPTOTIC efficiencies , *DISTRIBUTION (Probability theory) , *INSURANCE claims , *CONFIDENCE intervals , *PARAMETER estimation , *RELIABILITY in engineering - Abstract
In recent years, several probability distributions have been introduced in the literature to analyze the data exhibiting an upside-down bathtub–shaped failure rate; an inverse Pareto distribution (IPD) is an appropriate choice among them. For stress–strength reliability models, estimation of parameters is an interesting area of research. In this article, we estimate the stress–strength reliability parameter R = P (X > Y) (where X and Y are strength and stress variables, respectively) of the IPD, whereby we focus on the problem of fixed-accuracy confidence interval estimation of R. It is established that the proposed interval estimation problem cannot be solved with the help of any fixed sample technique. As a result, we propose a two-stage sequential sampling strategy (which reduces the sample size significantly) to solve the given estimation problem. We obtain the expressions of several exact operating characteristics associated with our two-stage sampling technique. We also establish that the proposed two-stage procedure enjoys interesting first-order asymptotic properties. The detailed simulation analyses support our theoretical findings, and two real data sets based on insurance claims reinforce the practical utility of the proposed technique. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
27. Nonparametric multivariate control chart for numerical and categorical variables.
- Author
-
Jin, Jiayun and Loosveldt, Geert
- Subjects
- *
QUALITY control charts , *PROBABILITY density function , *STATISTICAL process control , *PRINCIPAL components analysis , *FALSE alarms , *SAMPLE size (Statistics) - Abstract
Multivariate statistical process control (MSPC) was developed for the monitoring of variables that are either all numerical or all categorical. In the present paper, we describe a nonparametric control scheme that can be used to monitor a mixture of numerical and categorical variables simultaneously. It integrates Principal Component Analysis Mix (PCA Mix), a multivariate statistical tool, with the conventional Hotelling T2 chart. To estimate the control limit for the PCA Mix-based T2 statistic, two nonparametric approaches – kernel density estimation (KDE) and bootstrap – are employed, because of the unknown nature of the underlying distribution. The simulation results demonstrate that with an appropriate number of principal components, both bootstrap and KDE exhibit convincing performance in terms of generating the same, or nearly the same, number of false alarms (ARL0) as expected, and being able to detect process shifts efficiently (ARL1). Compared with bootstrap, KDE is shown to work better with small sample sizes (n < 800) and to be slightly more sensitive to small shifts. However, the results also show the instability of the estimated nonparametric control limit when highly imbalanced categorical variables are included, which indicates the need for further research on this topic. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
28. Improving reliability system in radar model based on two parameters Weibull distribution
- Author
-
Adel A. El-Faheem, Muhammed Ashraf Denguir, and Abdelfattah Mustafa
- Subjects
Weibull distribution ,Reliability equivalence factors ,Radar system ,improving methods ,AMS Subject Classification: 62N05 ,62P30 ,Science - Abstract
In this paper, the two parameters Weibull distribution lifetime model has been used to improve the reliability of a system in radar structure. This model is flexible to any lifetime data analysis and describing all stages of failure rate. Redundant and reduction methods of improving the system reliability are provided, including cold, hot, and imperfect methods. The lifetimes of units are assumed to be independent and non-identical components. To distinguish between different methods, we computed the mean time to failure and γ -fractiles. Reliability equivalence factors are also derived. Simulation examples were given to differentiate between different improving methods.
- Published
- 2024
- Full Text
- View/download PDF
29. Process Visualization of Manufacturing Execution System (MES) Data
- Author
-
O'Neill, Meadhbh, Morgan, Jeff, and Burke, Kevin
- Subjects
Computer Science - Human-Computer Interaction ,Statistics - Applications ,62P30 - Abstract
Process visualizations of data from manufacturing execution systems (MESs) provide the ability to generate valuable insights for improved decision-making. Industry 4.0 is awakening a digital transformation where advanced analytics and visualizations are critical. Exploiting MESs with data-driven strategies can have a major impact on business outcomes. The advantages of employing process visualizations are demonstrated through an application to real-world data. Visualizations, such as dashboards, enable the user to examine the performance of a production line at a high level. Furthermore, the addition of interactivity facilitates the user to customize the data they want to observe. Evidence of process variability between shifts and days of the week can be investigated with the goal of optimizing production.
- Published
- 2022
- Full Text
- View/download PDF
30. Registration-free localization of defects in 3-D parts from mesh metrology data using functional maps
- Author
-
Zhao, Xueqi and del Castillo, Enrique
- Subjects
Statistics - Methodology ,62P30 - Abstract
Spectral Laplacian methods, widely used in computer graphics and manifold learning, have been recently proposed for the Statistical Process Control (SPC) of a sequence of manufactured parts, whose 3-dimensional metrology is acquired with non-contact sensors. These techniques provide an {\em intrinsic} solution to the SPC problem, that is, a solution exclusively based on measurements on the scanned surfaces or 2-manifolds without making reference to their ambient space. These methods, therefore, avoid the computationally expensive, non-convex registration step needed to align the parts, as required by previous methods for SPC based on 3-dimensional measurements. Once a SPC mechanism triggers and out-of-control alarm, however, an additional problem remains: that of locating where on the surface of the part that triggered the SPC alarm there is a significant shape difference with respect to either an in-control part or its nominal (CAD) design. In the past, only registration-based solutions existed for this problem. In this paper, we present a new registration-free solution to the part localization problem. Our approach uses a functional map between the manifolds to be compared, that is, a map between functions defined on each manifold based on intrinsic differential operators, in particular, the Laplace-Beltrami operator, in order to construct a point to point mapping between the two manifolds and be able to locate defects on the suspected part. A recursive partitioning algorithm is presented to define a region of interest on the surface of the part where defects are likely to occur, which results in considerable computational advantages. The functional map method involves a very large number of point-to-point comparisons based on noisy measurements, and a statistical thresholding method is presented to filter the false positives in the underlying massive multiple comparisons problem., Comment: 32 pages, 12 figures
- Published
- 2021
31. Overview and Introduction to Development of Non-Ergodic Earthquake Ground-Motion Models
- Author
-
Lavrentiadis, Grigorios, Abrahamson, Norman A., Nicolas, Kuehn M., Bozorgnia, Yousef, Goulet, Christine A., Babič, Anže, Macedo, Jorge, Dolšek, Matjaž, Gregor, Nicholas, Kottke, Albert R., Lacour, Maxime, Liu, Chenying, Meng, Xiaofeng, Phung, Van-Bang, Sung, Chih-Hsuan, and Walling, Melanie
- Subjects
Statistics - Applications ,Physics - Geophysics ,62P30 - Abstract
This paper provides an overview and introduction to the development of non-ergodic ground-motion models, GMMs. It is intended for a reader who is familiar with the standard approach for developing ergodic GMMs. It starts with a brief summary of the development of ergodic GMMs and then describes different methods that are used in the development of non-ergodic GMMs with an emphasis on Gaussian Process (GP) regression, as that is currently the method preferred by most researchers contributing to this special issue. Non-ergodic modeling requires the definition of locations for the source and site characterizing the systematic source and site effects; the non-ergodic domain is divided into cells for describing the systematic path effects. Modeling the cell-specific anelastic attenuation as a GP and considerations on constraints for extrapolation of the non-ergodic GMMs are also discussed. An updated unifying notation for non-ergodic GMMs is also presented, which has been adopted by the authors of this issue., Comment: 28 pages, 2 figures
- Published
- 2021
32. A Review and Critique of Auxiliary Information-Based Process Monitoring Methods
- Author
-
Saleh, Nesma A., Mahmoud, Mahmoud A., Woodall, William H., and Knoth, Sven
- Subjects
Statistics - Methodology ,62P30 - Abstract
We review the rapidly growing literature on auxiliary information-based (AIB) process monitoring methods. Under this approach, there is an assumption that the auxiliary variable, which is correlated with the quality variable of interest, has a known mean, or some other parameter, which cannot change over time. We demonstrate that violations of this assumption can have serious adverse effects both when the process is stable and when there has been a process shift. Some process shifts can become undetectable. We also show that the basic AIB approach is a special case of simple linear regression profile monitoring. The AIB charting techniques require strong assumptions. Based on our results, we warn against the use of AIB approach in quality control applications., Comment: 21 pages, 2 figures
- Published
- 2021
33. Divergent Effects of Factors on Crashes under Autonomous and Conventional Driving Modes Using A Hierarchical Bayesian Approach
- Author
-
Ren, Weixi, Yu, Bo, Chen, Yuren, Gao, Kun, and Bao, Shan
- Subjects
Statistics - Applications ,62P30 ,G.3.1 - Abstract
Influencing factors on crashes involved with autonomous vehicles (AVs) have been paid increasing attention. However, there is a lack of comparative analyses between influencing factors on crashes of AVs and human-driven vehicles. To fill this research gap, the study aims to explore the divergent effects of factors on crashes under autonomous and conventional driving modes. This study obtained 154 publicly available autonomous vehicle crash data (70 for the autonomous driving mode and 84 for the conventional driving mode), and 36 explanatory variables were extracted from three categories, including environment, roads, and vehicles. Then, a hierarchical Bayesian approach was applied to analyze the impacting factors on crash type and severity under both driving modes. The results showed that some factors affected both driving modes, but their degrees were different. For example, the presence of turning movement had a greater impact on the crash severity under the conventional driving mode, while the presence of turning movement led to a larger decrease in the likelihood of rear-end crashes under the autonomous driving mode. More influencing factors only had a significant impact on one of the driving modes. For example, in the autonomous driving mode, two sidewalks decreased the severity of crashes, and on-street parking was positively associated with rear-end crashes, but they were not significant in the conventional driving mode. This study could contribute to the understanding and development of autonomous driving systems and the better coordination between autonomous driving and conventional driving., Comment: 42 pages,10 figures
- Published
- 2021
34. A Non-ergodic Spectral Acceleration Ground Motion Model for California Developed with Random Vibration Theory
- Author
-
Lavrentiadis, Grigorios and Abrahamson, Norman A.
- Subjects
Statistics - Applications ,62P30 - Abstract
A new approach for creating a non-ergodic $PSA$ ground-motion model (GMM) is presented which account for the magnitude dependence of the non-ergodic effects. In this approach, the average $PSA$ scaling is controlled by an ergodic $PSA$ GMM, and the non-ergodic effects are captured with non-ergodic $PSA$ factors, which are the adjustment that needs to be applied to an ergodic $PSA$ GMM to incorporate the non-ergodic effects. The non-ergodic $PSA$ factors are based on $EAS$ non-ergodic effects and are converted to $PSA$ through Random Vibration Theory (RVT). The advantage of this approach is that it better captures the non-ergodic source, path, and site effects through the small magnitude earthquakes. Due to the linear properties of Fourier Transform, the $EAS$ non-ergodic effects of the small events can be applied directly to the large magnitude events. This is not the case for $PSA$, as response spectrum is controlled by a range of frequencies, making $PSA$ non-ergodic effects depended on the spectral shape which is magnitude dependent. Two $PSA$ non-ergodic GMMs are derived using the ASK14 and CY14 GMMs as backbone models, respectively. The non-ergodic $EAS$ effects are estimated with the LAK21 GMM. The RVT calculations are performed with the V75 peak factor model, the $D_{a0.05-0.85}$ estimate of AS96 for the ground-motion duration, and BT15 oscillator-duration model. The California subset of the NGAWest2 database is used for both models. The total aleatory standard deviation of the two non-ergodic $PSA$ GMMs is approximately $30$ to $35\%$ smaller than the total aleatory standard deviation of the corresponding ergodic $PSA$ GMMs. This reduction has a significant impact on hazard calculations at large return periods. In remote areas, far from stations and past events, the reduction of aleatory variability is accompanied by an increase of epistemic uncertainty., Comment: 32 pages, 34 figures
- Published
- 2021
35. Application of Additive Exponential Distribution in Design of X¯ Control Chart-Statistical and Economic Perspective
- Author
-
Kalisetti, Yogendra, Katneni, Nirupama Devi, and Kraleti, Srinivasa Rao
- Published
- 2024
- Full Text
- View/download PDF
36. A novel XGamma extension: applications and actuarial risk analysis under the reinsurance data
- Author
-
Alizadeh, Morad, Afshari, Mahmoud, Ranjbar, Vahid, Merovci, Faton, and Yousof, Haitham M.
- Published
- 2024
- Full Text
- View/download PDF
37. Auxiliary information based HEWMA chart using variable sampling interval
- Author
-
Ng, Peh Sang, Yeong, Wai Chung, Saha, Sajal, Lim, Sok Li, and Lau, Wai Kwan
- Published
- 2024
- Full Text
- View/download PDF
38. Measuring One-Sided Process Capability Index for Autocorrelated Data in the Presence of Random Measurement Errors.
- Author
-
Bera, Kuntal and Anis, M. Z.
- Subjects
PROCESS capability ,MEASUREMENT errors ,STATIONARY processes ,GAUSSIAN processes ,GAUSSIAN distribution ,CHEMICAL processes - Abstract
Many quality characteristics in manufacturing industry are of one sided specifications. The well-known process capability indices C P U and C P L are often used to measure the performance of such type of production process. It is usually assumed that process observations are independent and measurement system is free of errors. But actually in many industry it has been proven that auto-correlation is an inherent nature of the production process, especially for chemical processes. Moreover, even with the use of highly sophisticated advanced measuring instruments some amount of measurement error is always present in the observed data. Hence gauge measurement error also needs to be considered. In this paper we discuss some inferential properties of one-sided process capability indices for a stationary Gaussian process in the presence of measurement errors. As a particular case of a stationary Gaussian process, we discuss the case of a stationary AR (1) process where measurement error follows an independent Gaussian distribution. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
39. Quality Control Using Convolutional Neural Networks Applied to Samples of Very Small Size.
- Author
-
Chatzimichail, Rallou A. and Hatjimihail, Aristides T.
- Subjects
CONVOLUTIONAL neural networks ,QUALITY control ,ARTIFICIAL neural networks - Abstract
Artificial neural networks (NNs) have been extensively studied for their application to quality control (QC) to monitor the conformity of processes to quality specifications. However, the requirement of at least five QC measurements increases the associated costs. This study explores the potential of using NNs on samples of QC measurements of very small size. To achieve this, three one-dimensional (1-D) convolutional NNs (CNNs) were designed, trained, and tested on datasets of n-tuples of simulated, standardized, normally distributed QC measurements, where 2 ≤ n ≤ 4 . The performance of the designed CNNs was compared to that of statistical QC functions applied to samples of equal sizes, maintaining equal probabilities for false rejection. The results demonstrated that for n-tuples of QC measurements distributed as 풩 (0 , s 2) , where 1.2 < s ≤ 9.0 , the designed CNNs outperformed their statistical QC functions counterparts. Therefore, the use of 1-D CNNs applied to samples of two to four QC measurements can effectively enhance the detection of nonconformity of a process to quality specifications. This approach has the potential to significantly reduce the costs of QC measurements and improve the overall efficiency of the QC process. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
40. Design and Optimization of c-Control Chart Using a Triple Sampling Scheme.
- Author
-
Muñoz, José Jorge, Campuzano, Manuel J., and Deibe-Blanco, Verónica
- Subjects
QUALITY control charts ,DESIGN - Abstract
In this paper, a c-control chart using a triple sampling scheme (TS-c) is studied. The chart design, procedure, and a bi-objective optimization model are given to optimize the TS-c-chart. The Average Run Length for in-control and out-of-control ( ARL 1 ), and Average Sample Number metrics are calculated. A Comparison among TS-c, Fixed parameters c (FP-c), VSS-c, EWMA-c, and Double Sampling c (DS-c) control charts are carried out in terms of ARL 1 . The proposed TS-c-chart has lower ARL 1 values for detecting small and moderate shifts in the mean number of non-conformities in control compared with FP-c, VSS-c, EWMA-c, and DS-c. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
41. Modelling of Functional Profiles and Explainable Shape Shifts Detection: An Approach Combining the Notion of the Fréchet Mean with the Shape-Invariant Model.
- Author
-
Papayiannis, Georgios I., Psarakis, Stelios, and Yannacopoulos, Athanasios N.
- Subjects
- *
QUALITY control charts , *AIR pollutants - Abstract
A modelling framework suitable for detecting shape shifts in functional profiles combining the notion of the Fréchet mean and the concept of deformation models is developed and proposed. The generalized mean sense offered by the Fréchet mean notion is employed to capture the typical pattern of the profiles under study, while the concept of deformation models, and in particular of the shape-invariant model, allows for interpretable parameterizations of the profile's deviations from the typical shape. The EWMA-type control charts compatible with the functional nature of data and the employed deformation model are built and proposed, exploiting certain shape characteristics of the profiles under study with respect to the generalized mean sense, allowing for the identification of potential shifts concerning the shape and/or the deformation process. Potential shifts in the shape deformation process are further distinguished into significant shifts with respect to amplitude and/or the phase of the profile under study. The proposed modeling and shift detection framework is implemented to a real-world case study, where daily concentration profiles concerning air pollutants from an area in the city of Athens are modeled, while profiles indicating hazardous concentration levels are successfully identified in most cases. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
42. Homogeneously weighted moving average-variance control chart using auxiliary information.
- Author
-
Noor-ul-Amin, Muhammad and Arshad, Asma
- Subjects
- *
QUALITY control charts , *MONTE Carlo method , *MOVING average process - Abstract
In this study, a variance control chart is proposed with some modifications in terms of the rapid detection of process shifts by using the homogenously weighted moving average ( HWMA) statistic by using the information from an auxiliary variable. The proposed HWMA chart is a new memory type variance control chart with a three-parameter log transformation that operates by assigning the specified weights to the most recent value. All previous samples are allotted with an equal proportion of the remaining smoothing parameter weight. The remaining weights put their effect as a counterpart within the smoothing parameter and the current observations take the maximum of it. To study the performance of the proposed chart the respective run length properties are determined by Monte Carlo simulations and extensively presented in the tables. The comparison has been made with the existing exponentially weighted moving average ( EWMA) variance control chart. It is figured out that the proposed control chart is sensitive to process shift by using the auxiliary information with a high degree of correlation. The real-life dataset is selected from an industrial production environment to demonstrate the implementation of the proposed chart. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
43. Exploration of high-quality development path of tourism and culture industry in the context of smart tourism
- Author
-
Hou Weigang
- Subjects
intelligent tourism ,tourism and culture industry ,coupling coordination degree ,entropy value method ,62p30 ,Mathematics ,QA1-939 - Abstract
The emergence of intelligent tourism as a new trend in the travel industry has raised the bar for the excellent growth of the travel and cultural sectors. The merger of the artistic and tourist industries significantly advances tourism’s modernization and evolution. Finding a high-quality growth path for the tourist and culture industries within the context of intelligent tourism is a significant problem that must be overcome in order to achieve the two sectors’ in-depth integration and synergistic growth. A scientific assessment system and model analysis is necessary to uncover practical ways to support industry integration and upgrading based on the current industrial development status quo. A strong connection between this province’s cultural and tourism sectors is indicated by the coupling degrees generally higher than 0.85. Although the degree of coupling coordination has improved annually, many cities are still unable to achieve a higher level of coordination, according to an analysis. This study utilizes the degree of coupling coordination model as a basis for creating a coupling index for the culture and tourist industries and examining the growth situation of an area from 2015 to 2022. In a particular province, the median annual growth rates for the tourism and cultural sectors between 2015 and 2022 are 19.18% and 32.11%, respectively. This suggests that there is still more room for growth and integration between the two sectors in the context of intelligent tourism. The tourism industry is developing more slowly than the cultural industry. The research suggests ways to aid the integration and superior growth of the tourist and cultural sectors, including boosting infrastructure construction and enhancing employee knowledge.
- Published
- 2024
- Full Text
- View/download PDF
44. Innovation and Informatization of Civics Teaching Mode in Colleges and Universities under Integrated Transfer Learning Strategies
- Author
-
Zou Liegang
- Subjects
fmbcc-ft ,heterogeneous transfer learning ,pedagogical knowledge base ,source domain ,target domain ,62p30 ,Mathematics ,QA1-939 - Abstract
This study addresses the problem of predicting student performance by proposing a heterogeneous transfer learning method based on FMBCC-FT, which aims to accurately predict students’ Civics knowledge mastery. The study achieves effective knowledge transfer by defining university Civics teaching as the source domain and considering the university’s demand for students’ Civics dynamics as the target domain. After collecting and standardizing students’ Civics education data, this study constructed a knowledge base model of Civics teaching based on the source domain data and migrated it to the target domain. To compensate for the bias in the migration, the study designed a three-stage iterative optimization migration weight learning method, which significantly improved the quality of Civics teaching by updating the parameter weights of the source domain model. The process was verified through the experimental control group pre and post-tests. The results showed that the total score of critical thinking ability pre and post-tests was −3.39. The students’ innovative consciousness score in the experimental group was 3.96, which was significantly higher than that of the control group, which was 0.18. In addition, the scores of the quality dimensions of the students in the control group and the experimental group ranged from 3.22 to 3.40 and 3.43 to 3.68, respectively. This empirical result effectively confirms that the Civics teaching model proposed in this paper can effectively improve students’ critical thinking ability, innovation consciousness and overall quality, which is of great guiding significance to the current practice of Civics teaching in colleges and universities.
- Published
- 2024
- Full Text
- View/download PDF
45. Research on the Quality Assessment Model of 'Dual-Teacher' Teachers in Higher Vocational Colleges and Universities Based on Big Data Technology
- Author
-
Lou Yunjia and Xiao Huayi
- Subjects
support vector machine ,dbt-svm evaluation ,classifier combination ,kernel function ,“dual-teacher” teachers ,62p30 ,Mathematics ,QA1-939 - Abstract
To further analyze the content of “dual-teacher” teacher team construction in higher vocational colleges and universities, this paper focuses on teacher moral construction, teacher training, assessment and evaluation, and incentives for in-depth investigation. It is clear that higher vocational colleges and universities should distinguish between “dual-teacher” teacher quality assessment and ordinary teacher assessment, and put forward the gradient “dual-teacher” teacher quality team construction. Improve the support vector machine, use the binary tree to propose an evaluation model based on incomplete DBT-SVM, and combine multiple binary classifiers to solve the multi-classification problem of “dual-teacher” teacher quality evaluation. The optimal parameter combinations, i.e., C =28 and γ = 0.0534, are obtained using the kernel function and parameter tuning experiments, and the accuracy of the model prediction results reaches 94.325%, taking the “dual-teacher” teachers in a university in Y province as the specific evaluation object. This shows that the accuracy of this DBT-SVM-based evaluation method of “dual-teacher” teacher quality is good.
- Published
- 2024
- Full Text
- View/download PDF
46. Reflection on Engineering Myth from the Viewpoint of the Technological Culture
- Author
-
Xu Xin
- Subjects
technological culture ,engineering myth ,engineering ethic ,moral ideal ,62p30 ,Mathematics ,QA1-939 - Abstract
Myth influences the thinking and imagination of engineering activities with its rich narrative and symbolic meaning, and engineering transforms technological evolution into reality, providing new possibilities for the inheritance and reshaping of myth. From the perspective of technological culture, we analyze the engineering myths that are intertwined and fused with mythological narratives and engineering practices, breaking the “sacred” limitations in the process, exploring new engineering fields and creating a sustainable future with a broader, more open, and more flexible vision. To examine the social influence of engineering myths, a knowledge map of traditional myths and technological culture was created. Taking the Chang’e moon exploration project as an example, academic qualifications have significant correlations with political influence, military influence, psychosocial influence, and educational influence. After analyzing the technological culture exhibition of the engineering myth, the technological culture of engineering myth purports to strengthen the interaction between culture and engineering, encourage engineering practice to transcend itself and guide the realization of technological ethics in engineering practice.
- Published
- 2024
- Full Text
- View/download PDF
47. Deep Learning Based Image Recognition Technology for Civil Engineering Applications
- Author
-
Yang Delan
- Subjects
deep learning ,image recognition technology ,civil engineering ,faster r-cnn ,resnet-101+fpn+roi align ,62p30 ,Mathematics ,QA1-939 - Abstract
In this paper, we use Caffe framework to implement the improved Faster R-CNN recognition technique for building images in civil engineering under Linux system and add feature pyramid network and regional feature aggregation into the ResNet-50 network and ResNet-101 network, respectively, to strengthen the training effect, and establish ResNet-101+FPN+ROI Align image recognition technique. Simulated crack experiments and concrete surface quality defect detection experiments confirm that the ResNet-101 FPN ROI Align method is accurate and detects defects at a high rate. The method established in this paper has a minimum error of only 0.4% in the simulated crack experiment, and the detection rate is much higher than that of other detection methods when detecting quality defects on the concrete surface, and the accuracy can reach up to 94% at the same time. In civil engineering, the image recognition technology established in this paper has practical significance and high application value, as demonstrated by the experiment.
- Published
- 2024
- Full Text
- View/download PDF
48. Exploring the reform of live e-commerce talent training based on the four-way collaboration of 'government, industry, enterprise and school'
- Author
-
Wang Xin
- Subjects
evolutionary game models ,replicated dynamic equations ,stable equilibrium points ,sensitivity level ,live e-commerce ,62p30 ,Mathematics ,QA1-939 - Abstract
This paper takes universities and enterprises as the basic game subjects of collaborative innovation of live e-commerce talent cultivation and defines the involved interest subjects as government, school, enterprise and industry to construct a multi-party evolutionary game model. The model’s evolutionary stable equilibrium point can be calculated using the replicated dynamic equations of multiple parties. MATLAB was used to analyze the numerical simulation to explore the sensitivity of the relevant parameters for the three parties. Finally, according to the game results, a four-point practical path to optimize the live e-commerce talent training program is proposed. The results show that under the satisfaction of the equilibrium point (0, 0, 0), the critical value of the active participation willingness of the four parties is between 0.4 and 0.5, and the final equilibrium point tends to (1, 1, 1). In this paper, the cooperation between schools and enterprises is prompted to change from limited rationality to full rationality in order to realize the Pareto optimal equilibrium of the final game.
- Published
- 2024
- Full Text
- View/download PDF
49. Research on brand operation strategy of integrating traditional culture into the modern tourism industry based on big data analysis
- Author
-
Zhao Xiang, Wei Jinchan, and Chen Jiaping
- Subjects
big data ,lfv-lda model ,topic representation ,tourism branding ,theme words ,62p30 ,Mathematics ,QA1-939 - Abstract
The rich traditional culture is an inexhaustible treasure trove of resources for the brand operation of the modern tourism industry. This paper proposes an improved method for the LDA model of text big data mining, optimizing the topic representation of LDA by introducing the Topic2Vec model, optimizing the probability calculation of generated words by combining the original Dilley distribution and Bernoulli distribution and optimizing the assignment of topic words in the Gibbs adoption process of LDA model with the help of likelihood function. Then, the big data of tourism brand design text, promotion, and cultural products are mined and analyzed based on the improved LFV-LDA model. In the design of tourism brands focusing on traditional culture, the frequency of the keyword “pattern” was the highest, accounting for 8.20%. In brand promotion, the most distributed channel from 2017 to 2021 was a live broadcast, and the growth rate in 2020 and 2021 reached 22.71% and 36.29%. Regarding traditional cultural and creative products, tourism cultural and creative products have maintained a high growth momentum from 2012 to 2021, with an average annual growth rate of 154.25%. Based on the analysis of big data, it can guide the operation of tourism industry brands in all aspects, such as design, promotion, and landing, and give full play to the value connotation of traditional culture.
- Published
- 2024
- Full Text
- View/download PDF
50. Corporate Utilization of Bank Credit Financing Advantages under Supply Chain Finance and Industrial Policy Paths
- Author
-
Zhao Xiaolu
- Subjects
strategy selection ,nash equilibrium point ,regression model ,game evolution model ,bank credit financing ,62p30 ,Mathematics ,QA1-939 - Abstract
Bank credit financing is an effective method to promote the development of enterprises, and this paper discusses the use of enterprises in the credit financing advantage supported by supply chain financial networks and industrial policy in the new era. In the paper, the strategic choice of enterprises using bank credit financing is analyzed, the Nash equilibrium point of the bank and the enterprise is modeled by calculus, and the game evolution model of enterprise use of credit financing advantage is constructed to explore the equilibrium distribution of bank credit financing. On this basis, the regression model of bank credit financing and enterprise reputation is established, and it is combined with relevant data for empirical analysis. The results show that under the initial conditions, there are two evolutionary stable strategies of (0,0,0) and (1,1,1) in the enterprise bank credit financing game system, which verifies the analysis for the stability of equilibrium point under the dual mechanism of supply chain finance and industrial policy. The regression coefficient of bank credit financing of enterprises with greater corporate reputation is 1.26, and when the explanatory variable is the indicator of short-term lending and long-term investment, the coefficient of long-term borrowing ratio is -0.003, with a significance level of 1%, and the increase of long-term borrowing ratio can significantly reduce the behavior of short-term lending and long-term investment of enterprises.
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.