86 results on '"Xueying, Zhang"'
Search Results
2. Detecting Disorders of Consciousness in Brain Injuries From EEG Connectivity Through Machine Learning
- Author
-
Fengyun Hu, Fang Wang, Yu-Chu Tian, and Xueying Zhang
- Subjects
Majority rule ,Control and Optimization ,medicine.diagnostic_test ,business.industry ,Computer science ,Disorders of consciousness ,Electroencephalography ,Machine learning ,computer.software_genre ,medicine.disease ,Computer Science Applications ,Support vector machine ,Computational Mathematics ,Artificial Intelligence ,Classifier (linguistics) ,medicine ,Wakefulness ,Artificial intelligence ,business ,computer - Abstract
Disorders of consciousness (DoC) happen frequently in various brain injuries. Their detection helps timely treatment for better survival outcomes of DoC patients. It is conventionally undertaken via clinical examinations, typically behavioural assessments. However, these neurological examinations consume significant resources of manpower and time, making continuous DoC monitoring practically infeasible. To address this issue, a computer-aided approach is proposed in this article for automated DoC detection through extracting knowledge from electroencephalogram (EEG) signals. It introduces a new connectivity measure: Power Spectral Density Difference (PSDD) incorporating with a recursive Cosine function (CPSDD). Then, the approach classifies brain-injured patients into DoC (i.e., positive) and wakefulness (i.e., negative) classes through an ensemble of support vector machines (EOSVM), which is a type of machine-learning methods. It is further applied to a dataset of 607 patients with brain injuries. Our classification results show that the EOSVM classifier with the new connectivity measure CPSDD has achieved the best classification performance among 12 connectivity measures. For a setting of 97% majority voting from all SVMs, the EOSVM has diagnosed, in high confidence, 35% of patients with the accuracy, sensitivity, and specificity of 98.21%, 100%, and 95.79%, respectively. Thus, the classifier EOSVM incorporating with the new connectivity measure CPSDD is a promising tool for automatic detection of DoC in brain injuries.
- Published
- 2022
- Full Text
- View/download PDF
3. Robust whole slide image analysis for cervical cancer screening using deep learning
- Author
-
Gong Rao, Wenjie Zhu, Shaoqun Zeng, Ziquan Wei, Jiabo Ma, Ning Li, Wei Han, Xiuli Liu, Li Chen, Xu Li, Jing Cai, Xi Feng, Shenghua Cheng, Tingwei Quan, Sibo Liu, Xiebo Geng, Zehua Wang, Xiao Yuwei, Xueying Zhang, Xiaohua Lv, Jingya Yu, Fei Yang, and Junbo Hu
- Subjects
Computer science ,Cytodiagnosis ,Science ,General Physics and Astronomy ,Uterine Cervical Neoplasms ,Image processing ,Cervical cancer screening ,General Biochemistry, Genetics and Molecular Biology ,Article ,Cancer screening ,Deep Learning ,Machine learning ,Image Processing, Computer-Assisted ,Humans ,Diagnosis, Computer-Assisted ,Recognition algorithm ,Early Detection of Cancer ,Multidisciplinary ,Artificial neural network ,business.industry ,Deep learning ,Reproducibility of Results ,Pattern recognition ,General Chemistry ,Translational research ,Recurrent neural network ,ROC Curve ,Whole slide image ,Female ,Artificial intelligence ,Neural Networks, Computer ,business - Abstract
Computer-assisted diagnosis is key for scaling up cervical cancer screening. However, current recognition algorithms perform poorly on whole slide image (WSI) analysis, fail to generalize for diverse staining and imaging, and show sub-optimal clinical-level verification. Here, we develop a progressive lesion cell recognition method combining low- and high-resolution WSIs to recommend lesion cells and a recurrent neural network-based WSI classification model to evaluate the lesion degree of WSIs. We train and validate our WSI analysis system on 3,545 patient-wise WSIs with 79,911 annotations from multiple hospitals and several imaging instruments. On multi-center independent test sets of 1,170 patient-wise WSIs, we achieve 93.5% Specificity and 95.1% Sensitivity for classifying slides, comparing favourably to the average performance of three independent cytopathologists, and obtain 88.5% true positive rate for highlighting the top 10 lesion cells on 447 positive slides. After deployment, our system recognizes a one giga-pixel WSI in about 1.5 min., Computer-assisted diagnosis is key for scaling up cervical cancer screening, but current algorithms perform poorly on whole slide image analysis and generalization. Here, the authors present a WSI classification and top lesion cell recommendation system using deep learning, and achieve comparable results with cytologists.
- Published
- 2021
4. A novel brain-computer interface based on audio-assisted visual evoked EEG and spatial-temporal attention CNN
- Author
-
Guijun, Chen, Xueying, Zhang, Jing, Zhang, Fenglian, Li, and Shufei, Duan
- Subjects
Artificial Intelligence ,Biomedical Engineering - Abstract
ObjectiveBrain-computer interface (BCI) can translate intentions directly into instructions and greatly improve the interaction experience for disabled people or some specific interactive applications. To improve the efficiency of BCI, the objective of this study is to explore the feasibility of an audio-assisted visual BCI speller and a deep learning-based single-trial event related potentials (ERP) decoding strategy.ApproachIn this study, a two-stage BCI speller combining the motion-onset visual evoked potential (mVEP) and semantically congruent audio evoked ERP was designed to output the target characters. In the first stage, the different group of characters were presented in the different locations of visual field simultaneously and the stimuli were coded to the mVEP based on a new space division multiple access scheme. And then, the target character can be output based on the audio-assisted mVEP in the second stage. Meanwhile, a spatial-temporal attention-based convolutional neural network (STA-CNN) was proposed to recognize the single-trial ERP components. The CNN can learn 2-dimentional features including the spatial information of different activated channels and time dependence among ERP components. In addition, the STA mechanism can enhance the discriminative event-related features by adaptively learning probability weights.Main resultsThe performance of the proposed two-stage audio-assisted visual BCI paradigm and STA-CNN model was evaluated using the Electroencephalogram (EEG) recorded from 10 subjects. The average classification accuracy of proposed STA-CNN can reach 59.6 and 77.7% for the first and second stages, which were always significantly higher than those of the comparison methods (p < 0.05).SignificanceThe proposed two-stage audio-assisted visual paradigm showed a great potential to be used to BCI speller. Moreover, through the analysis of the attention weights from time sequence and spatial topographies, it was proved that STA-CNN could effectively extract interpretable spatiotemporal EEG features.
- Published
- 2022
- Full Text
- View/download PDF
5. Case-based reasoning adaptation based on fuzzy gravitational search algorithm for disaster emergency plan
- Author
-
Xueying Zhang, Xiaobing Yu, and Xianrui Yu
- Subjects
Statistics and Probability ,0209 industrial biotechnology ,business.industry ,Computer science ,Gravitational search algorithm ,General Engineering ,Emergency plan ,02 engineering and technology ,Fuzzy logic ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Case-based reasoning ,Artificial intelligence ,Adaptation (computer science) ,business - Abstract
Disasters can result in substantial destructive damages to the world. Emergency plan is vital to deal with these disasters. It is still difficult for the traditional CBR to generate emergency plans to meet requirements of rapid responses. An integrated system including Case-based reasoning (CBR) and gravitational search algorithm (GSA) is proposed to generate the disaster emergency plan. Fuzzy GSA (FGSA) is developed to enhance the convergence ability and accomplish the case adaptation in CBR. The proposed algorithm dynamically updates the main parameters of GSA by introducing a fuzzy system. The FGSA-CBR system is proposed, in which fitness function is defined based on the effectiveness of disaster emergency management. The comparison results have revealed that the proposed algorithm has good performances compared with the original GSA and other algorithms. A gas leakage accident is taken as an empirical study. The results have demonstrated that the FGSA-CBR has good performances when generating the disaster emergency plan. The combination of CBR and FGSA can realize the case adaptation, which provides a useful approach to the real applications.
- Published
- 2021
- Full Text
- View/download PDF
6. Hyperspectral Remote Sensing Image Classification Using Deep Convolutional Capsule Network
- Author
-
Chunju Zhang, Yucheng Yang, Wencong Liu, Jianwei Huang, Zhenxuan Li, Runmin Lei, Zhiyi Zhou, Lei Zhang, and Xueying Zhang
- Subjects
Atmospheric Science ,convolutional neural network (CNN) ,Contextual image classification ,Artificial neural network ,Computer science ,business.industry ,QC801-809 ,Deep learning ,hyperspectral image classification ,Feature extraction ,Geophysics. Cosmic physics ,Hyperspectral imaging ,Convolutional neural network ,Convolution ,Ocean engineering ,Robustness (computer science) ,Artificial intelligence ,Computers in Earth Sciences ,Capsule neural network ,business ,TC1501-1800 ,Remote sensing - Abstract
Deep learning models have shown excellent performance in the hyperspectral remote sensing image (HSI) classification. In particular, convolutional neural networks (CNNs) have received widespread attention because of their powerful feature-extraction ability. Recently, a capsule network (CapsNet) was introduced to boost the performance of CNNs, marking a remarkable progress in the field of HSI classification. In this article, we propose a novel deep convolutional capsule neural network (DC-CapsNet) based on spectral–spatial features to improve the performance of CapsNet in the HSI classification while significantly reducing the computation cost of the model. Specifically, a convolutional capsule layer based on the extension of dynamic routing using 3-D convolution is used to reduce the number of parameters and enhance the robustness of the learned spectral–spatial features. Furthermore, a lighter and stronger decoder network composed of deconvolutional layers as a better regularization term and capable of acquiring more spatial relationships is used to further improve the HSI classification accuracy with low computation cost. In this study, we tested the performance of the proposed model on four widely used HSI datasets: the Kennedy Space Center, Indian Pines, Pavia University, and Salinas datasets. We found that the DC-CapsNet achieved high classification accuracy with limited training samples and effectively reduced the computation cost.
- Published
- 2021
7. A Partitioning-Stacking Prediction Fusion Network Based on an Improved Attention U-Net for Stroke Lesion Segmentation
- Author
-
Haisheng Hui, Xueying Zhang, Fenglian Li, Xiaobi Mei, and Yuling Guo
- Subjects
General Computer Science ,Computer science ,computer.software_genre ,prediction fusion ,030218 nuclear medicine & medical imaging ,Lesion ,03 medical and health sciences ,0302 clinical medicine ,Voxel ,medicine ,General Materials Science ,Segmentation ,Acute ischemic stroke ,medicine.diagnostic_test ,business.industry ,General Engineering ,deep learning ,Magnetic resonance imaging ,Pattern recognition ,Image segmentation ,ATLAS ,stroke ,Partition (database) ,attention U-net ,Identification (information) ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,Artificial intelligence ,medicine.symptom ,business ,lcsh:TK1-9971 ,computer ,030217 neurology & neurosurgery ,MRI - Abstract
Due to the narrow time window for the treatment of acute ischemic stroke, the stroke lesion area in the patient must be identified as quickly and accurately as possible to evaluate the risks and get the most timely and effective treatment. Therefore, there is great clinical significance in the study of automatic identification and segmentation methods for stroke lesions. In this paper, we propose a partitioning-stacking prediction fusion (PSPF) method based on an improved attention U-net to solve the problems of 3D-CNN-based networks, including their high computational cost and insufficient training data, and to achieve accurate segmentation of 3D stroke lesions. Our proposed PSPF method includes three steps in the first part. In Step 1, partitioning, we partition the slices obtained in a certain plane direction by slicing a Magnetic Resonance Imaging (MRI) into subsets according to the 2D graph similarity, then use each partitioned subset to perform training and prediction separately. In Step 2, stacking, we stack the 2D slice results of all subsets according to the position order in MRI before slicing and partitioning to form a 3D lesion result. In Step 3, fusion, we use soft voting to fuse the three orthogonal planes' 3D results that were obtained voxel by voxel in Steps 1 and 2. In the second part, we propose an improved attention U-net, which uses the features from three different scales to generate the attention gating coefficients that further improve training efficiency and segmentation accuracy. We implement a 6-fold cross-validation on the Anatomical Tracings of Lesions After Stroke (ATLAS) dataset to validate our method and model using metrics such as Dice Coefficient (DC), F2 score, and Area under Precision-Recall curve (APR). The results show that compared to the existing methods, our proposed method can not only improve the segmentation precision on unbalanced data but also improve the detailed performance of lesion segmentation. Our proposed method and model are generalized and accurate, demonstrating the good potential for clinical routines. The source codes and models in our method have been made publicly available at [available.upon.acceptance].
- Published
- 2020
- Full Text
- View/download PDF
8. Emotion Feature Analysis and Recognition Based on Reconstructed EEG Sources
- Author
-
Xueying Zhang, Jing Zhang, Guijun Chen, and Ying Sun
- Subjects
General Computer Science ,Computer science ,02 engineering and technology ,Electroencephalography ,EEG source reconstruction ,Field (computer science) ,03 medical and health sciences ,0302 clinical medicine ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,General Materials Science ,inverse solution ,Emotion recognition ,Affective computing ,medicine.diagnostic_test ,business.industry ,time- and frequency-domain features ,General Engineering ,Pattern recognition ,Support vector machine ,020201 artificial intelligence & image processing ,Artificial intelligence ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,difference analysis of active source ,business ,lcsh:TK1-9971 ,030217 neurology & neurosurgery - Abstract
Emotion plays a significant role in perceiving external events or situations in daily life. Due to ease of use and relative accuracy, Electroencephalography (EEG)-based emotion recognition has become a hot topic in the affective computing field. However, scalp EEG is a mixed-signal and cannot directly indicate the exact information about active cortex sources of different emotions. In this paper, we analyze the significant differences of active source regions and frequency bands for pairs of emotions-based reconstructed EEG sources using sLORETA, and 26 Brodmann areas are selected as the regions of interest (ROI). And then, six kinds of time- and frequency-domain features from significant active regions and frequency bands are extracted to classify different emotions using support vector machines. Furthermore, we compare the classification performances of emotion features extracted from active source regions and EEG sensors. We have demonstrated that the features from selected source regions can improve the classification accuracy by extensive experiments on the DEAP and TYUT 2.0 EEG-based datasets.
- Published
- 2020
9. Deep Feature Aggregation Network for Hyperspectral Remote Sensing Image Classification
- Author
-
Hui Zheng, Zhaofu Wu, Chunju Zhang, Runmin Lei, Xueying Zhang, Guandong Li, and Shihong Du
- Subjects
Atmospheric Science ,Dense connectivity ,Computer science ,hyperspectral image classification ,Feature extraction ,Geophysics. Cosmic physics ,0211 other engineering and technologies ,02 engineering and technology ,Overfitting ,Convolutional neural network ,Convergence (routing) ,0202 electrical engineering, electronic engineering, information engineering ,feature fusion ,Computers in Earth Sciences ,TC1501-1800 ,021101 geological & geomatics engineering ,Remote sensing ,Contextual image classification ,business.industry ,QC801-809 ,Deep learning ,Hyperspectral imaging ,Ocean engineering ,Feature (computer vision) ,3-D convolutional neural network (3D-CNN) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,residual learning - Abstract
Hyperspectral remote sensing images (HSIs) are rich in spectral–spatial information. The deep learning models can help to automatically extract and discover this rich information from HSIs for classifying HSIs. However, the sampling of the models and the design of the hyperparameters depend on the number of samples and the size of each sample's input space. In the case of limited samples, the description dimension of features is also limited and overfitting to other remote sensing image datasets is evident. This study proposes a novel deep feature aggregation network for HSI classification based on a 3-D convolutional neural network from the perspective of feature aggregation patterns. By introducing the residual learning and dense connectivity strategies, we established a deep feature residual network (DFRN) and a deep feature dense network (DFDN) to exploit the low-, middle-, and high-level features in HSIs. For the Indian Pines and Kennedy Space Center datasets, the DFRN model was determined to be more accurate. On the Pavia University dataset, both the DFDN and DFRN have basically the same accuracy, but the DFDN has faster convergence speed and more stable performance on the validation set than the DFRN. Therefore, when faced with different HSI data, the corresponding aggregation method can be chosen more flexibly according to the requirements on number of training samples and the convergence speed. This is beneficial in the HSI classification.
- Published
- 2020
10. Object-Oriented Method Combined with Deep Convolutional Neural Networks for Land-Use-Type Classification of Remote Sensing Images
- Author
-
Weiwei Song, Peng Ye, Shihua Li, Baoxuan Jin, and Xueying Zhang
- Subjects
Computer science ,business.industry ,Deep learning ,Geography, Planning and Development ,Feature extraction ,0211 other engineering and technologies ,02 engineering and technology ,Image segmentation ,Overfitting ,Convolutional neural network ,Feature (computer vision) ,Earth and Planetary Sciences (miscellaneous) ,Segmentation ,Artificial intelligence ,business ,021101 geological & geomatics engineering ,Curse of dimensionality ,Remote sensing - Abstract
Land-use information provides a direct representation of the effect of human activities on the environment, and an accurate and efficient land-use classification of remote sensing images is an important element of land-use and land-cover change research. To solve the problems associated with traditional land-use classification methods (e.g., rapid increase in dimensionality of data, inadequate feature extraction, and low running efficiency), a method that combines object-oriented approach with deep convolutional neural network (COCNN) is presented. First, a multi-scale segmentation algorithm is used to segment images to generate image segmentation regions with high homogeneity. Second, a typical rule set of feature objects is constructed on the basis of the object-oriented segmentation results, and the segmentation objects are classified and extracted to form a training sample set. Third, a convolutional neural network (CNN) model structure is modified to improve classification performance, and the training algorithm is optimized to avoid the overfitting phenomenon that occurs during training using small datasets. Ten land-use types are classified by using the remote sensing images covering the area around Fuxian Lake as an example. By comparing the COCNN method with the method based solely on CNN, precision and kappa index were selected to evaluate the classification accuracy of the two methods. For the COCNN method, on the basis of the classification statistics, precision and kappa index coefficients are 96.2% and 0.96, respectively, which are 8.98% and 0.1 higher than those of the method based solely on CNN. Experimental results show that the COCNN method reasonably and efficiently combines object-oriented and deep learning approaches, thereby effectively solving the problem of the inaccurate classification of typical features with better classification accuracy than the simple use of CNN.
- Published
- 2019
- Full Text
- View/download PDF
11. An ensemble of Xgboost models for detecting disorders of consciousness in brain injuries through EEG connectivity
- Author
-
Fang Wang, Yu-Chu Tian, Xueying Zhang, and Fengyun Hu
- Subjects
Artificial Intelligence ,General Engineering ,Computer Science Applications - Published
- 2022
- Full Text
- View/download PDF
12. Texas Public Agencies’ Tweets and Public Engagement During the COVID-19 Pandemic: Natural Language Processing Approach
- Author
-
Wenlin Liu, Hong Thoai Nga Tran, Degui Zhi, Wenxue Zou, Lu Tang, Benjamin Thomas, and Xueying Zhang
- Subjects
medicine.medical_specialty ,content analysis ,020205 medical informatics ,social media ,Twitter ,050801 communication & media studies ,Health Informatics ,02 engineering and technology ,computer.software_genre ,dissemination ,public engagement ,0508 media and communications ,Political science ,Pandemic ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,Health belief model ,Humans ,public health agencies ,health belief model ,Social media ,Public engagement ,natural language processing ,Pandemics ,Original Paper ,business.industry ,communication ,Public health ,05 social sciences ,public health ,Public Health, Environmental and Occupational Health ,COVID-19 ,Texas ,Action (philosophy) ,Content analysis ,Artificial intelligence ,Rural area ,Public aspects of medicine ,RA1-1270 ,business ,strategy ,computer ,Natural language processing ,belief ,engagement - Abstract
Background The ongoing COVID-19 pandemic is characterized by different morbidity and mortality rates across different states, cities, rural areas, and diverse neighborhoods. The absence of a national strategy for battling the pandemic also leaves state and local governments responsible for creating their own response strategies and policies. Objective This study examines the content of COVID-19–related tweets posted by public health agencies in Texas and how content characteristics can predict the level of public engagement. Methods All COVID-19–related tweets (N=7269) posted by Texas public agencies during the first 6 months of 2020 were classified in terms of each tweet’s functions (whether the tweet provides information, promotes action, or builds community), the preventative measures mentioned, and the health beliefs discussed, by using natural language processing. Hierarchical linear regressions were conducted to explore how tweet content predicted public engagement. Results The information function was the most prominent function, followed by the action or community functions. Beliefs regarding susceptibility, severity, and benefits were the most frequently covered health beliefs. Tweets that served the information or action functions were more likely to be retweeted, while tweets that served the action and community functions were more likely to be liked. Tweets that provided susceptibility information resulted in the most public engagement in terms of the number of retweets and likes. Conclusions Public health agencies should continue to use Twitter to disseminate information, promote action, and build communities. They need to improve their strategies for designing social media messages about the benefits of disease prevention behaviors and audiences’ self-efficacy.
- Published
- 2021
13. Dual-Path Attention Compensation U-Net for Stroke Lesion Segmentation
- Author
-
Wu Zelin, Haisheng Hui, Xueying Zhang, and Fenlian Li
- Subjects
Article Subject ,General Computer Science ,Computer science ,General Mathematics ,Computer applications to medicine. Medical informatics ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,R858-859.7 ,Core network ,Binary number ,Neurosciences. Biological psychiatry. Neuropsychiatry ,Image (mathematics) ,Compensation (engineering) ,Image Processing, Computer-Assisted ,Humans ,Segmentation ,business.industry ,General Neuroscience ,Pattern recognition ,General Medicine ,Function (mathematics) ,Magnetic Resonance Imaging ,Stroke ,Task (computing) ,Path (graph theory) ,Artificial intelligence ,business ,Research Article ,RC321-571 - Abstract
For the segmentation task of stroke lesions, using the attention U-Net model based on the self-attention mechanism can suppress irrelevant regions in an input image while highlighting salient features useful for specific tasks. However, when the lesion is small and the lesion contour is blurred, attention U-Net may generate wrong attention coefficient maps, leading to incorrect segmentation results. To cope with this issue, we propose a dual-path attention compensation U-Net (DPAC-UNet) network, which consists of a primary network and auxiliary path network. Both networks are attention U-Net models and identical in structure. The primary path network is the core network that performs accurate lesion segmentation and outputting of the final segmentation result. The auxiliary path network generates auxiliary attention compensation coefficients and sends them to the primary path network to compensate for and correct possible attention coefficient errors. To realize the compensation mechanism of DPAC-UNet, we propose a weighted binary cross-entropy Tversky (WBCE-Tversky) loss to train the primary path network to achieve accurate segmentation and propose another compound loss function called tolerance loss to train the auxiliary path network to generate auxiliary compensation attention coefficient maps with expanded coverage area to perform compensate operations. We conducted segmentation experiments using the 239 MRI scans of the anatomical tracings of lesions after stroke (ATLAS) dataset to evaluate the performance and effectiveness of our method. The experimental results show that the DSC score of the proposed DPAC-UNet network is 6% higher than the single-path attention U-Net. It is also higher than the existing segmentation methods of the related literature. Therefore, our method demonstrates powerful abilities in the application of stroke lesion segmentation.
- Published
- 2021
14. DSGPT: Domain-Specific Generative Pre-Training of Transformers for Text Generation in E-commerce Title and Review Summarization
- Author
-
Xiaochuan Fan, Bo Long, Xueying Zhang, Chi Zhang, Yunjiang Jiang, Yue Shang, Zhaomeng Cheng, and Yun Xiao
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Machine translation ,Computer Science - Artificial Intelligence ,business.industry ,Computer science ,E-commerce ,computer.software_genre ,Automatic summarization ,Domain (software engineering) ,Task (project management) ,Resource (project management) ,Artificial Intelligence (cs.AI) ,Language model ,Artificial intelligence ,business ,computer ,Computation and Language (cs.CL) ,Natural language processing ,Transformer (machine learning model) - Abstract
We propose a novel domain-specific generative pre-training (DSGPT) method for text generation and apply it to the product title and review summarization problems on E-commerce mobile display. First, we adopt a decoder-only transformer architecture, which fits well for fine-tuning tasks by combining input and output all together. Second, we demonstrate utilizing only small amount of pre-training data in related domains is powerful. Pre-training a language model from a general corpus such as Wikipedia or the Common Crawl requires tremendous time and resource commitment, and can be wasteful if the downstream tasks are limited in variety. Our DSGPT is pre-trained on a limited dataset, the Chinese short text summarization dataset (LCSTS). Third, our model does not require product-related human-labeled data. For title summarization task, the state of art explicitly uses additional background knowledge in training and predicting stages. In contrast, our model implicitly captures this knowledge and achieves significant improvement over other methods, after fine-tuning on the public Taobao.com dataset. For review summarization task, we utilize JD.com in-house dataset, and observe similar improvement over standard machine translation methods which lack the flexibility of fine-tuning. Our proposed work can be simply extended to other domains for a wide range of text generation tasks.
- Published
- 2021
- Full Text
- View/download PDF
15. Dynamic Pricing and Placement for Distributed Machine Learning Jobs
- Author
-
John C. S. Lui, Xueying Zhang, Zongpeng Li, and Ruiting Zhou
- Subjects
business.industry ,Computer science ,Cloud computing ,Time horizon ,Regret ,02 engineering and technology ,010501 environmental sciences ,Machine learning ,computer.software_genre ,01 natural sciences ,Multi-armed bandit ,020202 computer hardware & architecture ,Server ,Dynamic pricing ,Fixed price ,0202 electrical engineering, electronic engineering, information engineering ,Profitability index ,Artificial intelligence ,business ,computer ,0105 earth and related environmental sciences - Abstract
Nowadays distributed machine learning (ML) jobs usually adopt a parameter server (PS) framework to train models over large-scale datasets. Such ML job deploys hundreds of concurrent workers, and model parameter updates are exchanged frequently between workers and PSs. Current practice is that workers and PSs may be placed on different physical servers, bringing uncertainty in jobs’ runtime. Also, existing cloud pricing policy often charges a fixed price according to the job’s runtime. Although this pricing strategy is simple to implement, such pricing mechanism is not suitable for distributed ML jobs whose runtime is stochastic and can only be estimated according to its placement after job admission. To supplement existing cloud pricing schemes, we design a dynamic pricing and placement algorithm, DPS, for distributed ML jobs. DPS aims to maximize cloud provider’s profit, which dynamically calculates unit resource price upon a job’s arrival, and determines job’s placement to minimize its runtime if offered price is accepted to users. Our design exploits the multi-armed bandit (MAB) technique to learn unknown information based on past sales. DPS balances the exploration and exploitation stage, and selects the best price based on the reward which is related to job runtime. Our learning-based algorithm increases the provider’s profit, and achieves a sub-linear regret with both the time horizon and the total job number, compared to benchmark pricing schemes. Extensive evaluations also validates the efficacy of DPS.
- Published
- 2020
- Full Text
- View/download PDF
16. Simulation research on high precision multimode GNSS positioning algorithm
- Author
-
K. Kiatsupaibul, Xueying Zhang, and Kexun Chen
- Subjects
Statistics and Probability ,Multi-mode optical fiber ,010504 meteorology & atmospheric sciences ,Artificial Intelligence ,GNSS applications ,Computer science ,0103 physical sciences ,General Engineering ,Electronic engineering ,010303 astronomy & astrophysics ,01 natural sciences ,0105 earth and related environmental sciences - Published
- 2018
- Full Text
- View/download PDF
17. Multi-Feature Fusion Method Based on EEG Signal and its Application in Stroke Classification
- Author
-
Wenhui Jia, Hui Haisheng, Fengyun Hu, Fan Yuzhou, Can Wang, Xueying Zhang, and Fenglian Li
- Subjects
020205 medical informatics ,Computer science ,Quantitative Biology::Tissues and Organs ,Decision tree ,Wavelet Analysis ,Medicine (miscellaneous) ,Health Informatics ,02 engineering and technology ,Electroencephalography ,Approximate entropy ,Pattern Recognition, Automated ,Wavelet ,Health Information Management ,Fuzzy Logic ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,Humans ,Entropy (energy dispersal) ,Quantitative Biology::Neurons and Cognition ,medicine.diagnostic_test ,business.industry ,Pattern recognition ,Signal Processing, Computer-Assisted ,Random forest ,Sample entropy ,Support vector machine ,Stroke ,ComputingMethodologies_PATTERNRECOGNITION ,Artificial intelligence ,business ,Algorithms ,Information Systems - Abstract
Electroencephalogram (EEG) analysis has been widely used in the diagnosis of stroke diseases for its low cost and noninvasive characteristics. In order to classify the EEG signals of stroke patients with cerebral infarction and cerebral hemorrhage, this paper proposes a novel EEG stroke signal classification method. This method has two highlights. The first is that a multi-feature fusion method is given by combining wavelet packet energy, fuzzy entropy and hierarchical theory. The second highlight is that a suitable classification model based on ensemble classifier is constructed for perfectly classification stroke signals. Entropy is an accessible way to measure information and uncertainty of time series. Many entropy-based methods have been developed these years. By comparing with the performances of permutation entropy, sample entropy, approximate entropy in measuring the characteristic of stroke patient's EEG signals, it can be found that fuzzy entropy has best performance in characterization stroke EEG signal. By combining hierarchical theory, wavelet packet energy and fuzzy entropy, a multi-feature fusion method is proposed. The method first calculates wavelet packet energy of EEG stroke signal, then extracts hierarchical fuzzy entropy feature by combining hierarchical theory and fuzzy entropy. The experimental results show that, compared with the fuzzy entropy feature, the classification accuracy based on the fusion feature of wavelet packet energy and hierarchical fuzzy entropy is much higher than benchmark methods. It means that the proposed multi-feature fusion method based on stroke EEG signal is an efficient measure in classifying ischemic and hemorrhagic stroke. Support vector machine (SVM), decision tree and random forest are further used as the stroke signal classification models for classifying ischemic stroke and hemorrhagic stroke. Experimental results show that, based on the proposed multi-feature fusion method, the ensemble method of random forest can get the best classification performance in accuracy among three models.
- Published
- 2019
18. Classification of Improved Cross-Correlation Function to Determine Speaker Location from Microphone Array
- Author
-
Lixia Huang, Wang Jie, Xueying Zhang, and Suisui Zhang
- Subjects
Normalization (statistics) ,Microphone array ,Cross-correlation ,business.industry ,Computer science ,Pattern recognition ,Acoustic source localization ,01 natural sciences ,Weighting ,030507 speech-language pathology & audiology ,03 medical and health sciences ,0103 physical sciences ,Artificial intelligence ,0305 other medical science ,business ,010301 acoustics ,Classifier (UML) - Abstract
The approaches based on machine learning technology were used to source localization in adverse acoustic environment in recent years. However, the features extracted from the GCC vectors are no significant differences in different location, resulting in an increase in the misclassification. In order to mitigate this problem, this paper introduced the algorithm based on classification of novel cross-correlation function. Firstly, the cross-correlation function is calculated by joint weighting of normalized Phase Transformation and Smooth Coherence Transform (PHAT-SCOT) in each location. Then, acoustic source location is estimated by the Fisher Weighted Naive-Bayes Classifier(FWNBC), which gives different weights according to the importance of features in classification to improve the accuracy. At the same time, the new method was verified in real environment. The experiments all showed that our new method outperforms the baseline algorithm not only in simulating situation but also in real localizing system.
- Published
- 2019
- Full Text
- View/download PDF
19. Features of Hierarchical Fuzzy Entropy of Stroke Based on EEG Signal and Its Application in Stroke Classification
- Author
-
Fengyun Hu, Fan Yuzhou, Wenhui Jia, Xueying Zhang, Can Wang, and Fenglian Li
- Subjects
Computer science ,Quantitative Biology::Tissues and Organs ,Feature extraction ,02 engineering and technology ,Electroencephalography ,Approximate entropy ,03 medical and health sciences ,symbols.namesake ,0302 clinical medicine ,Robustness (computer science) ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,Gaussian function ,Quantitative Biology::Neurons and Cognition ,medicine.diagnostic_test ,business.industry ,Pattern recognition ,Support vector machine ,Sample entropy ,ComputingMethodologies_PATTERNRECOGNITION ,symbols ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Classifier (UML) ,030217 neurology & neurosurgery - Abstract
Electroencephalogram (EEG) analysis has been widely used in the diagnosis of stroke diseases for its low cost and noninvasive characteristics. In order to classify the electroencephalogram (EEG) signals of stroke patients with cerebral infarction and cerebral hemorrhage, this paper proposed a novel EEG stroke signal feature extraction method by combining fuzzy entropy and hierarchical theory. Fuzzy entropy not only took the advantages of sample entropy, but also had less dependence on the length of time series and possessed better robustness to noise signals. It measured the similarity of two vectors based on Gaussian function instead of Heaviside function, avoiding discontinuity problems of sample entropy and approximate entropy. Hierarchical theory efficiently took advantages of the approximation information in low-frequency and the detail information in high-frequency. This was benefit for capturing a wealth of dynamic information and retaining redundant components. Support vector machine (SVM) was further used as the stroke signal classification model for classifying ischemic stroke and hemorrhagic stroke. The experimental results showed that, compared with other benchmarks, the classification accuracy based on the features of hierarchical fuzzy entropy is much higher than those benchmarks methods. Compared with the features of fuzzy entropy without using hierarchical theory, the classifier based on the features of hierarchical fuzzy entropy gave a much more improvement in classification performance by increasing accuracy from 68.03% to 96.72%. It meant that the proposed EEG stroke signal hierarchical fuzzy entropy feature extraction method was an efficient measure in classifying ischemic and hemorrhagic stroke.
- Published
- 2019
- Full Text
- View/download PDF
20. Fuzzy Support Vector Machine with Imbalanced Regulator and its Application in Stroke Classification
- Author
-
Chao Wang, Fenglian Li, Fengyun Hu, Wenhui Jia, Wei Xin, and Xueying Zhang
- Subjects
Computer science ,Regulator ,02 engineering and technology ,Disease ,Machine learning ,computer.software_genre ,030218 nuclear medicine & medical imaging ,03 medical and health sciences ,0302 clinical medicine ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,Entropy (information theory) ,Stroke ,Fuzzy support vector machine ,business.industry ,High mortality ,Blood flow ,medicine.disease ,Class (biology) ,Transcranial Doppler ,Data set ,cardiovascular system ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
As a typical cardiovascular disease, stroke is the world's second most deadly disease with the characteristics of high mortality, incidence, recurrence rate and disability rate. To improve the stroke predicting accuracy is a necessary measure for reducing the happening of stroke. As a non-invasive and portable method? Transcranial Doppler (TCD) can provide valuable information in diagnosis of cerebrovascular disease by measuring the blood flow speed in intracranial arteries. How to improve diagnosis accuracy performance based on TCD data set is a necessary demand for realizing cerebrovascular disease TCD intelligent diagnosis system. But TCD data set often presents an imbalanced phenomenon for less patients' data over all examiners. This imbalanced phenomenon easily makes a lower predicting accuracy of cerebrovascular patients. In this paper, an improved fuzzy support vector machine model is proposed by combining class center distance and information entropy. The model efficiently utilizes the advantages of class center distance and information entropy. Moreover, the imbalanced fuzzy support vector machine is used by combining an adaptive imbalanced regulator. The different influence of the imbalance regulating factor value on the classification effect were discussed in detail for finding a suitable regulating factor value. A practical clinical imbalanced TCD dataset is used by collecting clinical TCD data for identify the performance of the proposed method. The experimental results showed that the proposed method can get a better classification performance with higher accuracies and G-means over imbalanced TCD dataset than other benchmark methods.
- Published
- 2019
- Full Text
- View/download PDF
21. Combining supervised classifiers with unlabeled data
- Author
-
Lixia Huang, Fenglian Li, Xueying Zhang, and Xue-yan Liu
- Subjects
Computer science ,business.industry ,Metals and Alloys ,General Engineering ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,Semi-supervised learning ,Machine learning ,computer.software_genre ,Ensemble learning ,Regularization (mathematics) ,ComputingMethodologies_PATTERNRECOGNITION ,Metallic materials ,0202 electrical engineering, electronic engineering, information engineering ,Graph (abstract data type) ,Labeled data ,020201 artificial intelligence & image processing ,Artificial intelligence ,Benchmark data ,business ,computer ,Label propagation - Abstract
Ensemble learning is a wildly concerned issue. Traditional ensemble techniques are always adopted to seek better results with labeled data and base classifiers. They fail to address the ensemble task where only unlabeled data are available. A label propagation based ensemble (LPBE) approach is proposed to further combine base classification results with unlabeled data. First, a graph is constructed by taking unlabeled data as vertexes, and the weights in the graph are calculated by correntropy function. Average prediction results are gained from base classifiers, and then propagated under a regularization framework and adaptively enhanced over the graph. The proposed approach is further enriched when small labeled data are available. The proposed algorithms are evaluated on several UCI benchmark data sets. Results of simulations show that the proposed algorithms achieve satisfactory performance compared with existing ensemble methods.
- Published
- 2016
- Full Text
- View/download PDF
22. Speaker Localization with Smoothing Generalized Cross Correlation Based on Naive Bayes Classifier
- Author
-
Suisui Zhang, Fenglian Li, Zan Danfei, Xueying Zhang, and Lixia Huang
- Subjects
Microphone ,business.industry ,Computer science ,Feature vector ,Word error rate ,Pattern recognition ,02 engineering and technology ,Acoustic source localization ,Bayes classifier ,030507 speech-language pathology & audiology ,03 medical and health sciences ,Naive Bayes classifier ,Statistical classification ,Computer Science::Sound ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,0305 other medical science ,business ,Smoothing - Abstract
Conventional approaches to acoustic source localization simply based on the received microphone signals, are often vulnerable to adverse acoustic conditions, such as low signal-to-noise ratio (SNR) or high reverberation. But, approaches based on Pattern Recognition and Machine Learning Technology can increase accuracy to locate source in adverse acoustic environment. The advantage of the algorithm is that it requires no calibration of microphone arrays. And Naive Bayes Classifier is simple, fast, and has a small error rate. This paper proposed an improved localization algorithm based on classification of cross-correlation functions (GCC). The weighted cross power spectrum of GCC is smoothed by a smooth filter to formed smooth generalized cross-correlation (SGCC). Then, the classifier model is obtained in each location and form the feature vector. Finally, acoustic source location is estimated by Naive- Bayes classifier. We also proposed in this study the source localization system that based on merely two microphones to input sound signals, combined with improved and optimal methods proposed above. Real-data experiments have demonstrated that algorithm with SGCC has higher localization accuracy than with GCC by 20% in the proposed system at least. The system has good ability to acoustic source localization.
- Published
- 2018
- Full Text
- View/download PDF
23. Deep Belief Networks Based Toponym Recognition for Chinese Text
- Author
-
Shu Wang, Xueying Zhang, Peng Ye, and Du Mi
- Subjects
Conditional random field ,Computer science ,Process (engineering) ,Geography, Planning and Development ,Big data ,0211 other engineering and technologies ,Deep Belief Networks ,lcsh:G1-922 ,toponym recognition ,02 engineering and technology ,computer.software_genre ,geographic information retrieval ,Deep belief network ,0202 electrical engineering, electronic engineering, information engineering ,Earth and Planetary Sciences (miscellaneous) ,Information system ,Word2vec ,Computers in Earth Sciences ,021101 geological & geomatics engineering ,business.industry ,place names ,Geographic information retrieval ,Complementarity (molecular biology) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer ,Natural language processing ,lcsh:Geography (General) ,Chinese text - Abstract
In Geographical Information Systems, geo-coding is used for the task of mapping from implicitly geo-referenced data to explicitly geo-referenced coordinates. At present, an enormous amount of implicitly geo-referenced information is hidden in unstructured text, e.g., Wikipedia, social data and news. Toponym recognition is the foundation of mining this useful geo-referenced information by identifying words as toponyms in text. In this paper, we propose an adapted toponym recognition approach based on deep belief network (DBN) by exploring two key issues: word representation and model interpretation. A Skip-Gram model is used in the word representation process to represent words with contextual information that are ignored by current word representation models. We then determine the core hyper-parameters of the DBN model by illustrating the relationship between the performance and the hyper-parameters, e.g., vector dimensionality, DBN structures and probability thresholds. The experiments evaluate the performance of the Skip-Gram model implemented by the Word2Vec open-source tool, determine stable hyper-parameters and compare our approach with a conditional random field (CRF) based approach. The experimental results show that the DBN model outperforms the CRF model with smaller corpus. When the corpus size is large enough, their statistical metrics become approaching. However, their recognition results express differences and complementarity on different kinds of toponyms. More importantly, combining their results can directly improve the performance of toponym recognition relative to their individual performances. It seems that the scale of the corpus has an obvious effect on the performance of toponym recognition. Generally, there is no adequate tagged corpus on specific toponym recognition tasks, especially in the era of Big Data. In conclusion, we believe that the DBN-based approach is a promising and powerful method to extract geo-referenced information from text in the future.
- Published
- 2018
24. Cost-sensitive and hybrid-attribute measure multi-decision tree over imbalanced data sets
- Author
-
Chunlei Du, Fenglian Li, Yu-Chu Tian, Xueying Zhang, Yue Xu, and Xiqian Zhang
- Subjects
Root (linguistics) ,Information Systems and Management ,Decision tree ,02 engineering and technology ,computer.software_genre ,Machine learning ,Theoretical Computer Science ,Artificial Intelligence ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,080605 Decision Support and Group Support Systems ,Mathematics ,080109 Pattern Recognition and Data Mining ,multi-decision tree ,Incremental decision tree ,Measure (data warehouse) ,Binary decision diagram ,business.industry ,minority class ,Node (networking) ,Decision tree learning ,hybrid attribute measure ,cost sensitivity ,Computer Science Applications ,Tree (data structure) ,ComputingMethodologies_PATTERNRECOGNITION ,Control and Systems Engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,Data mining ,business ,computer ,Software ,imbalanced data set - Abstract
One of the most popular algorithms for classification is the decision tree. However, existing binary decision tree models do not handle well the minority class over imbalanced data sets. To address this difficulty, a Cost-sensitive and Hybrid attribute measure Multi-Decision Tree (CHMDT) approach is presented in this paper. It penalizes misclassification through a hybrid attribute measure, which is defined from the combination of the Gini index and information gain measure. It further builds a multi-decision tree consisting of multiple decision trees each with different root node information. The overall objective of the approach is to maximize the classification performance with the hybrid attribute measure while minimizing the total misclassification cost. Experiments are conducted over twelve KEEL imbalanced data sets to demonstrate the CHMDT approach. They show that the classification performance of the minority class is improved significantly without sacrifice of the overall classification accuracy of the majority class.
- Published
- 2018
25. Robust support vector data description for outlier detection with noise or uncertain data
- Author
-
Zizhong John Wang, Xueying Zhang, Fenglian Li, and Chen Guijun
- Subjects
Information Systems and Management ,Uncertain data ,business.industry ,Computer science ,Pattern recognition ,Overfitting ,computer.software_genre ,Management Information Systems ,Constant false alarm rate ,Support vector machine ,Artificial Intelligence ,Robustness (computer science) ,Outlier ,Anomaly detection ,Artificial intelligence ,Data mining ,business ,computer ,Software - Abstract
We propose two new SVDD models which improve the robustness to noise.Cutoff distance-based local density can mitigate the effect of noise towards SVDD.Tolerated gap of SVDD with e-insensitive loss can improve generalization performance. As an example of one-class classification methods, support vector data description (SVDD) offers an opportunity to improve the performance of outlier detection and reduce the loss caused by outlier occurrence in many real-world applications. However, due to limited outliers, the SVDD model is built only by using the normal data. In this situation, SVDD may easily lead to over fitting when the normal data contain noise or uncertainty. This paper presents two types of new SVDD methods, named R-SVDD and eNR-SVDD, which are constructed by introducing cutoff distance-based local density of each data sample and the e-insensitive loss function with negative samples. We have demonstrated that the proposed methods can improve the robustness of SVDD for data with noise or uncertainty by extensive experiments on ten UCI datasets. The experimental results have shown that the proposed eNR-SVDD is superior to other existing outlier detection methods in terms of the detection rate and the false alarm rate. Meanwhile, the proposed R-SVDD can also achieve a better outlier detection performance with only normal data. Finally, the proposed methods are successfully used to detect the image-based conveyor belt fault.
- Published
- 2015
- Full Text
- View/download PDF
26. Power Load Forecasting Based on Swarm Intelligence Algorithm
- Author
-
Jun Yu and Xueying Zhang
- Subjects
Computational Mathematics ,Power load ,business.industry ,Computer science ,General Materials Science ,General Chemistry ,Artificial intelligence ,Electrical and Electronic Engineering ,Condensed Matter Physics ,business ,Swarm intelligence - Published
- 2015
- Full Text
- View/download PDF
27. ELM-based spammer detection in social networks
- Author
-
Tahar Kechadi, Yuanlong Yu, Xueying Zhang, Xianghan Zheng, and Chunming Rong
- Subjects
Social network ,Computer science ,business.industry ,02 engineering and technology ,Crawling ,computer.software_genre ,Machine learning ,Theoretical Computer Science ,Spamming ,Support vector machine ,ComputingMethodologies_PATTERNRECOGNITION ,Hardware and Architecture ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Malware ,020201 artificial intelligence & image processing ,Data mining ,Artificial intelligence ,business ,computer ,Dissemination ,Software ,Information Systems - Abstract
Online social networks, such as Facebook, Twitter, and Weibo have played an important role in people's common life. Most existing social network platforms, however, face the challenges of dealing with undesirable users and their malicious spam activities that disseminate content, malware, viruses, etc. to the legitimate users of the service. The spreading of spam degrades user experience and also negatively impacts server-side functions such as data mining, user behavior analysis, and resource recommendation. In this paper, an extreme learning machine (ELM)-based supervised machine is proposed for effective spammer detection. The work first constructs the labeled dataset through crawling Sina Weibo data and manually classifying corresponding users into spammer and non-spammer categories. A set of features is then extracted from message content and user behavior and applies them to the ELM-based spammer classification algorithm. The experiment and evaluation show that the proposed solution provides excellent performance with a true positive rate of spammers and non-spammers reaching 99 and 99.95 %, respectively. As the results suggest, the proposed solution could achieve better reliability and feasibility compared with existing SVM-based approaches.
- Published
- 2015
- Full Text
- View/download PDF
28. Single-trial event-related potential emotional classification based on compressed sensing
- Author
-
Ying Sun, Lixia Huang, Xueying Zhang, Jiang Chang, Fenglian Li, and Shufei Duan
- Subjects
Computer science ,business.industry ,020208 electrical & electronic engineering ,Feature extraction ,Process (computing) ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,Signal ,Electronic mail ,Statistical classification ,Compressed sensing ,Compression (functional analysis) ,0202 electrical engineering, electronic engineering, information engineering ,Artificial intelligence ,business ,Curse of dimensionality - Abstract
In this study, a robust classification method for emotional speech single-trial event-related potential (ERP) signal was developed. The classification method based on compression sensing (CS) theory. First, we use CS theory to reduce the dimensionality of the ERP signal. Second, the ERP signal was reconstructed by using K-SVD method to construct the over-complete redundant dictionary. Finally, the ERP signal was classified by calculating the residuals between the reconstructed samples and the test samples. The experimental results show that the proposed algorithm can effectively classify the noisy ERP signal and avoid the feature extraction process in the signal recognition.
- Published
- 2017
- Full Text
- View/download PDF
29. Extraction of EEG Components Based on Time - Frequency Blind Source Separation
- Author
-
Huang Lixia, Xueying Zhang, Wei-Rong Wang, Cheng-Ye Shen, and Sun Ying
- Subjects
medicine.diagnostic_test ,business.industry ,Computer science ,Wavelet transform ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,Electroencephalography ,Signal ,Blind signal separation ,Time–frequency analysis ,Support vector machine ,Distribution (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,020201 artificial intelligence & image processing ,Artificial intelligence ,Beta wave ,business - Abstract
In order to extract EEG characteristic waves better, this paper adopts the method of combining wavelet transform with time-frequency blind source separation based on smooth pseudo Wigner-Ville distribution. Firstly, the EEG signal is extracted by wavelet transform to reconstruct the β wave band signal and reconstructed as the initial extracted characteristic wave. Then, to remove the other components which are less relevant to get the enhanced beta wave signal, the time-frequency blind source separation technique based on the smooth pseudo-Wigner distribution is used for the initial extracted Target wave. Finally, the features are extracted, and the support vector machine is used to classify and identify the emotional categories. The experimental results show that the recognition rate is improved when the characteristic wave is extracted by using wavelet transform only.
- Published
- 2017
- Full Text
- View/download PDF
30. A New Fuzzy Cognitive Map Learning Algorithm for Speech Emotion Recognition
- Author
-
Wei Zhang, Ying Sun, and Xueying Zhang
- Subjects
Article Subject ,Computer science ,business.industry ,General Mathematics ,Speech recognition ,lcsh:Mathematics ,General Engineering ,Pattern recognition ,02 engineering and technology ,lcsh:QA1-939 ,01 natural sciences ,Fuzzy cognitive map ,Simple (abstract algebra) ,lcsh:TA1-2040 ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,020201 artificial intelligence & image processing ,Artificial intelligence ,Emotion recognition ,business ,Scale (map) ,lcsh:Engineering (General). Civil engineering (General) ,010301 acoustics ,Algorithm - Abstract
Selecting an appropriate recognition method is crucial in speech emotion recognition applications. However, the current methods do not consider the relationship between emotions. Thus, in this study, a speech emotion recognition system based on the fuzzy cognitive map (FCM) approach is constructed. Moreover, a new FCM learning algorithm for speech emotion recognition is proposed. This algorithm includes the use of the pleasure-arousal-dominance emotion scale to calculate the weights between emotions and certain mathematical derivations to determine the network structure. The proposed algorithm can handle a large number of concepts, whereas a typical FCM can handle only relatively simple networks (maps). Different acoustic features, including fundamental speech features and a new spectral feature, are extracted to evaluate the performance of the proposed method. Three experiments are conducted in this paper, namely, single feature experiment, feature combination experiment, and comparison between the proposed algorithm and typical networks. All experiments are performed on TYUT2.0 and EMO-DB databases. Results of the feature combination experiments show that the recognition rates of the combination features are 10%–20% better than those of single features. The proposed FCM learning algorithm generates 5%–20% performance improvement compared with traditional classification networks.
- Published
- 2017
31. A dissimilarity-based imbalance data classification algorithm
- Author
-
Guangtao Wang, Xiaolin Jia, Kaiyuan Zhang, Xueying Zhang, Qinbao Song, and Liang He
- Subjects
Artificial neural network ,Computer science ,business.industry ,Data classification ,Decision tree ,Context (language use) ,Feature selection ,computer.software_genre ,Machine learning ,Data set ,Naive Bayes classifier ,Statistical classification ,ComputingMethodologies_PATTERNRECOGNITION ,Discriminative model ,Artificial Intelligence ,Data mining ,Artificial intelligence ,business ,computer - Abstract
Class imbalances have been reported to compromise the performance of most standard classifiers, such as Naive Bayes, Decision Trees and Neural Networks. Aiming to solve this problem, various solutions have been explored mainly via balancing the skewed class distribution or improving the existing classification algorithms. However, these methods pay more attention on the imbalance distribution, ignoring the discriminative ability of features in the context of class imbalance data. In this perspective, a dissimilarity-based method is proposed to deal with the classification of imbalanced data. Our proposed method first removes the useless and redundant features by feature selection from the given data set; and then, extracts representative instances from the reduced data as prototypes; finally, projects the reduced data into a dissimilarity space by constructing new features, and builds the classification model with data in the dissimilarity space. Extensive experiments over 24 benchmark class imbalance data sets show that, compared with seven other imbalance data tackling solutions, our proposed method greatly improves the performance of imbalance learning, and outperforms the other solutions with all given classification algorithms.
- Published
- 2014
- Full Text
- View/download PDF
32. A Generic Multilabel Learning-Based Classification Algorithm Recommendation Method
- Author
-
Kaiyuan Zhang, Xueying Zhang, Guangtao Wang, and Qinbao Song
- Subjects
General Computer Science ,Learning problem ,business.industry ,Contrast (statistics) ,Machine learning ,computer.software_genre ,Ranking (information retrieval) ,Set (abstract data type) ,Statistical classification ,Multiple comparison procedure ,Benchmark (computing) ,Learning based ,Artificial intelligence ,business ,computer ,Algorithm ,Mathematics - Abstract
As more and more classification algorithms continue to be developed, recommending appropriate algorithms to a given classification problem is increasingly important. This article first distinguishes the algorithm recommendation methods by two dimensions: (1) meta-features, which are a set of measures used to characterize the learning problems, and (2) meta-target, which represents the relative performance of the classification algorithms on the learning problem. In contrast to the existing algorithm recommendation methods whose meta-target is usually in the form of either the ranking of candidate algorithms or a single algorithm, this article proposes a new and natural multilabel form to describe the meta-target. This is due to the fact that there would be multiple algorithms being appropriate for a given problem in practice. Furthermore, a novel multilabel learning-based generic algorithm recommendation method is proposed, which views the algorithm recommendation as a multilabel learning problem and solves the problem by the mature multilabel learning algorithms. To evaluate the proposed multilabel learning-based recommendation method, extensive experiments with 13 well-known classification algorithms, two kinds of meta-targets such as algorithm ranking and single algorithm, and five different kinds of meta-features are conducted on 1,090 benchmark learning problems. The results show the effectiveness of our proposed multilabel learning-based recommendation method.
- Published
- 2014
- Full Text
- View/download PDF
33. Predicting the number of nearest neighbors for the k-NN classification algorithm
- Author
-
Xueying Zhang and Qinbao Song
- Subjects
Artificial neural network ,Computer science ,business.industry ,Pattern recognition ,Theoretical Computer Science ,k-nearest neighbors algorithm ,Back propagation neural network ,Statistical classification ,Artificial Intelligence ,Computer Vision and Pattern Recognition ,Artificial intelligence ,Benchmark data ,business ,Classifier (UML) ,Algorithm - Abstract
k-Nearest Neighbor k-NN is one of the most widely used classification algorithms. When classifying a new instance, k-NN first finds out its k nearest neighbors, and then classifies it by voting for the categories of the k nearest neighbors. Therefore, an appropriate number of nearest neighbors is critical for the k-NN classifier. However, in present, there is no systematical solution to determine the specific value of k. In order to address this problem, we propose a novel method of using back-propagation neural networks to explore the relationship between data set characteristics and the optimal values of k, then the relationship and the data set characteristics of a new data set are used to recommend the value of k for this data set. The experimental results on the 49 UCI benchmark data sets show that compared with the optimal k values, although there is a decrease of 1.61% in the average classification accuracy for the k-NN classifier with the recommended k values, the time for determining the k values is greatly shortened.
- Published
- 2014
- Full Text
- View/download PDF
34. A Joint Optimized Robust Acoustic Echo Cancellation Algorithm
- Author
-
Jianghao Feng, Hongyan Li, Yue Wang, and Xueying Zhang
- Subjects
0209 industrial biotechnology ,Computer science ,Echo (computing) ,020206 networking & telecommunications ,02 engineering and technology ,Impulse noise ,Least mean squares filter ,020901 industrial engineering & automation ,Artificial Intelligence ,Convergence (routing) ,0202 electrical engineering, electronic engineering, information engineering ,Computer Vision and Pattern Recognition ,Joint (audio engineering) ,Algorithm ,Software - Abstract
When the input signals for acoustic echo cancellation (AEC) are related signals, the convergence speed of the traditional normalized least mean square (NLMS) algorithms is significantly reduced. In this paper, a joint optimization robust AEC algorithm is proposed to solve this problem. Based on the analysis of the convergence of the normalized subband adaptive filtering (NSAF) algorithm, the algorithm is optimized by minimizing the mean square error (MSE) of the NSAF algorithm, combining sub-band time-varying step factor and time-varying regularization parameter to update the filter weight vectors. And when the impulse noise occurs, the sub-band cut-off parameter is updated in a recursive manner, which makes the algorithm achieve fast convergence speed and low steady-state error, and has strong robustness to impulse noise. In a series of experiments on AEC, simulation results show that the performance of the algorithm is better than the existing algorithms.
- Published
- 2019
- Full Text
- View/download PDF
35. Evaluation of a set of new ORF kernel functions of SVM for speech recognition
- Author
-
Zizhong John Wang, Xiaofeng Liu, and Xueying Zhang
- Subjects
business.industry ,Computer science ,Speech recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,Quantitative Biology::Genomics ,Support vector machine ,Kernel (linear algebra) ,ComputingMethodologies_PATTERNRECOGNITION ,Kernel method ,Computer Science::Sound ,Artificial Intelligence ,Control and Systems Engineering ,Kernel embedding of distributions ,String kernel ,Polynomial kernel ,Computer Science::Computer Vision and Pattern Recognition ,Kernel (statistics) ,Least squares support vector machine ,Radial basis function kernel ,Radial basis function ,Artificial intelligence ,Electrical and Electronic Engineering ,Tree kernel ,business - Abstract
The kernel function is the core of the Support Vector Machine (SVM), and its selection directly affects the performance of SVM. There has been no theoretical basis on choosing a kernel function for speech recognition. In order to improve the learning ability and generalization ability of SVM for speech recognition, this paper presents the Optimal Relaxation Factor (ORF) kernel function, which is a set of new SVM kernel functions for speech recognition, and proves that the ORF function is a Mercer kernel function. The experiments show the ORF kernel function's effectiveness on mapping trend, bi-spiral, and speech recognition problems. The paper draws the conclusion that the ORF kernel function performs better than the Radial Basis Function (RBF), the Exponential Radial Basis Function (ERBF) and the Kernel with Moderate Decreasing (KMOD). Furthermore, the results of speech recognition with the ORF kernel function illustrate higher recognition accuracy.
- Published
- 2013
- Full Text
- View/download PDF
36. Parameters Optimization and Application Research of v-Support Vector Machine
- Author
-
Peiyun Xue, Xueying Zhang, Jie Wang, and Jing Bai
- Subjects
Relevance vector machine ,Support vector machine ,General Computer Science ,Computer science ,business.industry ,Artificial intelligence ,Machine learning ,computer.software_genre ,business ,computer - Published
- 2013
- Full Text
- View/download PDF
37. Comparison of Text Sentiment Analysis Based on Machine Learning
- Author
-
Xianghan Zheng and Xueying Zhang
- Subjects
User information ,Artificial neural network ,Computer science ,business.industry ,Sentiment analysis ,Supervised learning ,Feature selection ,02 engineering and technology ,Machine learning ,computer.software_genre ,Support vector machine ,ComputingMethodologies_PATTERNRECOGNITION ,020204 information systems ,Kernel (statistics) ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
Sentiment analysis is a technology with great practical value, it can solve the phenomenon of network comment information disorderly to a certain extent, and accurate positioning of user information required. Currently for Chinese sentiment analysis research is relatively small, including a variety of supervised learning method of classification result and the text feature representation methods and feature selection mechanism and other factors impact on the classification performance is an urgent problem. In this paper, we taken the verb, adjectives and adverbs as text features, used TF-IDF to calculate weight of words. Then we adopted the SVM and ELM with kernels to analyze the text emotion tendentiousness. The experimental results show that ELM with kernels can be obtained a better classification result in a relatively short period of time than SVM.
- Published
- 2016
- Full Text
- View/download PDF
38. Context Enhanced Keyword Extraction for Sparse Geo-Entity Relation from Web Texts
- Author
-
Xueying Zhang, Xiliang Liu, Feng Lu, and Li Yu
- Subjects
Information retrieval ,Computer science ,business.industry ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,Supervised learning ,0211 other engineering and technologies ,Keyword extraction ,02 engineering and technology ,computer.software_genre ,Dynamic feature ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Entropy (information theory) ,Unsupervised learning ,Artificial intelligence ,business ,computer ,Natural language processing ,021101 geological & geomatics engineering - Abstract
Geo-entity relation recognition from rich texts requires robust and effective solutions on keyword extraction. Compared with supervised learning methods, unsupervised learning methods attract more attention for their capability to capture the dynamic feature variation in text and to discover additional relation types. The frequency-based methods of keyword extraction have been widely studied. However, it is difficult to be applied into geo-entity keyword extraction directly because of the sparse distribution of geo-entity relations in texts. Besides, there are few studies on Chinese keyword extraction. This paper proposes a context enhanced keyword extraction method. Firstly the contexts for geo-entities are enhanced to reduce the sparseness of terms. Secondly two well-known frequency-based statistical methods (i.e., DF and Entropy) are used to build a large-scale corpus automatically from the enhanced contexts. Thirdly the lexical features and their weights are statistically determined based on the corpus to enhance the distinction of the terms. Finally, all terms in the enhanced contexts are measured with the lexical features, and the most important terms are selected as the keywords of geo-entity pairs. Experiments are conducted with mass real Chinese web texts. Compared with DF and Entropy, the presented method improves the precision by 41 % and 36 % respectively in discovering the keywords with sparse distribution and generates additional 60 % correct keywords for geo-entity relation recognition.
- Published
- 2016
- Full Text
- View/download PDF
39. Anti-noise Speech Recognition System Based on Improved MFCC Features and Wavelet Kernel SVM
- Author
-
Peiyun Xue, Lihong Yang, Xueying Zhang, and Jing Bai
- Subjects
Support vector machine ,Noise ,General Computer Science ,business.industry ,Computer science ,General Mathematics ,Speech recognition ,Pattern recognition ,Mel-frequency cepstrum ,Artificial intelligence ,business ,Wavelet kernel - Published
- 2012
- Full Text
- View/download PDF
40. Fuzzy Support Vector Machine-Based Emotional Optimal Algorithm in Spoken Chinese
- Author
-
Yuqiang Qin and Xueying Zhang
- Subjects
Computational Mathematics ,Theoretical computer science ,Fuzzy support vector machine ,Computer science ,business.industry ,General Materials Science ,General Chemistry ,Artificial intelligence ,Electrical and Electronic Engineering ,Condensed Matter Physics ,business - Published
- 2012
- Full Text
- View/download PDF
41. Adaptive bands filter bank optimized by genetic algorithm for robust speech recognition system
- Author
-
Xueying Zhang, Lixia Huang, and Gianpaolo Evangelista
- Subjects
Engineering ,business.industry ,Speech recognition ,Feature extraction ,Metals and Alloys ,General Engineering ,Pattern recognition ,Filter bank ,Adaptive filter ,Filter design ,Robustness (computer science) ,Kernel adaptive filter ,Radial basis function ,Artificial intelligence ,business ,Root-raised-cosine filter - Abstract
Perceptual auditory filter banks such as Bark-scale filter bank are widely used as front-end processing in speech recognition systems. However, the problem of the design of optimized filter banks that provide higher accuracy in recognition tasks is still open. Owing to spectral analysis in feature extraction, an adaptive bands filter bank (ABFB) is presented. The design adopts flexible bandwidths and center frequencies for the frequency responses of the filters and utilizes genetic algorithm (GA) to optimize the design parameters. The optimization process is realized by combining the front-end filter bank with the back-end recognition network in the performance evaluation loop. The deployment of ABFB together with zero-crossing peak amplitude (ZCPA) feature as a front process for radial basis function (RBF) system shows significant improvement in robustness compared with the Bark-scale filter bank. In ABFB, several sub-bands are still more concentrated toward lower frequency but their exact locations are determined by the performance rather than the perceptual criteria. For the ease of optimization, only symmetrical bands are considered here, which still provide satisfactory results.
- Published
- 2011
- Full Text
- View/download PDF
42. Speech Enhancement Using Modified MMSE-LSA and Phase Reconstruction in Voiced and Unvoiced Speech
- Author
-
Hairong Jia, Xueying Zhang, Weimei Wang, and Dong Wang
- Subjects
Computer science ,business.industry ,Speech recognition ,020206 networking & telecommunications ,Improved method ,02 engineering and technology ,Speech enhancement ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,Software ,Phase reconstruction - Abstract
Aiming at the problem of auditory negative enhancement of typical phase reconstruction method, an improved method of phase reconstruction and MMSE-LSA estimation is proposed. First, the geometric relationship between noisy speech and clean speech in unvoiced segment is used to estimate the phase of the clean speech; Second, considering the randomness of speech appearance in the actual noise environment, a modified MMSE-LSA amplitude estimation is proposed by using the binary hypothesis model. Finally, the new phase reconstruction in voiced and unvoiced speech is combined with the modified MMSE-LSA. The simulation results show that the performance of the algorithm proposed in this paper is better than typical phase reconstruction method in terms of the SegSNR and PESQ.
- Published
- 2018
- Full Text
- View/download PDF
43. An Unsupervised Two-Talker Speech Separation System Based on CASA
- Author
-
Xueying Zhang, Rongrong Zhao, Hongyan Li, and Yue Wang
- Subjects
Offset (computer science) ,Computer science ,business.industry ,Speech recognition ,Brute-force search ,02 engineering and technology ,Monaural ,Speaker recognition ,030507 speech-language pathology & audiology ,03 medical and health sciences ,Artificial Intelligence ,Computational auditory scene analysis ,0202 electrical engineering, electronic engineering, information engineering ,Beam search ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Mel-frequency cepstrum ,Artificial intelligence ,0305 other medical science ,Cluster analysis ,business ,Software - Abstract
On the basis of the theory about blind separation of monaural speech based on computational auditory scene analysis (CASA), a two-talker speech separation system combining CASA and speaker recognition was proposed to separate speech from other speech interferences in this paper. First, a tandem algorithm is used to organize voiced speech, then based on the clustering of gammatone frequency cepstral coefficients (GFCCs), an object function is established to recognize the speaker, and the best group is achieved through exhaustive search or beam search, so that voiced speech is organized sequentially. Second, unvoiced segments are generated by estimating onset/offset, and then unvoiced–voiced (U–V) segments and unvoiced–unvoiced (U–U) segments are separated respectively. The U–V segments are managed via the binary mask of the separated voiced speech, while the U–V segments are separated evenly. So far the unvoiced segments are separated. The simulation and performance evaluation verify the feasibility and effectiveness of the proposed algorithm.
- Published
- 2018
- Full Text
- View/download PDF
44. A novel method for spammer detection in social networks
- Author
-
Xueying Zhang and Xianghan Zheng
- Subjects
Social network ,Artificial neural network ,business.industry ,Computer science ,Reliability (computer networking) ,computer.software_genre ,Machine learning ,Spamming ,Support vector machine ,Malware ,Artificial intelligence ,business ,computer ,Dissemination ,Extreme learning machine - Abstract
Online social networks have played an important role in people's common life. Most existing social network platforms, however, face the challenges of dealing with undesirable users and their malicious spam activities that disseminate content, malware, viruses, etc. to the legitimate users of the service. In this paper, an Extreme Learning Machine based supervised machine is proposed for effective spammer detection. The experiment and evaluation show that the proposed solution provides excellent performance with a true positive rate of spammers and non-spammers reaching 99% and 99.95%, respectively. As the results suggest, the proposed solution could achieve better reliability and feasibility compared with existing SVM based approaches.
- Published
- 2015
- Full Text
- View/download PDF
45. An Enhanced Artificial Bee Colony-Based Support Vector Machine for Image-Based Fault Detection
- Author
-
Chen Guijun, Xueying Zhang, Zizhong John Wang, and Fenglian Li
- Subjects
Engineering ,Article Subject ,business.industry ,Machine vision ,lcsh:Mathematics ,General Mathematics ,General Engineering ,Chaotic ,Initialization ,Conveyor belt ,Pattern recognition ,lcsh:QA1-939 ,Fault (power engineering) ,Machine learning ,computer.software_genre ,Fault detection and isolation ,Support vector machine ,lcsh:TA1-2040 ,Feature (computer vision) ,Artificial intelligence ,lcsh:Engineering (General). Civil engineering (General) ,business ,computer - Abstract
Fault detection has become extremely important in industrial production so that numerous potential losses caused from equipment failures could be saved. As a noncontact method, machine vision can satisfy the needs of real-time fault monitoring. However, image-based fault features often have the characteristics of high-dimensionality and redundant correlation. To optimize feature subsets and SVM parameters, this paper presents an enhanced artificial bee colony-based support vector machine (EABC-SVM) approach. The method is applied to the image-based fault detection for the conveyor belt. To improve the optimized capability of original ABC, the EABC algorithm introduces two enhanced strategies including the Cat chaotic mapping initialization and current optimum based search equations. Several UCI datasets have been used to evaluate the performance of EABC-SVM and the experimental results show that this approach has better classification accuracy and convergence performance than the ABC-SVM and other ABC variants-based SVM. Furthermore, the EABC-SVM can achieve a significant detection accuracy of 95% and reduce the amount of features about 65% in the conveyor belt fault detection.
- Published
- 2015
- Full Text
- View/download PDF
46. A multi-label learning based kernel automatic recommendation method for support vector machine
- Author
-
Xueying Zhang and Qinbao Song
- Subjects
Graph kernel ,Support Vector Machine ,Computer science ,Feature vector ,lcsh:Medicine ,Kernel principal component analysis ,Kernel (linear algebra) ,String kernel ,Polynomial kernel ,Least squares support vector machine ,Humans ,Radial basis function ,lcsh:Science ,Multidisciplinary ,business.industry ,lcsh:R ,Pattern recognition ,Support vector machine ,Kernel method ,Kernel embedding of distributions ,Variable kernel density estimation ,Kernel (statistics) ,Radial basis function kernel ,Kernel smoother ,Principal component regression ,lcsh:Q ,Artificial intelligence ,Kernel Fisher discriminant analysis ,Tree kernel ,business ,Algorithms ,Research Article - Abstract
Choosing an appropriate kernel is very important and critical when classifying a new problem with Support Vector Machine. So far, more attention has been paid on constructing new kernels and choosing suitable parameter values for a specific kernel function, but less on kernel selection. Furthermore, most of current kernel selection methods focus on seeking a best kernel with the highest classification accuracy via cross-validation, they are time consuming and ignore the differences among the number of support vectors and the CPU time of SVM with different kernels. Considering the tradeoff between classification success ratio and CPU time, there may be multiple kernel functions performing equally well on the same classification problem. Aiming to automatically select those appropriate kernel functions for a given data set, we propose a multi-label learning based kernel recommendation method built on the data characteristics. For each data set, the meta-knowledge data base is first created by extracting the feature vector of data characteristics and identifying the corresponding applicable kernel set. Then the kernel recommendation model is constructed on the generated meta-knowledge data base with the multi-label classification method. Finally, the appropriate kernel functions are recommended to a new data set by the recommendation model according to the characteristics of the new data set. Extensive experiments over 132 UCI benchmark data sets, with five different types of data set characteristics, eleven typical kernels (Linear, Polynomial, Radial Basis Function, Sigmoidal function, Laplace, Multiquadric, Rational Quadratic, Spherical, Spline, Wave and Circular), and five multi-label classification methods demonstrate that, compared with the existing kernel selection methods and the most widely used RBF kernel function, SVM with the kernel function recommended by our proposed method achieved the highest classification performance.
- Published
- 2015
47. A Neural Learning Algorithm of Blind Separation of Noisy Mixed Images Based on Independent Component Analysis
- Author
-
Hongyan Li and Xueying Zhang
- Subjects
Signal processing ,General Computer Science ,Artificial neural network ,Computer science ,business.industry ,Pattern recognition ,Blind signal separation ,Independent component analysis ,Noise ,Windage ,Convergence (routing) ,Artificial intelligence ,business ,Algorithm ,Active noise control - Abstract
Blind source separation problem has recently received a great deal of attention in signal processing and unsupervised neural learning. In the current approaches, the additive noise is negligible so that it can be omitted from the consideration. To be applicable in realistic scenarios, blind source separation approaches should deal evenly with the presence of noise. In this contribution, we proposed approaches to independent component analysis when the measured signals are contaminated by additive noise. A noisy multiple channels neural learning algorithm of blind separation is proposed based on independent component analysis. The data have no noise are used to whiten the noisy data, and the windage wipe off technique is used to correct the infection of noise, a neural network model having denoise capability is adopted to recover some original signals from their noisy mixtures observed by the same number of sensors. And a relaxation factor is introduced into the iteration algorithm, thus the new algorithm can implement convergence. Computer simulations and experiment results prove the feasibility and validity of the neural network modeling and control method based on independent component analysis, which can renew the original images effectively.
- Published
- 2014
- Full Text
- View/download PDF
48. A Feature Subset Selection Algorithm Automatic Recommendation Method
- Author
-
Yuming Zhou, Baowen Xu, Xueying Zhang, Qinbao Song, Heli Sun, and Guangtao Wang
- Subjects
FOS: Computer and information sciences ,Meta learning (computer science) ,Computer science ,Feature selection ,computer.software_genre ,Machine Learning (cs.LG) ,Data set ,Computer Science - Learning ,Artificial Intelligence ,Metric (mathematics) ,Feature (machine learning) ,Data mining ,Selection algorithm ,computer ,Selection (genetic algorithm) - Abstract
Many feature subset selection (FSS) algorithms have been proposed, but not all of them are appropriate for a given feature selection problem. At the same time, so far there is rarely a good way to choose appropriate FSS algorithms for the problem at hand. Thus, FSS algorithm automatic recommendation is very important and practically useful. In this paper, a meta learning based FSS algorithm automatic recommendation method is presented. The proposed method first identifies the data sets that are most similar to the one at hand by the k-nearest neighbor classification algorithm, and the distances among these data sets are calculated based on the commonly-used data set characteristics. Then, it ranks all the candidate FSS algorithms according to their performance on these similar data sets, and chooses the algorithms with best performance as the appropriate ones. The performance of the candidate FSS algorithms is evaluated by a multi-criteria metric that takes into account not only the classification accuracy over the selected features, but also the runtime of feature selection and the number of selected features. The proposed recommendation method is extensively tested on 115 real world data sets with 22 well-known and frequently-used different FSS algorithms for five representative classifiers. The results show the effectiveness of our proposed FSS algorithm recommendation method.
- Published
- 2014
- Full Text
- View/download PDF
49. CRFs in Music Chord Recognition Algorithm Application Research
- Author
-
Feng Wang and Xueying Zhang
- Subjects
General Computer Science ,Computer science ,business.industry ,Speech recognition ,Chord recognition ,Pattern recognition ,ComputingMethodologies_PATTERNRECOGNITION ,Character (mathematics) ,Recognition system ,Music recognition ,Artificial intelligence ,Mel-frequency cepstrum ,business ,CRFS - Abstract
In this paper, five different methods in music recognition are discussed, a new character MPCP is proposed in Chord Recognition. The new character overcome the limitation of the traditional PCP and MFCC, apply for recognition system by combining both characteristics. For features we use MPCP vectors, it is trained by CRFs, finally we acquired accurate chords by Viterbi.The experiment show that the proposed strategy can reach a good performance in chord recognition and proved our strategy to be quite promising.
- Published
- 2013
- Full Text
- View/download PDF
50. Parameter Optimization and Application of Support Vector Machine Based on Parallel Artificial Fish Swarm Algorithm
- Author
-
Lihong Yang, Xueying Zhang, and Jing Bai
- Subjects
Linde–Buzo–Gray algorithm ,Structured support vector machine ,business.industry ,Computer science ,Population-based incremental learning ,Swarm behaviour ,Pattern recognition ,Human-Computer Interaction ,Support vector machine ,Relevance vector machine ,Kernel (linear algebra) ,Kernel method ,Artificial Intelligence ,Robustness (computer science) ,Artificial intelligence ,business ,Algorithm ,Software - Abstract
Parameters selection of support vector machine is a very important problem, which has great influence on its performance. In order to improve the learning and generalization ability of support vector machine, in this paper, proposed a new algorithm -parallel artificial fish swarm algorithm to optimize kernel parameter and penalty factor of support vector machine, improved the loop body of artificial fish swarm algorithm to avoid the missing of the optimum solution, and proved its validity by testing with some test functions; used the optimal parameters in a non-specific persons, isolated words, and medium-vocabulary speech recognition system. The experimental results show that the rates of speech recognition based on support vector machine using the new algorithm are better than those of using the traditional artificial fish swarm algorithm in different signal to noise ratios and different words. Especially, the support vector machine model based on the new algorithm can still maintain better recognition rates in lower signal to noise ratios. So the new algorithm is an effective support vector machine parameter optimization method, which makes the support vector machine not only have good generalization ability, but have better robustness.
- Published
- 2013
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.