30 results on '"Uddin, Md Palash"'
Search Results
2. Empirical curvelet transform based deep DenseNet model to predict NDVI using RGB drone imagery data
- Author
-
Diykh, Mohammed, Ali, Mumtaz, Jamei, Mehdi, Abdulla, Shahab, Uddin, Md Palash, Farooque, Aitazaz Ahsan, Labban, Abdulhaleem H., and Alabdally, Hussein
- Published
- 2024
- Full Text
- View/download PDF
3. Towards privacy-preserving Alzheimer's disease classification: Federated learning on T1-weighted magnetic resonance imaging data.
- Author
-
Sahid, Md Abdus, Uddin, Md Palash, Saha, Hasi, and Islam, Md Rashedul
- Published
- 2024
- Full Text
- View/download PDF
4. A Multi-level ensemble approach for skin lesion classification using Customized Transfer Learning with Triple Attention.
- Author
-
Efat, Anwar Hossain, Hasan, S. M. Mahedy, Uddin, Md. Palash, and Mamun, Md. Al
- Subjects
CONVOLUTIONAL neural networks ,DEEP learning ,SKIN cancer ,CELL growth ,SKIN diseases - Abstract
Skin lesions encompass a variety of skin abnormalities, including skin diseases that affect structure and function, and skin cancer, which can be fatal and arise from abnormal cell growth. Early detection of lesions and automated prediction is crucial, yet accurately identifying responsible regions post-dominance dispersion remains a challenge in current studies. Thus, we propose a Convolutional Neural Network (CNN)-based approach employing a Customized Transfer Learning (CTL) model and Triple Attention (TA) modules in conjunction with Ensemble Learning (EL). While Ensemble Learning has become an integral component of both Machine Learning (ML) and Deep Learning (DL) methodologies, a specific technique ensuring optimal allocation of weights for each model's prediction is currently lacking. Consequently, the primary objective of this study is to introduce a novel method for determining optimal weights to aggregate the contributions of models for achieving desired outcomes. We term this approach "Information Gain Proportioned Averaging (IGPA)," further refining it to "Multi-Level Information Gain Proportioned Averaging (ML-IGPA)," which specifically involves the utilization of IGPA at multiple levels. Empirical evaluation of the HAM1000 dataset demonstrates that our approach achieves 94.93% accuracy with ML-IGPA, surpassing state-of-the-art methods. Given previous studies' failure to elucidate the exact focus of black-box models on specific regions, we utilize the Gradient Class Activation Map (GradCAM) to identify responsible regions and enhance explainability. Our study enhances both accuracy and interpretability, facilitating early diagnosis and preventing the consequences of neglecting skin lesion detection, thereby addressing issues related to time, accessibility, and costs. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. A Comprehensive Survey on Edge Data Integrity Verification: Fundamentals and Future Trends.
- Author
-
Zhao, Yao, Qu, Youyang, Xiang, Yong, Uddin, Md Palash, Peng, Dezhong, and Gao, Longxiang
- Published
- 2025
- Full Text
- View/download PDF
6. Enhancing tertiary students' programming skills with an explainable Educational Data Mining approach.
- Author
-
Islam, Md Rashedul, Nitu, Adiba Mahjabin, Marjan, Md Abu, Uddin, Md Palash, Afjal, Masud Ibn, and Mamun, Md Abdulla Al
- Subjects
DATA mining ,MACHINE learning ,ARTIFICIAL intelligence ,RECEIVER operating characteristic curves ,CLASSIFICATION - Abstract
Educational Data Mining (EDM) holds promise in uncovering insights from educational data to predict and enhance students' performance. This paper presents an advanced EDM system tailored for classifying and improving tertiary students' programming skills. Our approach emphasizes effective feature engineering, appropriate classification techniques, and the integration of Explainable Artificial Intelligence (XAI) to elucidate model decisions. Through rigorous experimentation, including an ablation study and evaluation of six machine learning algorithms, we introduce a novel ensemble method, Stacking-SRDA, which outperforms others in accuracy, precision, recall, f1-score, ROC curve, and McNemar test. Leveraging XAI tools, we provide insights into model interpretability. Additionally, we propose a system for identifying skill gaps in programming among weaker students, offering tailored recommendations for skill enhancement. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Leveraging textual information for social media news categorization and sentiment analysis.
- Author
-
Hasan, Mahmudul, Ahmed, Tanver, Islam, Md. Rashedul, and Uddin, Md. Palash
- Subjects
SENTIMENT analysis ,MACHINE learning ,SOCIAL media ,USER-generated content ,K-nearest neighbor classification ,DECISION trees - Abstract
The rise of social media has changed how people view connections. Machine Learning (ML)-based sentiment analysis and news categorization help understand emotions and access news. However, most studies focus on complex models requiring heavy resources and slowing inference times, making deployment difficult in resource-limited environments. In this paper, we process both structured and unstructured data, determining the polarity of text using the TextBlob scheme to determine the sentiment of news headlines. We propose a Stochastic Gradient Descent (SGD)-based Ridge classifier (RC) for blending SGDR with an advanced string processing technique to effectively classify news articles. Additionally, we explore existing supervised and unsupervised ML algorithms to gauge the effectiveness of our SGDR classifier. The scalability and generalization capability of SGD and L2 regularization techniques in RCs to handle overfitting and balance bias and variance provide the proposed SGDR with better classification capability. Experimental results highlight that our string processing pipeline significantly boosts the performance of all ML models. Notably, our ensemble SGDR classifier surpasses all state-of-the-art ML algorithms, achieving an impressive 98.12% accuracy. McNemar's significance tests reveal that our SGDR classifier achieves a 1% significance level improvement over K-Nearest Neighbor, Decision Tree, and AdaBoost and a 5% significance level improvement over other algorithms. These findings underscore the superior proficiency of linear models in news categorization compared to tree-based and nonlinear counterparts. This study contributes valuable insights into the efficacy of the proposed methodology, elucidating its potential for news categorization and sentiment analysis. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Band reordering heuristics for lossless satellite image compression with 3D-CALIC and CCSDS
- Author
-
Afjal, Masud Ibn, Mamun, Md. Al, and Uddin, Md. Palash
- Published
- 2019
- Full Text
- View/download PDF
9. Improving Hyperspectral Image Classification with Compact Multi-Branch Deep Learning.
- Author
-
Islam, Md. Rashedul, Islam, Md. Touhid, Uddin, Md Palash, and Ulhaq, Anwaar
- Subjects
IMAGE recognition (Computer vision) ,DEEP learning ,FEATURE extraction ,FACTOR analysis - Abstract
The progress in hyperspectral image (HSI) classification owes much to the integration of various deep learning techniques. However, the inherent 3D cube structure of HSIs presents a unique challenge, necessitating an innovative approach for the efficient utilization of spectral data in classification tasks. This research focuses on HSI classification through the adoption of a recently validated deep-learning methodology. Challenges in HSI classification encompass issues related to dimensionality, data redundancy, and computational expenses, with CNN-based methods prevailing due to architectural limitations. In response to these challenges, we introduce a groundbreaking model known as "Crossover Dimensionality Reduction and Multi-branch Deep Learning" (CMD) for hyperspectral image classification. The CMD model employs a multi-branch deep learning architecture incorporating Factor Analysis and MNF for crossover feature extraction, with the selection of optimal features from each technique. Experimental findings underscore the CMD model's superiority over existing methods, emphasizing its potential to enhance HSI classification outcomes. Notably, the CMD model exhibits exceptional performance on benchmark datasets such as Salinas Scene (SC), Pavia University (PU), Kennedy Space Center (KSC), and Indian Pines (IP), achieving impressive overall accuracy rates of 99.35% and 99.18% using only 5% of the training data. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Predictive modeling of multi-class diabetes mellitus using machine learning and filtering iraqi diabetes data dynamics.
- Author
-
Sahid, Md Abdus, Babar, Mozaddid Ul Hoque, and Uddin, Md Palash
- Subjects
MACHINE learning ,DIABETES ,HYPERGLYCEMIA ,BLOOD sugar ,SUPPORT vector machines ,FEATURE selection ,K-nearest neighbor classification ,INSULIN - Abstract
Diabetes is a persistent metabolic disorder linked to elevated levels of blood glucose, commonly referred to as blood sugar. This condition can have detrimental effects on the heart, blood vessels, eyes, kidneys, and nerves as time passes. It is a chronic ailment that arises when the body fails to produce enough insulin or is unable to effectively use the insulin it produces. When diabetes is not properly managed, it often leads to hyperglycemia, a condition characterized by elevated blood sugar levels or impaired glucose tolerance. This can result in significant harm to various body systems, including the nerves and blood vessels. In this paper, we propose a multiclass diabetes mellitus detection and classification approach using an extremely imbalanced Laboratory of Medical City Hospital data dynamics. We also formulate a new dataset that is moderately imbalanced based on the Laboratory of Medical City Hospital data dynamics. To correctly identify the multiclass diabetes mellitus, we employ three machine learning classifiers namely support vector machine, logistic regression, and k-nearest neighbor. We also focus on dimensionality reduction (feature selection—filter, wrapper, and embedded method) to prune the unnecessary features and to scale up the classification performance. To optimize the classification performance of classifiers, we tune the model by hyperparameter optimization with 10-fold grid search cross-validation. In the case of the original extremely imbalanced dataset with 70:30 partition and support vector machine classifier, we achieved maximum accuracy of 0.964, precision of 0.968, recall of 0.964, F1-score of 0.962, Cohen kappa of 0.835, and AUC of 0.99 by using top 4 feature according to filter method. By using the top 9 features according to wrapper-based sequential feature selection, the k-nearest neighbor provides an accuracy of 0.935 and 1.0 for the other performance metrics. For our created moderately imbalanced dataset with an 80:20 partition, the SVM classifier achieves a maximum accuracy of 0.938, and 1.0 for other performance metrics. For the multiclass diabetes mellitus detection and classification, our experiments outperformed conducted research based on the Laboratory of Medical City Hospital data dynamics. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Improving hyperspectral image classification through spectral-spatial feature reduction with a hybrid approach and deep learning.
- Author
-
Islam, Md. Rashedul, Islam, Md. Touhid, and Uddin, Md. Palash
- Subjects
IMAGE recognition (Computer vision) ,DEEP learning ,CONVOLUTIONAL neural networks ,FEATURE selection ,FEATURE extraction ,REMOTE sensing - Abstract
The Hyperspectral Image (HSI) is a great source of information for observing the earth's elements due to its numerous narrow and continuous spectral wavelength bands. There are some key difficulties, such as the fact that the image bands are spectrally and spatially highly correlated, and the 'curse of dimensionality' in using the original HSI for classification. Band (dimensionality/feature) reduction is necessary to improve classification performance through feature extraction and selection. As such, this paper proposes a hybrid HSI classification paradigm, termed Hybrid-2DNET. Specifically, our proposed Hybrid-2DNET incorporates factor analysis-based feature extraction along the Minimum-Redundancy-Maximum-Relevance (mRMR)-based feature selection criterion and a 2D-wavelet Convolutional Neural Network (CNN) to reduce both spectral and spatial dimensionalities. The experimental analyses performed on three remote sensings HSI datasets, namely Indian Pines, University of Pavia, and Salinas Scene, manifest that our proposed Hybrid-2DNET outperforms the handcrafted and deep learning methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. Performance discrepancy mitigation in heart disease prediction for multisensory inter-datasets.
- Author
-
Hasan, Mahmudul, Sahid, Md Abdus, Uddin, Md Palash, Marjan, Md Abu, Kadry, Seifedine, and Jungeun Kim
- Subjects
HEART diseases ,MACHINE learning ,FEATURE selection ,SUPPORT vector machines ,K-nearest neighbor classification ,RANDOM forest algorithms ,FEATURE extraction - Abstract
Heart disease is one of the primary causes of morbidity and death worldwide. Millions of people have had heart attacks every year, and only early-stage predictions can help to reduce the number. Researchers are working on designing and developing early-stage prediction systems using different advanced technologies, and machine learning (ML) is one of them. Almost all existing ML-based works consider the same dataset (intra-dataset) for the training and validation of their method. In particular, they do not consider inter-dataset performance checks, where different datasets are used in the training and testing phases. In inter-dataset setup, existing ML models show a poor performance named the inter-dataset discrepancy problem. This work focuses on mitigating the inter-dataset discrepancy problem by considering five available heart disease datasets and their combined form. All potential training and testing mode combinations are systematically executed to assess discrepancies before and after applying the proposed methods. Imbalance data handling using SMOTE-Tomek, feature selection using random forest (RF), and feature extraction using principle component analysis (PCA) with a long preprocessing pipeline are used to mitigate the inter-dataset discrepancy problem. The preprocessing pipeline builds on missing value handling using RF regression, log transformation, outlier removal, normalization, and data balancing that convert the datasets to more ML-centric. Support vector machine, K-nearest neighbors, decision tree, RF, eXtreme Gradient Boosting, Gaussian naive Bayes, logistic regression, and multilayer perceptron are used as classifiers. Experimental results show that feature selection and classification using RF produce better results than other combination strategies in both single- and inter-dataset setups. In certain configurations of individual datasets, RF demonstrates 100% accuracy and 96% accuracy during the feature selection phase in an inter-dataset setup, exhibiting commendable precision, recall, F1 score, specificity, and AUC score. The results indicate that an effective preprocessing technique has the potential to improve the performance of the ML model without necessitating the development of intricate prediction models. Addressing inter-dataset discrepancies introduces a novel research avenue, enabling the amalgamation of identical features from various datasets to construct a comprehensive global dataset within a specific domain. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. A novel data balancing technique via resampling majority and minority classes toward effective classification.
- Author
-
Hasan, Mahmudul, Rabbi, Md. Fazle, Sultan, Md. Nahid, Nitu, Adiba Mahjabin, and Uddin, Md. Palash
- Subjects
MEDICAL informatics ,RANDOM forest algorithms ,HEART diseases ,CLASSIFICATION ,RESAMPLING (Statistics) ,LOGISTIC regression analysis ,MACHINE learning ,LIFTING & carrying (Human mechanics) - Abstract
Classification is a predictive modelling task in machine learning (ML), where the class label is determined for a specific example of predefined features. In determining handwriting characters, identifying spam, detecting disease, identifying signals, and so on, classification requires training data with many features and label instances. In medical informatics, high precision and recall are mandatory issues besides the high accuracy of the ML classifiers. Most of the real-life datasets have imbalanced characteristics that hamper the overall performance of the classifiers. Existing data balancing techniques perform the whole dataset at a time that sometimes causes overfitting and underfitting. We propose a data balancing technique that follows the divide and conquer procedure to cluster the dataset into several segments, and both oversampling and undersampling operation is performed on each cluster. Finally, the cluster joined together and built a balanced dataset. We chose the sample data of two heart disease datasets: Hungarian and Long Beach. Logistic regression and random forest classifier are the representatives of ML algorithms. We compare our proposed techniques with existing SMOTE, NearMiss, and SMOTETomek data balancing techniques. Both algorithms perform better on the proposed technique-balanced dataset. This technique can be the optimal solution for the imbalanced data handling strategy. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
14. An IoT-Enabled Ground Loop Detection System: Design, Implementation and Testing.
- Author
-
Rahman, Md. Saifur, Uddin, Md. Palash, and Kim, Sikyung
- Subjects
- *
ONLINE identities , *INTERNET of things , *DETECTOR circuits , *CABLE structures - Abstract
The ground loop is a solemn problem in complex environments including laboratories and industries. In particular, it creates spurious signals, which interfere with low-level signals of instrumentation, and often imperil the human community. Manual ground loop detection is inefficient and requires more diagnosis time. As such, automatic ground loop detection is demanding although it is still a complex task in an environment of massive instruments. In this paper, we exploit the Internet of Things (IoT) technology to present a novel ground loop detection system to cope with such a difficult scenario. Specifically, the proposed scheme comprises an exciter block along with the IoT device to generate up to 100 kHz ground loop current, and a detector module to regulate the affected cable by receiving the test current. We also use multiple detectors to give a virtual cable identity (ID) number in a complex area for recognizing the faulty cable accurately. After detecting the ground loop, the affected cable ID number is sent to the server for immediate action for prevention through the use of a smartphone (Android) application and website. The test results clarify the superiority of the proposed ground loop detection scheme in terms of accuracy, dependency and robustness. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
15. Ensemble machine learningbased recommendation system for effective prediction of suitable agricultural crop cultivation.
- Author
-
Hasan, Mahmudul, Marjan, Md Abu, Uddin, Md Palash, Ibn Afjal, Masud, Kardy, Seifedine, Shaoqi Ma, and Yunyoung Nam
- Subjects
CROPS ,MACHINE learning ,RECOMMENDER systems ,AGRICULTURAL productivity ,POTATOES ,FOOD supply - Abstract
Agriculture is the most critical sector for food supply on the earth, and it is also responsible for supplying raw materials for other industrial productions. Currently, the growth in agricultural production is not sufficient to keep up with the growing population, which may result in a food shortfall for the world’s inhabitants. As a result, increasing food production is crucial for developing nations with limited land and resources. It is essential to select a suitable crop for a specific region to increase its production rate. Effective crop production forecasting in that area based on historical data, including environmental and cultivation areas, and crop production amount, is required. However, the data for such forecasting are not publicly available. As such, in this paper, we take a case study of a developing country, Bangladesh, whose economy relies on agriculture. We first gather and preprocess the data from the relevant research institutions of Bangladesh and then propose an ensemble machine learning approach, called K-nearest Neighbor Random Forest Ridge Regression (KRR), to effectively predict the production of the major crops (three different kinds of rice, potato, and wheat). KRR is designed after investigating five existing traditional machine learning (Support Vector Regression, Naïve Bayes, and Ridge Regression) and ensemble learning (Random Forest and CatBoost) algorithms. We consider four classical evaluation metrics, i.e., mean absolute error, mean square error (MSE), root MSE, and R², to evaluate the performance of the proposed KRR over the other machine learning models. It shows 0.009 MSE, 99% R² for Aus; 0.92 MSE, 90% R² for Aman; 0.246 MSE, 99% R² for Boro; 0.062 MSE, 99% R² for wheat; and 0.016 MSE, 99% R² for potato production prediction. The Diebold–Mariano test is conducted to check the robustness of the proposed ensemble model, KRR. In most cases, it shows 1% and 5% significance compared to the benchmark ML models. Lastly, we design a recommender system that suggests suitable crops for a specific land area for cultivation in the next season. We believe that the proposed paradigm will help the farmers and personnel in the agricultural sector leverage proper crop cultivation and production. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
16. A Deep Learning-Based Hyperspectral Object Classification Approach via Imbalanced Training Samples Handling.
- Author
-
Islam, Md Touhid, Islam, Md Rashedul, Uddin, Md Palash, and Ulhaq, Anwaar
- Subjects
DEEP learning ,IMAGE recognition (Computer vision) ,CLASSIFICATION ,ELECTRONIC data processing ,MATRIX decomposition ,NONNEGATIVE matrices - Abstract
Object classification in hyperspectral images involves accurately categorizing objects based on their spectral characteristics. However, the high dimensionality of hyperspectral data and class imbalance pose significant challenges to object classification performance. To address these challenges, we propose a framework that incorporates dimensionality reduction and re-sampling as preprocessing steps for a deep learning model. Our framework employs a novel subgroup-based dimensionality reduction technique to extract and select the most informative features with minimal redundancy. Additionally, the data are resampled to achieve class balance across all categories. The reduced and balanced data are then processed through a hybrid CNN model, which combines a 3D learning block and a 2D learning block to extract spectral–spatial features and achieve satisfactory classification accuracy. By adopting this hybrid approach, we simplify the model while improving performance in the presence of noise and limited sample size. We evaluated our proposed model on the Salinas scene, Pavia University, and Kennedy Space Center benchmark hyperspectral datasets, comparing it to state-of-the-art methods. Our object classification technique achieves highly promising results, with overall accuracies of 99.98%, 99.94%, and 99.46% on the three datasets, respectively. This proposed approach offers a compelling solution to overcome the challenges of high dimensionality and class imbalance in hyperspectral object classification. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
17. Blockchain-enabled Federated Learning: A Survey.
- Author
-
YOUYANG QU, UDDIN, MD PALASH, CHENQUAN GAN, YONG XIANG, LONGXIANG GAO, and YEARWOOD, JOHN
- Subjects
- *
BLOCKCHAINS , *ARTIFICIAL intelligence , *FALSIFICATION of data , *MACHINE learning , *LEARNING , *PRIVACY - Abstract
Federated learning (FL) has experienced a boom in recent years, which is jointly promoted by the prosperity of machine learning and Artificial Intelligence along with emerging privacy issues. In the FL paradigm, a central server and local end devices maintain the same model by exchanging model updates instead of raw data, with which the privacy of data stored on end devices is not directly revealed. In this way, the privacy violation caused by the growing collection of sensitive data can be mitigated. However, the performance of FL with a central server is reaching a bottleneck, while new threats are emerging simultaneously. There are various reasons, among which the most significant ones are centralized processing, data falsification, and lack of incentives. To accelerate the proliferation of FL, blockchain-enabled FL has attracted substantial attention from both academia and industry. A considerable number of novel solutions are devised to meet the emerging demands of diverse scenarios. Blockchain-enabled FL provides both theories and techniques to improve the performance of FL from various perspectives. In this survey, we will comprehensively summarize and evaluate existing variants of blockchain-enabled FL, identify the emerging challenges, and propose potentially promising research directions in this under-explored domain. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
18. An Educational Data Mining System For Predicting And Enhancing Tertiary Students' Programming Skill.
- Author
-
Marjan, Md Abu, Uddin, Md Palash, and Afjal, Masud Ibn
- Subjects
- *
DATA mining , *NAIVE Bayes classification , *COMPUTER literacy , *K-nearest neighbor classification , *LEARNING , *SUPPORT vector machines - Abstract
Educational Data Mining (EDM) has become a promising research field for improving the quality of students and the education system. Although EDM dates back to several years, there is still lack of works for measuring and enhancing the computer programming skills of tertiary students. As such, we, in this paper, propose an EDM system for evaluating and improving tertiary students' programming skills. The proposed EDM system comprises two key modules for (i) classification process and (ii) learning process,. The classification module predicts the current status of a student and the learning process module helps generate respective suggestions and feedback to enhance the student's quality. In particular, for the classification module, we prepare a real dataset related to this task and evaluate the dataset to investigate six key Machine Learning (ML) algorithms, Support Vector Machine (SVM), decision tree, artificial neural network, Random Forest (RF), k -nearest neighbor and naive Bayes classifier, using accuracy-related performance measure metrics and goodness of the fit. The experimental results manifest that RF and SVM can predict the students more accurately than the other models. In addition, critical factors analysis is accomplished to identify the critical features toward achieving high classification accuracy. At last, we design an improvement mechanism in the learning process module that helps the students enhance their programming skills. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
19. Hyperspectral Image Classification via Information Theoretic Dimension Reduction.
- Author
-
Islam, Md Rashedul, Siddiqa, Ayasha, Ibn Afjal, Masud, Uddin, Md Palash, and Ulhaq, Anwaar
- Subjects
FEATURE selection ,FEATURE extraction ,SUPPORT vector machines ,DATA mining ,HIGH-dimensional model representation ,CUBES - Abstract
Hyperspectral images (HSIs) are one of the most successfully used tools for precisely and potentially detecting key ground surfaces, vegetation, and minerals. HSIs contain a large amount of information about the ground scene; therefore, object classification becomes the most difficult task for such a high-dimensional HSI data cube. Additionally, the HSI's spectral bands exhibit a high correlation, and a large amount of spectral data creates high dimensionality issues as well. Dimensionality reduction is, therefore, a crucial step in the HSI classification pipeline. In order to identify a pertinent subset of features for effective HSI classification, this study proposes a dimension reduction method that combines feature extraction and feature selection. In particular, we exploited the widely used denoising method minimum noise fraction (MNF) for feature extraction and an information theoretic-based strategy, cross-cumulative residual entropy (CCRE), for feature selection. Using the normalized CCRE, minimum redundancy maximum relevance (mRMR)-driven feature selection criteria were used to enhance the quality of the selected feature. To assess the effectiveness of the extracted features' subsets, the kernel support vector machine (KSVM) classifier was applied to three publicly available HSIs. The experimental findings manifest a discernible improvement in classification accuracy and the qualities of the selected features. Specifically, the proposed method outperforms the traditional methods investigated, with overall classification accuracies on Indian Pines, Washington DC Mall, and Pavia University HSIs of 97.44%, 99.71%, and 98.35%, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
20. Mutual Information-Driven Feature Reduction for Hyperspectral Image Classification.
- Author
-
Islam, Md Rashedul, Ahmed, Boshir, Hossain, Md Ali, and Uddin, Md Palash
- Subjects
IMAGE registration ,FEATURE extraction ,FEATURE selection ,PRINCIPAL components analysis ,SUPPORT vector machines ,GROUND cover plants - Abstract
A hyperspectral image (HSI), which contains a number of contiguous and narrow spectral wavelength bands, is a valuable source of data for ground cover examinations. Classification using the entire original HSI suffers from the "curse of dimensionality" problem because (i) the image bands are highly correlated both spectrally and spatially, (ii) not every band can carry equal information, (iii) there is a lack of enough training samples for some classes, and (iv) the overall computational cost is high. Therefore, effective feature (band) reduction is necessary through feature extraction (FE) and/or feature selection (FS) for improving the classification in a cost-effective manner. Principal component analysis (PCA) is a frequently adopted unsupervised FE method in HSI classification. Nevertheless, its performance worsens when the dataset is noisy, and the computational cost becomes high. Consequently, this study first proposed an efficient FE approach using a normalized mutual information (NMI)-based band grouping strategy, where the classical PCA was applied to each band subgroup for intrinsic FE. Finally, the subspace of the most effective features was generated by the NMI-based minimum redundancy and maximum relevance (mRMR) FS criteria. The subspace of features was then classified using the kernel support vector machine. Two real HSIs collected by the AVIRIS and HYDICE sensors were used in an experiment. The experimental results demonstrated that the proposed feature reduction approach significantly improved the classification performance. It achieved the highest overall classification accuracy of 94.93% for the AVIRIS dataset and 99.026% for the HYDICE dataset. Moreover, the proposed approach reduced the computational cost compared with the studied methods. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
21. Improved folded-PCA for efficient remote sensing hyperspectral image classification.
- Author
-
Uddin, Md. Palash, Mamun, Md. Al, Hossain, Md. Ali, and Afjal, Masud Ibn
- Subjects
- *
IMAGE recognition (Computer vision) , *FEATURE extraction , *PRINCIPAL components analysis , *FEATURE selection , *REMOTE sensing , *HYPERSPECTRAL imaging systems , *AGRICULTURE - Abstract
Hyperspectral images (HSIs) contain notable information of land objects by acquiring an immense set of narrow and contiguous spectral bands. Feature extraction (FE) and feature selection (FS) as dimensionality (band) reduction strategies are performed to enhance the classification result of HSI. Principal component analysis (PCA) is frequently exploited for the FE of HSI. However, it often possesses the inability to extract local and subtle HSI structures. As such, segmented-PCA (SPCA), spectrally segmented-PCA (SSPCA) and folded-PCA (FPCA) are presented for local and useful FE from the HSI. In this paper, we propose two FE methods called segmented-FPCA (SFPCA) and spectrally segmented-FPCA (SSFPCA). SFPCA exploits SPCA and FPCA while SSFPCA exploits SSPCA and FPCA together. In particular, SFPCA and SSFPCA apply FPCA on highly correlated and spectrally grouped HSI bands, respectively. We consider nonlinear methods Kernel-PCA (KPCA) and Kernel entropy component analysis (KECA) for extended comparison. For the experimented agricultural Indian Pine and urban Washington DC Mall HSIs, the results manifest that SFPCA (95.6262% for the agricultural HSI and 97.4782% for the urban HSI) and SSFPCA (96.3221% for the agricultural HSI and 98.0116% for the urban HSI) outperform the conventional methods. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
22. Robust Federated Averaging via Outlier Pruning.
- Author
-
Uddin, Md Palash, Xiang, Yong, Yearwood, John, and Gao, Longxiang
- Subjects
AGGREGATION operators ,ARITHMETIC ,DEEP learning - Abstract
Federated Averaging (FedAvg) is the baseline Federated Learning (FL) algorithm that applies the stochastic gradient descent for local model training and the arithmetic averaging of the local models’ parameters for global model aggregation. Succeeding FL works commonly utilize the arithmetic averaging scheme of FedAvg for the aggregation. However, such arithmetic averaging is prone to the outlier model-updates, especially when the clients’ data are non-Independent and Identically Distributed (non-IID). As such, the classical aggregation approach suffers from the dominance of the outlier updates and, consequently, causes high communication costs towards producing a decent global model. In this letter, we propose a robust aggregation strategy to alleviate the above issues. In particular, we propose first pruning the node-wise outlier updates (weights) from the local trained models and then performing the aggregation on the selected effective weights-set at each node. We provide the theoretical result of our method and conduct extensive experiments on the MNIST, CIFAR-10, and Shakespeare datasets with IID and non-IID settings, which demonstrate that our aggregation approach outperforms the state-of-the-art methods in terms of communication speedup, test-set performance and training convergence. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
23. PCA-based Feature Reduction for Hyperspectral Remote Sensing Image Classification.
- Author
-
Uddin, Md. Palash, Mamun, Md. Al, and Hossain, Md. Ali
- Subjects
- *
REMOTE sensing , *FEATURE extraction , *FEATURE selection , *PRINCIPAL components analysis , *SUPPORT vector machines - Abstract
The hyperspectral remote sensing images (HSIs) are acquired to encompass the essential information of land objects through contiguous narrow spectral wavelength bands. The classification accuracy is not often satisfactory in a cost-effective way using the entire original HSI for practical applications. To enhance the classification result of HSIs the band reduction strategies are applied which can be divided into feature extraction and feature selection methods. PCA (Principal Component Analysis), a linear unsupervised statistical transformation, is frequently adopted for the extraction of features from HSIs. In this paper, PCA and SPCA (Segmented-PCA), SSPCA (Spectrally Segmented-PCA), FPCA (Folded-PCA) and MNF (Minimum Noise Fraction) as linear variants of PCA together with KPCA (Kernel-PCA) and KECA (kernel Entropy Component Analysis) as nonlinear variants of PCA have been investigated. The top transformed features were picked out using accumulation of variance for all other feature extraction methods except for MNF and KECA. MNF uses SNR (Signal-to-Noise Ratio) values and KECA employs Renyi quadratic entropy measurement for this purpose. The studied approaches are equated and analyzed for Indian Pine agricultural and urban Washington DC Mall HSI classification using SVM (Support Vector Machine) classifier. The experiment illustrates that the costly effective and improved classification performance of the feature extraction approaches over the performance using the entire original dataset. MNF offers the highest classification accuracy and FPCA offers the least space and time complexity with satisfactory classification result. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
24. Mutual Information Driven Federated Learning.
- Author
-
Uddin, Md Palash, Xiang, Yong, Lu, Xuequan, Yearwood, John, and Gao, Longxiang
- Subjects
- *
INFORMATION theory - Abstract
Federated Learning (FL) is an emerging research field that yields a global trained model from different local clients without violating data privacy. Existing FL techniques often ignore the effective distinction between local models and the aggregated global model when doing the client-side weight update, as well as the distinction of local models for the server-side aggregation. In this article, we propose a novel FL approach with resorting to mutual information (MI). Specifically, in client-side, the weight update is reformulated through minimizing the MI between local and aggregated models and employing Negative Correlation Learning (NCL) strategy. In server-side, we select top effective models for aggregation based on the MI between an individual local model and its previous aggregated model. We also theoretically prove the convergence of our algorithm. Experiments conducted on MNIST, CIFAR-10, ImageNet, and the clinical MIMIC-III datasets manifest that our method outperforms the state-of-the-art techniques in terms of both communication and testing performance. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
25. Information-theoretic feature selection with segmentation-based folded principal component analysis (PCA) for hyperspectral image classification.
- Author
-
Uddin, Md. Palash, Mamun, Md. Al, Afjal, Masud Ibn, and Hossain, Md. Ali
- Subjects
- *
PRINCIPAL components analysis , *FEATURE selection , *SHOPPING mall design & construction , *FEATURE extraction , *THEMATIC maps , *LAND cover - Abstract
Hyperspectral image (HSI) usually holds information of land cover classes as a set of many contiguous narrow spectral wavelength bands. For its efficient thematic mapping or classification, band (feature) reduction strategies through Feature Extraction (FE) and/or Feature Selection (FS) methods for finding the intrinsic bands' information are typically applied. Principal Component Analysis (PCA) is a frequently employed unsupervised linear FE method whereas cumulative-variance accumulation is used for selecting top features from PCA data. However, PCA can fail to extract intrinsic structure of HSI due to global variance consideration and domination by visible and near infrared bands while cumulative-variance accumulation has no capability to exploit non-linear relationships among the transformed features produced by PCA-based FE methods. Consequently, we propose an information theoretic normalized Mutual Information (nMI)-based minimum Redundancy Maximum Relevance (mRMR) non-linear measure to select the intrinsic features from the transformed space of our previously proposed Segmented-Folded-PCA (Seg-Fol-PCA) and Spectrally Segmented-Folded-PCA (SSeg-Fol-PCA) FE methods. We extensively analyse the effectiveness of the proposed unsupervised FE and supervised FS combinations Seg-Fol-PCA-mRMR and SSeg-Fol-PCA-mRMR with that of PCA-based existing linear and non-linear state-of-the-art methods. In addition, cumulative variance-based top features pick-up strategy is considered with all FE methods and Renyi quadratic entropy-based FS is used with Kernel Entropy Component Analysis (Ker-ECA). The experimental results illustrate that SSeg-Fol-PCA-mRMR and Seg-Fol-PCA-mRMR obtain highest classification result e.g. 95.39% and 95.03% respectively for agricultural Indian Pines HSI, and 96.58% and 95.30% respectively for urban Washington DC Mall HSI while the classification accuracies using all original features of the HSIs are 70.28% and 91.90% respectively. Moreover, the proposed SSeg-Fol-PCA-mRMR and Seg-Fol-PCA-mRMR outperform all investigated combinations of FE and FS using the real HSI datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
26. Effective subspace detection based on the measurement of both the spectral and spatial information for hyperspectral image classification.
- Author
-
Mishu, Sadia Zaman, Ahmed, Boshir, Hossain, Md. Ali, and Uddin, Md. Palash
- Subjects
HYPERSPECTRAL imaging systems ,PRINCIPAL components analysis ,SUPPORT vector machines ,FEATURE selection ,REMOTE sensing ,FEATURE extraction - Abstract
Subspace detection from high dimensional hyperspectral image (HSI) data cube has become an important area of research for efficient identification of ground objects. Standard feature extraction method such as Principal Component Analysis (PCA) has some drawbacks as it depends solely on global variance of the dataset generated. Folded-PCA (FPCA), an improvement of PCA, offers more benefits over PCA as it envisages both local and global structures of image contents and requires less computation and memory. These superior properties make FPCA more effective for feature extraction in high dimensional remote sensing images e.g. HSIs. Therefore, the proposed feature reduction method combines FPCA feature extraction with Normalized Cross Cumulative Residual Entropy (NCCRE) feature selection, termed as FPCA-NCCRE, for efficient features' subspace detection. NCCRE is utilised as a means of feature selection over the new features generated from FPCA to obtain a more informative subspace. It is experimented on a real mixed agricultural and an urban hyperspectral dataset. Finally, Kernel Support Vector Machine (KSVM) is implemented to calculate the classification accuracy using the detected subspace. From the experiments, it is observed that the proposed method outperforms the baseline approaches and obtains the highest accuracy of 97.67 and 98.57% on the two real hyperspectral images. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
27. Effective feature extraction through segmentation-based folded-PCA for hyperspectral image classification.
- Author
-
Uddin, Md. Palash, Mamun, Md. Al, and Hossain, Md. Ali
- Subjects
- *
FEATURE extraction , *IMAGE segmentation , *LAND cover , *FEATURE selection , *PRINCIPAL components analysis - Abstract
The remote sensing hyperspectral images (HSIs) usually comprise many important information of the land covers capturing through a set of hundreds of narrow and contiguous spectral wavelength bands. Appropriate classification performance can only offer the required knowledge from these immense bands of HSI since the classification result is not reasonable using all the original features (bands) of the HSI. Although it is not easy to calculate the intrinsic features from the bands, band (dimensionality) reduction techniques through feature extraction and feature selection are usually applied to increase the classification result and to fix the curse of dimensionality problem. Though the Principal Component Analysis (PCA) has been commonly adopted for the feature reduction of HSI, it can often fail to extract the local useful characteristics of the HSI for effective classification as it considers the global statistics of the HSI. Consequently, Segmented-PCA (SPCA), Spectrally-Segmented-PCA (SSPCA), Folded-PCA (FPCA) and Superpixelwise PCA (SuperPCA) have been introduced for better feature extraction of HSI in diverse ways. In this paper, feature extraction through SPCA & FPCA and SSPCA & FPCA, termed as Segmented-FPCA (SFPCA) and Spectrally-Segmented-FPCA (SSFPCA) respectively, has further been improved through applying FPCA on the highly correlated or spectrally separated bands' segments of the HSI rather than not applying the FPCA on the entire dataset directly. The proposed methods are compared and analysed for a real mixed agricultural and an urban HSI classification using per-pixel SVM classifier. The experimental result shows that the classification performance using SSFPCA and SFPCA outperforms that of using conventional PCA, SPCA, SSPCA, FPCA, SuperPCA and using the entire original dataset without employing any feature reduction. Moreover, the proposed feature extraction methods provide the least memory and computation cost complexity. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
28. A stroke prediction framework using explainable ensemble learning.
- Author
-
Mitu, Mostarina, Hasan, S. M. Mahedy, Uddin, Md Palash, Mamun, Md Al, Rajinikanth, Venkatesan, and Kadry, Seifedine
- Abstract
AbstractThe death of brain cells occurs when blood flow to a particular area of the brain is abruptly cut off, resulting in a stroke. Early recognition of stroke symptoms is essential to prevent strokes and promote a healthy lifestyle. FAST tests (looking for abnormalities in the face, arms, and speech) have limitations in reliability and accuracy for diagnosing strokes. This research employs machine learning (ML) techniques to develop and assess multiple ML models to establish a robust stroke risk prediction framework. This research uses a stacking-based ensemble method to select the best three machine learning (ML) models and combine their collective intelligence. An empirical evaluation of a publicly available stroke prediction dataset demonstrates the superior performance of the proposed stacking-based ensemble model, with only one misclassification. The experimental results reveal that the proposed stacking model surpasses other state-of-the-art research, achieving accuracy, precision, F1-score of 99.99%, recall of 100%, receiver operating characteristics (ROC), Mathews correlation coefficient (MCC), and Kappa scores 1.0. Furthermore, Shapley’s Additive Explanations (SHAP) are employed to analyze the predictions of the black-box machine learning (ML) models. The findings highlight that age, BMI, and glucose level are the most significant risk factors for stroke prediction. These findings contribute to the development of more efficient techniques for stroke prediction, potentially saving many lives. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. Segmentation-based truncated-SVD for effective feature extraction in hyperspectral image classification.
- Author
-
Rahman, Md Moshiur, Islam, Md Rashedul, Afjal, Masud Ibn, Marjan, Md Abu, Uddin, Md Palash, and Islam, Md Mominul
- Subjects
- *
SINGULAR value decomposition , *IMAGE recognition (Computer vision) , *FEATURE selection , *PRINCIPAL components analysis , *SUPPORT vector machines - Abstract
Remote sensing hyperspectral images (HSIs) are rich sources of information about land cover captured across hundreds of narrow, contiguous spectral wavelength bands. However, using the entire original HSI for practical applications can lead to suboptimal classification accuracy. To address this, band reduction techniques, categorized as feature extraction and feature selection methods, are employed to enhance classification results. One commonly used feature extraction approach for HSIs is Principal Component Analysis (PCA). However, PCA may fall short of capturing the local and specific characteristics present in the HSI data. In this paper, we introduce two novel feature extraction methods: Segmented Truncated Singular Value Decomposition (STSVD) and Spectrally Segmented Truncated Singular Value Decomposition (SSTSVD) to improve classification performance. Segmentation is carried out based on highly correlated bands’ segments and spectral bands’ segments within the HSI data. Our study evaluates and compares these newly proposed methods against classical feature extraction methods, including PCA, Incremental PCA, Sparse-PCA, Kernel PCA, Segmented-PCA (SPCA), and Truncated Singular Value Decomposition (TSVD). We perform this analysis on three distinct HSI datasets, namely the Indian Pines HSI, the Pavia University HSI, and the Kennedy Space Center HSI, using per-pixel Support Vector Machine (SVM) and Random Forest (RF) classification. The experimental results demonstrate the superiority of our proposed methods for all three datasets. The best-performing feature extraction methods when classification is performed using an SVM classifier are STSVD3 (89.03%), SSTSVD2 (95.55%), and STSVD3 (97.74%) for the Indian Pines, Pavia University, and Kennedy Space Center datasets, respectively. Similarly, for the RF classifier, the best-performing feature extraction methods are SSTSVD4 (88.98%), SSTSVD3 (96.04%), and SSTSVD4 (96.09%) for Indian Pines, Pavia University, and Kennedy Space Center datasets, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. Ensemble machine learning-based recommendation system for effective prediction of suitable agricultural crop cultivation.
- Author
-
Hasan M, Marjan MA, Uddin MP, Afjal M, Kardy S, Ma S, and Nam Y
- Abstract
Agriculture is the most critical sector for food supply on the earth, and it is also responsible for supplying raw materials for other industrial productions. Currently, the growth in agricultural production is not sufficient to keep up with the growing population, which may result in a food shortfall for the world's inhabitants. As a result, increasing food production is crucial for developing nations with limited land and resources. It is essential to select a suitable crop for a specific region to increase its production rate. Effective crop production forecasting in that area based on historical data, including environmental and cultivation areas, and crop production amount, is required. However, the data for such forecasting are not publicly available. As such, in this paper, we take a case study of a developing country, Bangladesh, whose economy relies on agriculture. We first gather and preprocess the data from the relevant research institutions of Bangladesh and then propose an ensemble machine learning approach, called K-nearest Neighbor Random Forest Ridge Regression (KRR), to effectively predict the production of the major crops (three different kinds of rice, potato, and wheat). KRR is designed after investigating five existing traditional machine learning (Support Vector Regression, Naïve Bayes, and Ridge Regression) and ensemble learning (Random Forest and CatBoost) algorithms. We consider four classical evaluation metrics, i.e., mean absolute error, mean square error (MSE), root MSE, and R
2 , to evaluate the performance of the proposed KRR over the other machine learning models. It shows 0.009 MSE, 99% R2 for Aus; 0.92 MSE, 90% R2 for Aman; 0.246 MSE, 99% R2 for Boro; 0.062 MSE, 99% R2 for wheat; and 0.016 MSE, 99% R2 for potato production prediction. The Diebold-Mariano test is conducted to check the robustness of the proposed ensemble model, KRR. In most cases, it shows 1% and 5% significance compared to the benchmark ML models. Lastly, we design a recommender system that suggests suitable crops for a specific land area for cultivation in the next season. We believe that the proposed paradigm will help the farmers and personnel in the agricultural sector leverage proper crop cultivation and production., Competing Interests: The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest., (Copyright © 2023 Hasan, Marjan, Uddin, Afjal, Kardy, Ma and Nam.)- Published
- 2023
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.