20 results on '"deep generative adversarial network"'
Search Results
2. Reinforced black widow algorithm with restoration technique based on optimized deep generative adversarial network.
- Author
-
Praveen Kumar, K., Venkata Narasimhulu, C., and Satya Prasad, K.
- Subjects
- *
GENERATIVE adversarial networks , *IMAGE reconstruction , *GRAYSCALE model , *IMAGING systems , *SMOOTHNESS of functions - Abstract
Image restoration is used to develop the quality of image that is triggered by various noises and blurring. During this causes, certain areas in the images are vanished. The existing works does not provide sufficient restoration process with high accuracy. Therefore, a new image restoration system based on Optimized Deep Generative Adversarial Network (DGAN) with Reinforced Black Widow algorithm (BWOA) is proposed in this paper to increase the restoration accuracy and reducing the noises. At first, the input image is converted as gray scale image and the multi-scale edge information is removed as damaged area of an image by constructing a smooth function. Here, the extracted multi-scale edge information is given to the DGAN model. After that, the images are trained to create the best fake images through continuous play among generator and discriminator. Then, the detected images are restored in the original image with high accuracy. The hyper parameters of the DGAN are optimized by using the BWOA. The major objective of this paper is 'to increase the restoration accuracy and the quality of the image by decreasing the noises occurred in the input image.' The simulation process is performed on the MATLAB platform. The proposed DGAN-BWOA-IR attains higher restoration accuracy of 9.3%, higher PSNR value 74.589(db), SSIM 9.023% the proposed system is likened with the existing approaches, such as plug-and-play image restoration including deep denoiser prior (DCNN-IR), Learning enriched features for rapid image restoration with enhancement (LEF-IR), Exploiting deep generative prior for versatile image restoration with manipulation (GAN-IR), respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Augmented Data-Driven Approach towards 3D Printed Concrete Mix Prediction.
- Author
-
Rehman, Saif Ur, Riaz, Raja Dilawar, Usman, Muhammad, and Kim, In-Ho
- Subjects
ARTIFICIAL neural networks ,MACHINE learning ,GENERATIVE adversarial networks ,DATA augmentation ,COMPRESSIVE strength - Abstract
Formulating a mix design for 3D concrete printing (3DCP) is challenging, as it involves an iterative approach, wasting a lot of resources, time, and effort to optimize the mix for strength and printability. A potential solution is mix formulation through artificial intelligence (AI); however, being a new and emerging field, the open-source availability of datasets is limited. Limited datasets significantly restrict the predictive performance of machine learning (ML) models. This research explores data augmentation techniques like deep generative adversarial network (DGAN) and bootstrap resampling (BR) to increase the available data to train three ML models, namely support vector machine (SVM), artificial neural network (ANN), and extreme gradient boosting regression (XGBoost). Their performance was evaluated using R
2 , MSE, RMSE, and MAE metrics. Models trained on BR-augmented data showed higher accuracy than those trained on the DGAN-augmented data. The BR-trained XGBoost exhibited the highest R2 scores of 0.982, 0.970, 0.972, 0.971, and 0.980 for cast compressive strength, printed compressive strength direction 1, 2, 3, and slump flow respectively. The proposed method of predicting the slump flow (mm), cast, and anisotropic compressive strength (MPa) can effectively predict the mix design for printable concrete, unlocking its full potential for application in the construction industry. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
4. A deep learning approach for prediction of air quality index in smart city
- Author
-
Adel Binbusayyis, Muhammad Attique Khan, Mohamed Mustaq Ahmed A, and W. R. Sam Emmanuel
- Subjects
Air quality index ,Pollutants ,Regression ,Deep generative adversarial network ,Modified stacked attention GRU ,Chennai ,Environmental sciences ,GE1-350 - Abstract
Abstract Industrial developments and consumption of massive amount of fossil fuels, vehicle pollution, and other calamities upsurges the AQI (Air Quality Index) of major cities in a drastic manner. Owing to these factors, it is important to take proactive measures for reducing the air pollution in order to avoid life- threatening consequence. Therefore, prediction of air quality is significant for improving the health of living beings as highly polluted regions have a higher concentration of pollutants mixed in the air, affecting the respiratory system and reducing the lifetime. To control pollution, AQI is used as a measure for estimating the pollutant content in the air. Even though many existing techniques have predicted AQI, enhancement is required in prediction algorithms with minimized loss. To address the challenges in traditional algorithms, the proposed smart cities-based AQI prediction intends to utilize the proposed regression algorithm in the dataset, namely Air- Quality-Data, which collected harmful pollutants on an hourly and daily basis from multiple cities in India between 2015 to 2020. To achieve prediction efficiency with reduced loss, pre-processing of input data is being performed using Deep GAN (Generative Adversarial Network). It performs the imputation of data in place of missing values to improve accurate prediction. Additionally, feature scaling normalizes independent real-data features to a fixed scale. With the processed data, regression is done through modified Stacked Attention GRU with KL divergence, which predicts Ernakulam, Chennai and Ahmedabad cities with higher, medium, and low levels of AQI in India. The performance of the proposed regression algorithm is measured using metrics such as MAE (Mean Absolute Error), MSE (Mean Square Error), R2 (Coefficient of determination), MAPE (Mean Absolute Percentage Error), and RMSE (Root Mean Square Error) and better MAE, MSE, R2, MAPE and RMSE obtained by the model is 0.1013, 0.0134, 0.9479, 0.1152 and 0.1156. Internal assessment and comparative analysis performed with existing regression algorithms exhibit lower loss values obtained from the present research, which determines the efficacy of the proposed model.
- Published
- 2024
- Full Text
- View/download PDF
5. A deep learning approach for prediction of air quality index in smart city.
- Author
-
Binbusayyis, Adel, Khan, Muhammad Attique, Ahmed A, Mohamed Mustaq, and Emmanuel, W. R. Sam
- Subjects
AIR quality indexes ,DEEP learning ,SMART cities ,GENERATIVE adversarial networks ,STANDARD deviations - Abstract
Industrial developments and consumption of massive amount of fossil fuels, vehicle pollution, and other calamities upsurges the AQI (Air Quality Index) of major cities in a drastic manner. Owing to these factors, it is important to take proactive measures for reducing the air pollution in order to avoid life- threatening consequence. Therefore, prediction of air quality is significant for improving the health of living beings as highly polluted regions have a higher concentration of pollutants mixed in the air, affecting the respiratory system and reducing the lifetime. To control pollution, AQI is used as a measure for estimating the pollutant content in the air. Even though many existing techniques have predicted AQI, enhancement is required in prediction algorithms with minimized loss. To address the challenges in traditional algorithms, the proposed smart cities-based AQI prediction intends to utilize the proposed regression algorithm in the dataset, namely Air- Quality-Data, which collected harmful pollutants on an hourly and daily basis from multiple cities in India between 2015 to 2020. To achieve prediction efficiency with reduced loss, pre-processing of input data is being performed using Deep GAN (Generative Adversarial Network). It performs the imputation of data in place of missing values to improve accurate prediction. Additionally, feature scaling normalizes independent real-data features to a fixed scale. With the processed data, regression is done through modified Stacked Attention GRU with KL divergence, which predicts Ernakulam, Chennai and Ahmedabad cities with higher, medium, and low levels of AQI in India. The performance of the proposed regression algorithm is measured using metrics such as MAE (Mean Absolute Error), MSE (Mean Square Error), R2 (Coefficient of determination), MAPE (Mean Absolute Percentage Error), and RMSE (Root Mean Square Error) and better MAE, MSE, R2, MAPE and RMSE obtained by the model is 0.1013, 0.0134, 0.9479, 0.1152 and 0.1156. Internal assessment and comparative analysis performed with existing regression algorithms exhibit lower loss values obtained from the present research, which determines the efficacy of the proposed model. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Augmented Data-Driven Approach towards 3D Printed Concrete Mix Prediction
- Author
-
Saif Ur Rehman, Raja Dilawar Riaz, Muhammad Usman, and In-Ho Kim
- Subjects
3D concrete printing ,compressive strength ,slump flow ,deep generative adversarial network ,bootstrap resampling ,mix design ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
Formulating a mix design for 3D concrete printing (3DCP) is challenging, as it involves an iterative approach, wasting a lot of resources, time, and effort to optimize the mix for strength and printability. A potential solution is mix formulation through artificial intelligence (AI); however, being a new and emerging field, the open-source availability of datasets is limited. Limited datasets significantly restrict the predictive performance of machine learning (ML) models. This research explores data augmentation techniques like deep generative adversarial network (DGAN) and bootstrap resampling (BR) to increase the available data to train three ML models, namely support vector machine (SVM), artificial neural network (ANN), and extreme gradient boosting regression (XGBoost). Their performance was evaluated using R2, MSE, RMSE, and MAE metrics. Models trained on BR-augmented data showed higher accuracy than those trained on the DGAN-augmented data. The BR-trained XGBoost exhibited the highest R2 scores of 0.982, 0.970, 0.972, 0.971, and 0.980 for cast compressive strength, printed compressive strength direction 1, 2, 3, and slump flow respectively. The proposed method of predicting the slump flow (mm), cast, and anisotropic compressive strength (MPa) can effectively predict the mix design for printable concrete, unlocking its full potential for application in the construction industry.
- Published
- 2024
- Full Text
- View/download PDF
7. Improved deep generative adversarial network with illuminant invariant local binary pattern features for facial expression recognition.
- Author
-
Gavade, Priyanka A., Bhat, Vandana S., and Pujari, Jagadeesh
- Subjects
GENERATIVE adversarial networks ,FACIAL expression ,FEATURE extraction ,VIDEO compression - Abstract
Facial Expression Recognition (FER) is one of the prevailing as well as demanding tasks in social communication. In general, face expressions are usual and straight ways for individuals to converse their intentions as well as emotions. This paper proposes the Taylor-Chicken Swarm Optimization-based Deep Generative Adversarial Network (Taylor-CSO-based Deep GAN) for FER. Here, the facial expressions are predicted through a series of steps, such as video frame extraction, pre-processing, face detection, feature extraction, as well as FER. The major contribution of the proposed work lies in the last step, where Taylor-CSO based Deep GAN is employed for recognizing facial expressions. Initially, video frames are extracted from the input video as well as pre-processing is done to extract the Region of Interest (RoI). Then, the Viola Jones algorithm is employed to detect the face images, and Illuminant Invariant Local Binary Pattern (IILBP) features are extracted. Finally, the FER is performed by the proposed model. The performance of the developed model is analyzed using the CK+ dataset, RAVDESS dataset, and SAVEE dataset. Also, the developed model's performance is evaluated with conventional FER methods. The performance analysis exhibits that the adopted model is precise and valuable. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. Landslide identification using water cycle particle swarm optimization‐based deep generative adversarial network.
- Author
-
Lekshmanan, Lijesh and George, Arockia Selva Saroja
- Subjects
LANDSLIDES ,GENERATIVE adversarial networks ,HYDROLOGIC cycle ,PARTICLE swarm optimization ,WATER use ,DATA augmentation - Abstract
Summary: A natural vulnerability called a landslide threatens people, property, and infrastructure in many different places all over the world. Various landslide identification techniques are developed to assess the landslide hazard, but accurate identification of landslide occurrence particularly in certain regions remains a challenging issue, due to the complex scale‐dependent processes among geomorphometric features and landslides. To solve the shortcomings of landslide identification, the water cycle particle swarm optimization‐based deep generative adversarial network (WCPSO‐based Deep GAN) hybridization technique is presented. To develop the developed WCPSO, the water cycle algorithm (WCA) and particle swarm optimization (PSO) were hybridized. The landslide rainfall dataset is used as the source of the input data in this instance, and the Manhattan distance approach is then used to do feature selection. After performing data augmentation based on the results of the feature selection, the proposed WCPSO‐based deep GAN approach accurately detects the landslides. The developed WCPSO‐driven deep GAN's performance is also measured using the three metrics accuracy, specificity, and sensitivity, with a maximum accuracy of 0.937, higher specificity of 0.919, and higher sensitivity of 0.952. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. Automated urban planning aware spatial hierarchies and human instructions.
- Author
-
Wang, Dongjie, Liu, Kunpeng, Huang, Yanyong, Sun, Leilei, Du, Bowen, and Fu, Yanjie
- Subjects
URBAN planning ,GENERATIVE adversarial networks ,AUTOMATED planning & scheduling ,ZONING ,DEEP learning - Abstract
Traditional urban planning demands urban experts to spend much time producing an optimal urban plan under many architectural constraints. The remarkable imaginative ability of deep generative learning provides hope for renovating this domain. Existing works are constrained by: (1) neglecting human requirements; (2) omitting spatial hierarchies, and (3) lacking urban plan samples. We propose a novel, deep human-instructed urban planner to fill these gaps and implement two practical frameworks. In the preliminary version, we formulate the task into an encoder–decoder paradigm. The encoder is to learn the information distribution of surrounding contexts, human instructions, and land-use configuration. The decoder is to reconstruct the land-use configuration and the associated urban functional zones. Although it has achieved good results, the generation performance is still unstable due to the complex optimization directions of the decoder. Thus, we propose a cascading deep generative adversarial network (GAN) in this paper, inspired by the workflow of urban experts. The first GAN is to build urban functional zones based on human instructions and surrounding contexts. The second GAN will produce the land-use configuration by considering the built urban functional zones. Finally, we conducted extensive experiments and case studies to validate the effectiveness and superiority of our work. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
10. Motion Blur Detection Using Deep Generative Adversarial Network Method
- Author
-
Ranjini, Kurian, M. Z., Chidanandamurthy, M. V., Ranjini, Kurian, M. Z., and Chidanandamurthy, M. V.
- Abstract
This paper presents a novel approach for blur detection and removal using Generative Adversarial Networks (GANs). The proposed method leverages the power of deep learning to automatically identify and eliminate blur in digital images. The first phase of the process involves training a GAN model on a dataset of paired images, where one image is sharp and the other is intentionally blurred. The GAN consists of a generator network that aims to generate sharp images from their blurred counterparts, and a discriminator network that distinguishes between real and generated sharp images. During the training phase, the generator network learns the mapping from blurred images to sharp images, while the discriminator network improves its ability to differentiate between real and generated sharp images. This adversarial training process helps the GAN model improve its performance in detecting and removing blur from images. In the testing phase, the trained GAN model can be used to enhance images by detecting and effectively removing blur. Experimental results demonstrate the effectiveness of the proposed approach in achieving high-quality image restoration and enhancement. Overall, the proposed blur detection and removal technique and removal technique using GANs showcases the potential of deep learning in addressing challenging image processing tasks and contributes to the ongoing advancements in the field of computer vision.
- Published
- 2024
11. Image DeHazing Using Deep Learning Techniques Ravi Raj Choudhary.
- Author
-
Jisnu, K K and Meena, Gaurav
- Subjects
COMPUTER vision ,VISUAL fields ,BRITISH occupation of India, 1765-1947 ,HAZE ,COST functions ,WEATHER ,DEEP learning - Abstract
The task of image de-hazing has been a challenge in the field of Computer Vision since its inception. The images captured during adverse weather conditions often appear to be of low quality due to the presence of various atmospheric particles, which results in the haze, fog etc. This, in turn, causes trouble in detecting objects in an image. This causes problems for many computer vision problems that rely on the visibility of these images. In this paper we are implementing a deep Generative Adversarial Network for image de-hazing. The original de-hazing approaches use per-pixel loss, which when calculating creates huge differences even when there is the only difference in a single pixel, even if these images are perceptually similar. The perceptual loss function extracts high level features of images using pre-trained models on ImageNet, which removes the problems of per-pixel loss functions. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
12. Restoration of Missing Patterns on Satellite Infrared Sea Surface Temperature Images Due to Cloud Coverage Using Deep Generative Inpainting Network
- Author
-
Song-Hee Kang, Youngjin Choi, and Jae Young Choi
- Subjects
sea surface temperature ,daily-averaged SST ,NOAA-19 infrared data ,loss of sea surface temperature ,restoration ,deep generative adversarial network ,Naval architecture. Shipbuilding. Marine engineering ,VM1-989 ,Oceanography ,GC1-1581 - Abstract
In this paper, we propose a novel deep generative inpainting network (GIN) trained under the framework of generative adversarial learning, which is optimized for the restoration of cloud-disturbed satellite sea surface temperature (SST) imagery. The proposed GIN architecture can achieve accurate and fast restoration results. The proposed GIN consists of rough and fine reconstruction stages to promote the details and textures of missing (clouded) regions in SST images. We also propose a nov el preprocessing strategy that replaces the land areas with the average value of daily oceanic surface temperatures for improving restoration accuracy. To learn the proposed GIN, we developed a novel approach that combines multiple loss functions well suited for improving the restoration quality over missing SST information. Our results show that the difference in temperature between restored and actual satellite image data was no larger than 0.7 °C in monthly average values, which suggests excellent resilience against the missing sea surface temperature data. The proposed GIN has a faster restoration time and is feasible for real-time ocean-related applications. Furthermore, the computational cost of restoring SST images is much lower than the popular interpolation methods.
- Published
- 2021
- Full Text
- View/download PDF
13. SLSNet: Skin lesion segmentation using a lightweight generative adversarial network
- Author
-
Universitat Rovira i Virgili, Sarker, Md Mostafa Kamal; Rashwan, Hatem A.; Akram, Farhan; Singh, Vivek Kumar; Banu, Syeda Furruka; Chowdhury, Forhad U. H.; Choudhury, Kabir Ahmed; Chambon, Sylvie; Radeva, Petia; Puig, Domenec; Abdel-Nasser, Mohamed, Universitat Rovira i Virgili, and Sarker, Md Mostafa Kamal; Rashwan, Hatem A.; Akram, Farhan; Singh, Vivek Kumar; Banu, Syeda Furruka; Chowdhury, Forhad U. H.; Choudhury, Kabir Ahmed; Chambon, Sylvie; Radeva, Petia; Puig, Domenec; Abdel-Nasser, Mohamed
- Abstract
The determination of precise skin lesion boundaries in dermoscopic images using automated methods faces many challenges, most importantly, the presence of hair, inconspicuous lesion edges and low contrast in dermoscopic images, and variability in the color, texture and shapes of skin lesions. Existing deep learning-based skin lesion segmentation algorithms are expensive in terms of computational time and memory. Consequently, running such segmentation algorithms requires a powerful GPU and high bandwidth memory, which are not available in dermoscopy devices. Thus, this article aims to achieve precise skin lesion segmentation with minimum resources: a lightweight, efficient generative adversarial network (GAN) model called SLSNet, which combines 1-D kernel factorized networks, position and channel attention, and multiscale aggregation mechanisms with a GAN model. The 1-D kernel factorized network reduces the computational cost of 2D filtering. The position and channel attention modules enhance the discriminative ability between the lesion and non-lesion feature representations in spatial and channel dimensions, respectively. A multiscale block is also used to aggregate the coarse-to-fine features of input skin images and reduce the effect of the artifacts. SLSNet is evaluated on two publicly available datasets: ISBI 2017 and the ISIC 2018. Although SLSNet has only 2.35 million parameters, the experimental results demonstrate that it achieves segmentation results on a par with the state-of-the-art skin lesion segmentation methods with an accuracy of 97.61%, and Dice and Jaccard similarity coefficients of 90.63% and 81.98%, respectively. SLSNet can run at more than 110 frames per second (FPS) in a single GTX1080Ti GPU, which is faster than well-known deep learning-based im
- Published
- 2021
14. Spatio-temporal silhouette sequence reconstruction for gait recognition against occlusion
- Author
-
Uddin, Md. Zasim, Muramatsu, Daigo, Takemura, Noriko, Ahad, Md. Atiqur Rahman, and Yagi, Yasushi
- Published
- 2019
- Full Text
- View/download PDF
15. SLSNet: Skin lesion segmentation using a lightweight generative adversarial network.
- Author
-
Sarker, Md. Mostafa Kamal, Rashwan, Hatem A., Akram, Farhan, Singh, Vivek Kumar, Banu, Syeda Furruka, Chowdhury, Forhad U.H., Choudhury, Kabir Ahmed, Chambon, Sylvie, Radeva, Petia, Puig, Domenec, and Abdel-Nasser, Mohamed
- Subjects
- *
GENERATIVE adversarial networks , *DEEP learning , *FEATURE extraction , *SKIN imaging , *ATTENTION control , *IMAGE segmentation - Abstract
The determination of precise skin lesion boundaries in dermoscopic images using automated methods faces many challenges, most importantly, the presence of hair, inconspicuous lesion edges and low contrast in dermoscopic images, and variability in the color, texture and shapes of skin lesions. Existing deep learning-based skin lesion segmentation algorithms are expensive in terms of computational time and memory. Consequently, running such segmentation algorithms requires a powerful GPU and high bandwidth memory, which are not available in dermoscopy devices. Thus, this article aims to achieve precise skin lesion segmentation with minimum resources: a lightweight, efficient generative adversarial network (GAN) model called SLSNet, which combines 1-D kernel factorized networks, position and channel attention, and multiscale aggregation mechanisms with a GAN model. The 1-D kernel factorized network reduces the computational cost of 2D filtering. The position and channel attention modules enhance the discriminative ability between the lesion and non-lesion feature representations in spatial and channel dimensions, respectively. A multiscale block is also used to aggregate the coarse-to-fine features of input skin images and reduce the effect of the artifacts. SLSNet is evaluated on two publicly available datasets: ISBI 2017 and the ISIC 2018. Although SLSNet has only 2.35 million parameters, the experimental results demonstrate that it achieves segmentation results on a par with the state-of-the-art skin lesion segmentation methods with an accuracy of 97.61%, and Dice and Jaccard similarity coefficients of 90.63% and 81.98%, respectively. SLSNet can run at more than 110 frames per second (FPS) in a single GTX1080Ti GPU, which is faster than well-known deep learning-based image segmentation models, such as FCN. Therefore, SLSNet can be used for practical dermoscopic applications. • A lightweight and fully automatic skin lesion segmentation model is proposed. • A multiscale mechanism is introduced to extract features at different scales. • The position attention module controls the spatial inter-dependencies. • The channel attention module controls the channel inter-dependencies. • The combination of binary cross-entropy, Jaccard index, and L1 loss is used. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
16. Restoration of Missing Patterns on Satellite Infrared Sea Surface Temperature Images Due to Cloud Coverage Using Deep Generative Inpainting Network.
- Author
-
Kang, Song-Hee, Choi, Youngjin, Choi, Jae Young, and Márquez, Fausto Pedro García
- Subjects
OCEAN temperature ,INPAINTING ,GENERATIVE adversarial networks ,REMOTE-sensing images ,ANTIQUE & classic car restoration ,SURFACE temperature - Abstract
In this paper, we propose a novel deep generative inpainting network (GIN) trained under the framework of generative adversarial learning, which is optimized for the restoration of cloud-disturbed satellite sea surface temperature (SST) imagery. The proposed GIN architecture can achieve accurate and fast restoration results. The proposed GIN consists of rough and fine reconstruction stages to promote the details and textures of missing (clouded) regions in SST images. We also propose a nov el preprocessing strategy that replaces the land areas with the average value of daily oceanic surface temperatures for improving restoration accuracy. To learn the proposed GIN, we developed a novel approach that combines multiple loss functions well suited for improving the restoration quality over missing SST information. Our results show that the difference in temperature between restored and actual satellite image data was no larger than 0.7 °C in monthly average values, which suggests excellent resilience against the missing sea surface temperature data. The proposed GIN has a faster restoration time and is feasible for real-time ocean-related applications. Furthermore, the computational cost of restoring SST images is much lower than the popular interpolation methods. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
17. SLSNet: skin lesion segmentation using a lightweight generative adversarial network.
- Author
-
Sarker, M. M. K. (Md. Mostafa Kamal), Rashwan, H. A. (Hatem A.), Akram, F. (Farhan), Singh, V. K. (Vivek Kumar), Banu, S. F. (Syeda Furruka), Chowdhury, F. U.H. (Forhad U.H.), Choudhury, K. A. (Kabir Ahmed), Chambon, S. (Sylvie), Radeva, P. (Petia), Puig, D. (Domenec), Abdel-Nasser, M. (Mohamed), Sarker, M. M. K. (Md. Mostafa Kamal), Rashwan, H. A. (Hatem A.), Akram, F. (Farhan), Singh, V. K. (Vivek Kumar), Banu, S. F. (Syeda Furruka), Chowdhury, F. U.H. (Forhad U.H.), Choudhury, K. A. (Kabir Ahmed), Chambon, S. (Sylvie), Radeva, P. (Petia), Puig, D. (Domenec), and Abdel-Nasser, M. (Mohamed)
- Abstract
The determination of precise skin lesion boundaries in dermoscopic images using automated methods faces many challenges, most importantly, the presence of hair, inconspicuous lesion edges and low contrast in dermoscopic images, and variability in the color, texture and shapes of skin lesions. Existing deep learning-based skin lesion segmentation algorithms are expensive in terms of computational time and memory. Consequently, running such segmentation algorithms requires a powerful GPU and high bandwidth memory, which are not available in dermoscopy devices. Thus, this article aims to achieve precise skin lesion segmentation with minimum resources: a lightweight, efficient generative adversarial network (GAN) model called SLSNet, which combines 1-D kernel factorized networks, position and channel attention, and multiscale aggregation mechanisms with a GAN model. The 1-D kernel factorized network reduces the computational cost of 2D filtering. The position and channel attention modules enhance the discriminative ability between the lesion and non-lesion feature representations in spatial and channel dimensions, respectively. A multiscale block is also used to aggregate the coarse-to-fine features of input skin images and reduce the effect of the artifacts. SLSNet is evaluated on two publicly available datasets: ISBI 2017 and the ISIC 2018. Although SLSNet has only 2.35 million parameters, the experimental results demonstrate that it achieves segmentation results on a par with the state-of-the-art skin lesion segmentation methods with an accuracy of 97.61%, and Dice and Jaccard similarity coefficients of 90.63% and 81.98%, respectively. SLSNet can run at more than 110 frames per second (FPS) in a single GTX1080Ti GPU, which is faster than well-known deep learning-based image segmentation models, such as FCN. Therefore, SLSNet can be used for practical dermoscopic applications.
18. SLSNet: skin lesion segmentation using a lightweight generative adversarial network.
- Abstract
The determination of precise skin lesion boundaries in dermoscopic images using automated methods faces many challenges, most importantly, the presence of hair, inconspicuous lesion edges and low contrast in dermoscopic images, and variability in the color, texture and shapes of skin lesions. Existing deep learning-based skin lesion segmentation algorithms are expensive in terms of computational time and memory. Consequently, running such segmentation algorithms requires a powerful GPU and high bandwidth memory, which are not available in dermoscopy devices. Thus, this article aims to achieve precise skin lesion segmentation with minimum resources: a lightweight, efficient generative adversarial network (GAN) model called SLSNet, which combines 1-D kernel factorized networks, position and channel attention, and multiscale aggregation mechanisms with a GAN model. The 1-D kernel factorized network reduces the computational cost of 2D filtering. The position and channel attention modules enhance the discriminative ability between the lesion and non-lesion feature representations in spatial and channel dimensions, respectively. A multiscale block is also used to aggregate the coarse-to-fine features of input skin images and reduce the effect of the artifacts. SLSNet is evaluated on two publicly available datasets: ISBI 2017 and the ISIC 2018. Although SLSNet has only 2.35 million parameters, the experimental results demonstrate that it achieves segmentation results on a par with the state-of-the-art skin lesion segmentation methods with an accuracy of 97.61%, and Dice and Jaccard similarity coefficients of 90.63% and 81.98%, respectively. SLSNet can run at more than 110 frames per second (FPS) in a single GTX1080Ti GPU, which is faster than well-known deep learning-based image segmentation models, such as FCN. Therefore, SLSNet can be used for practical dermoscopic applications.
19. SLSNet: skin lesion segmentation using a lightweight generative adversarial network.
- Author
-
Sarker, M. M. K. (Md. Mostafa Kamal), Rashwan, H. A. (Hatem A.), Akram, F. (Farhan), Singh, V. K. (Vivek Kumar), Banu, S. F. (Syeda Furruka), Chowdhury, F. U.H. (Forhad U.H.), Choudhury, K. A. (Kabir Ahmed), Chambon, S. (Sylvie), Radeva, P. (Petia), Puig, D. (Domenec), Abdel-Nasser, M. (Mohamed), Sarker, M. M. K. (Md. Mostafa Kamal), Rashwan, H. A. (Hatem A.), Akram, F. (Farhan), Singh, V. K. (Vivek Kumar), Banu, S. F. (Syeda Furruka), Chowdhury, F. U.H. (Forhad U.H.), Choudhury, K. A. (Kabir Ahmed), Chambon, S. (Sylvie), Radeva, P. (Petia), Puig, D. (Domenec), and Abdel-Nasser, M. (Mohamed)
- Abstract
The determination of precise skin lesion boundaries in dermoscopic images using automated methods faces many challenges, most importantly, the presence of hair, inconspicuous lesion edges and low contrast in dermoscopic images, and variability in the color, texture and shapes of skin lesions. Existing deep learning-based skin lesion segmentation algorithms are expensive in terms of computational time and memory. Consequently, running such segmentation algorithms requires a powerful GPU and high bandwidth memory, which are not available in dermoscopy devices. Thus, this article aims to achieve precise skin lesion segmentation with minimum resources: a lightweight, efficient generative adversarial network (GAN) model called SLSNet, which combines 1-D kernel factorized networks, position and channel attention, and multiscale aggregation mechanisms with a GAN model. The 1-D kernel factorized network reduces the computational cost of 2D filtering. The position and channel attention modules enhance the discriminative ability between the lesion and non-lesion feature representations in spatial and channel dimensions, respectively. A multiscale block is also used to aggregate the coarse-to-fine features of input skin images and reduce the effect of the artifacts. SLSNet is evaluated on two publicly available datasets: ISBI 2017 and the ISIC 2018. Although SLSNet has only 2.35 million parameters, the experimental results demonstrate that it achieves segmentation results on a par with the state-of-the-art skin lesion segmentation methods with an accuracy of 97.61%, and Dice and Jaccard similarity coefficients of 90.63% and 81.98%, respectively. SLSNet can run at more than 110 frames per second (FPS) in a single GTX1080Ti GPU, which is faster than well-known deep learning-based image segmentation models, such as FCN. Therefore, SLSNet can be used for practical dermoscopic applications.
20. SLSNet: skin lesion segmentation using a lightweight generative adversarial network.
- Abstract
The determination of precise skin lesion boundaries in dermoscopic images using automated methods faces many challenges, most importantly, the presence of hair, inconspicuous lesion edges and low contrast in dermoscopic images, and variability in the color, texture and shapes of skin lesions. Existing deep learning-based skin lesion segmentation algorithms are expensive in terms of computational time and memory. Consequently, running such segmentation algorithms requires a powerful GPU and high bandwidth memory, which are not available in dermoscopy devices. Thus, this article aims to achieve precise skin lesion segmentation with minimum resources: a lightweight, efficient generative adversarial network (GAN) model called SLSNet, which combines 1-D kernel factorized networks, position and channel attention, and multiscale aggregation mechanisms with a GAN model. The 1-D kernel factorized network reduces the computational cost of 2D filtering. The position and channel attention modules enhance the discriminative ability between the lesion and non-lesion feature representations in spatial and channel dimensions, respectively. A multiscale block is also used to aggregate the coarse-to-fine features of input skin images and reduce the effect of the artifacts. SLSNet is evaluated on two publicly available datasets: ISBI 2017 and the ISIC 2018. Although SLSNet has only 2.35 million parameters, the experimental results demonstrate that it achieves segmentation results on a par with the state-of-the-art skin lesion segmentation methods with an accuracy of 97.61%, and Dice and Jaccard similarity coefficients of 90.63% and 81.98%, respectively. SLSNet can run at more than 110 frames per second (FPS) in a single GTX1080Ti GPU, which is faster than well-known deep learning-based image segmentation models, such as FCN. Therefore, SLSNet can be used for practical dermoscopic applications.
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.