Comparative Analysis Of Convolutional Neural Network Models For Digital Image-Based Melanoma Classification

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Melanoma is one of the most malignant forms of skin cancer, with an incidence rate of 7.9% in Indonesia. Traditional biopsy-based diagnosis, though crucial, is invasive and time-consuming, creating barriers for early detection. To address this issue, this research compares two Convolutional Neural Network (CNN) models for digital image-based melanoma classification. The study utilized a publicly available dataset from Kaggle, consisting of 17,805 images (melanoma and non-melanoma), which were divided into training, validation, and testing subsets. The models were trained using the Adamax and SGD optimizers for 100 epochs. The performance of the models was evaluated based on accuracy, loss, precision, recall, and F1-score. The CNN model with the best architecture, which consisted of two fully connected layers, achieved an accuracy of 93.18% and a loss of 0.1636, outperforming the alternative model. These results confirm the effectiveness of CNN models in classifying melanoma images and support the development of a web-based platform that allows users to upload or capture images for rapid and non-invasive detection.

Similar Papers
  • Front Matter
  • Cite Count Icon 1
  • 10.1016/j.gie.2020.12.008
Artificial intelligence: finding the intersection of predictive modeling and clinical utility
  • Mar 7, 2021
  • Gastrointestinal Endoscopy
  • Karthik Ravi

Artificial intelligence: finding the intersection of predictive modeling and clinical utility

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 58
  • 10.2196/jmir.9413
Public Perception Analysis of Tweets During the 2015 Measles Outbreak: Comparative Study Using Convolutional Neural Network Models
  • Jul 9, 2018
  • Journal of Medical Internet Research
  • Jingcheng Du + 6 more

BackgroundTimely understanding of public perceptions allows public health agencies to provide up-to-date responses to health crises such as infectious diseases outbreaks. Social media such as Twitter provide an unprecedented way for the prompt assessment of the large-scale public response.ObjectiveThe aims of this study were to develop a scheme for a comprehensive public perception analysis of a measles outbreak based on Twitter data and demonstrate the superiority of the convolutional neural network (CNN) models (compared with conventional machine learning methods) on measles outbreak-related tweets classification tasks with a relatively small and highly unbalanced gold standard training set.MethodsWe first designed a comprehensive scheme for the analysis of public perception of measles based on tweets, including 3 dimensions: discussion themes, emotions expressed, and attitude toward vaccination. All 1,154,156 tweets containing the word “measles” posted between December 1, 2014, and April 30, 2015, were purchased and downloaded from DiscoverText.com. Two expert annotators curated a gold standard of 1151 tweets (approximately 0.1% of all tweets) based on the 3-dimensional scheme. Next, a tweet classification system based on the CNN framework was developed. We compared the performance of the CNN models to those of 4 conventional machine learning models and another neural network model. We also compared the impact of different word embeddings configurations for the CNN models: (1) Stanford GloVe embedding trained on billions of tweets in the general domain, (2) measles-specific embedding trained on our 1 million measles related tweets, and (3) a combination of the 2 embeddings.ResultsCohen kappa intercoder reliability values for the annotation were: 0.78, 0.72, and 0.80 on the 3 dimensions, respectively. Class distributions within the gold standard were highly unbalanced for all dimensions. The CNN models performed better on all classification tasks than k-nearest neighbors, naïve Bayes, support vector machines, or random forest. Detailed comparison between support vector machines and the CNN models showed that the major contributor to the overall superiority of the CNN models is the improvement on recall, especially for classes with low occurrence. The CNN model with the 2 embedding combination led to better performance on discussion themes and emotions expressed (microaveraging F1 scores of 0.7811 and 0.8592, respectively), while the CNN model with Stanford embedding achieved best performance on attitude toward vaccination (microaveraging F1 score of 0.8642).ConclusionsThe proposed scheme can successfully classify the public’s opinions and emotions in multiple dimensions, which would facilitate the timely understanding of public perceptions during the outbreak of an infectious disease. Compared with conventional machine learning methods, our CNN models showed superiority on measles-related tweet classification tasks with a relatively small and highly unbalanced gold standard. With the success of these tasks, our proposed scheme and CNN-based tweets classification system is expected to be useful for the analysis of tweets about other infectious diseases such as influenza and Ebola.

  • Research Article
  • Cite Count Icon 115
  • 10.1016/j.aca.2020.03.055
Understanding the learning mechanism of convolutional neural networks in spectral analysis
  • Apr 8, 2020
  • Analytica Chimica Acta
  • Xiaolei Zhang + 8 more

Understanding the learning mechanism of convolutional neural networks in spectral analysis

  • Conference Article
  • Cite Count Icon 86
  • 10.1109/biosig.2016.7736925
How Transferable Are CNN-Based Features for Age and Gender Classification?
  • Sep 1, 2016
  • Gokhan Ozbulak + 2 more

Age and gender are complementary soft biometric traits for face recognition. Successful estimation of age and gender from facial images taken under real-world conditions can contribute improving the identification results in the wild. In this study, in order to achieve robust age and gender classification in the wild, we have benefited from Deep Convolutional Neural Networks based representation. We have explored transferability of existing deep convolutional neural network (CNN) models for age and gender classification. The generic AlexNet-like architecture and domain specific VGG-Face CNN model are employed and fine-tuned with the Adience dataset prepared for age and gender classification in uncontrolled environments. In addition, task specific GilNet CNN model has also been utilized and used as a baseline method in order to compare with transferred models. Experimental results show that both transferred deep CNN models outperform the GilNet CNN model, which is the state-of-the-art age and gender classification approach on the Adience dataset, by an absolute increase of 7% and 4.5% in accuracy, respectively. This outcome indicates that transferring a deep CNN model can provide better classification performance than a task specific CNN model, which has a limited number of layers and trained from scratch using a limited amount of data as in the case of GilNet. Domain specific VGG-Face CNN model has been found to be more useful and provided better performance for both age and gender classification tasks, when compared with generic AlexNet-like model, which shows that transfering from a closer domain is more useful.

  • Research Article
  • Cite Count Icon 7
  • 10.1016/j.rinp.2021.104385
Application of PET/CT image under convolutional neural network model in postoperative pneumonia virus infection monitoring of patients with non-small cell lung cancer
  • May 27, 2021
  • Results in Physics
  • Jing Wei + 5 more

Application of PET/CT image under convolutional neural network model in postoperative pneumonia virus infection monitoring of patients with non-small cell lung cancer

  • Research Article
  • 10.1097/brs.0000000000005438
Enabling Early Identification of Malignant Vertebral Compression Fractures Through 2.5D Convolutional Neural Network Model With CT Image Analysis.
  • Jun 23, 2025
  • Spine
  • Chengbin Huang + 7 more

This study used a retrospective data analysis approach combined with model development and validation. The present study introduces a 2.5D convolutional neural network (CNN) model leveraging CT imaging to facilitate the early detection of malignant vertebral compression fractures (MVCFs), potentially reducing reliance on invasive biopsies. Vertebral histopathologic biopsy is recognized as the gold standard for differentiating between osteoporotic and malignant vertebral compression fractures (VCFs). Nevertheless, its application is restricted due to its invasive nature and high cost, highlighting the necessity for alternative methods to identify MVCFs. The clinical, imaging, and pathologic data of patients who underwent vertebral augmentation and biopsy at institution 1 and institution 2 were collected and analyzed. On the basis of the vertebral CT images of these patients, 2D, 2.5D, and 3D CNN models were developed to identify the patients with osteoporotic vertebral compression fractures (OVCF) and MVCF. To verify the clinical application value of the CNN model, two rounds of reader studies were performed. The 2.5D CNN model performed well, and its performance in identifying MVCF patients was significantly superior to that of the 2D and 3D CNN models. In the training data set, the area under the receiver operating characteristic curve (AUC) of the 2.5D CNN model was 0.996 and an F1 score of 0.915. In the external cohort test, the AUC was 0.815 and an F1 score of 0.714. And clinicians' ability to identify MVCF patients has been enhanced by the 2.5D CNN model. With the assistance of the 2.5D CNN model, the AUC of senior clinicians was 0.882, and the F1 score was 0.774. For junior clinicians, the 2.5D CNN model-assisted AUC was 0.784 and the F1 score was 0.667. The development of our 2.5D CNN model marks a significant step toward noninvasive identification of MVCF patients. The 2.5D CNN model may be a potential model to assist clinicians in better identifying MVCF patients.

  • Research Article
  • Cite Count Icon 5
  • 10.1080/01431161.2024.2343429
Pre- and post-fire forest canopy height mapping in Southeast Australia through the integration of multi-temporal GEDI data, satellite images, and Convolution Neural Network
  • May 7, 2024
  • International Journal of Remote Sensing
  • Tsung-Chi Chou + 2 more

This study leveraged Convolutional Neural Network (CNN) models to estimate canopy height in Southeast Australian forests before and after the 2019–2020 bushfire event, using inputs from Sentinel-1, Sentinel-2, spectral indices, and terrain features and GEDI canopy height (rh99) as target variables. Our research consisted of three primary objectives: (1) an evaluation of GEDI rh99 canopy height in relation to an nDSM derived from airborne LiDAR, (2) the development of two CNN models for pre- and post-fire scenarios, and (3) utilizing the trained CNN models to generate pre- and post-fire canopy height maps and canopy height change maps across the continuous landscape. GEDI rh99 accuracy analysis revealed R 2 = 0.85, RMSE = 7.25 m and nRMSE = 0.23 for pre-fire GEDI and R 2 = 0.78, RMSE = 10.1 m and nRMSE = 0.3 for post-fire GEDI. The pre-fire CNN model achieved R 2 = 0.75, RMSE = 7.29 m and nRMSE = 0.22, and post-fire CNN model achieved R 2 = 0.62, RMSE = 11.67 m, and nRMSE = 0.34 when validated against nDSM. Bias analysis was conducted, revealing that GEDI rh99 underestimated nDSM across all canopy height ranges, and slope has no significant impact on GEDI accuracy. Notably, post-fire GEDI rh99 exhibited significant underestimation in high to extreme fire severity. The CNN models’ prediction displayed overestimation in short forests < 20 m and increasing underestimation in taller forests > 20 m and underestimation in high and extreme fire severity. Subsequent investigation indicated that canopy layer thinning in high fire severity conditions resulted in weak GEDI waveform signals and consequent bias. Additionally, the GEDI rh99 bias was propagated to the CNN models, together with insensitivity of spectral band values to canopy height in dense and burnt forests potentially contributed to models’ bias. Finally, we demonstrated CNN models’ ability to monitor forest recovery by generating canopy height maps spanning East Gippsland from 2019 to 2023.

  • Research Article
  • Cite Count Icon 31
  • 10.1016/j.atmosres.2022.106037
Near real-time hurricane rainfall forecasting using convolutional neural network models with Integrated Multi-satellitE Retrievals for GPM (IMERG) product
  • Feb 9, 2022
  • Atmospheric Research
  • Taereem Kim + 3 more

Near real-time hurricane rainfall forecasting using convolutional neural network models with Integrated Multi-satellitE Retrievals for GPM (IMERG) product

  • Research Article
  • Cite Count Icon 1
  • 10.1016/j.compbiolchem.2025.108713
A comprehensive comparison of convolutional neural network and visual transformer models on skin cancer classification.
  • Feb 1, 2026
  • Computational biology and chemistry
  • Ibrahim Aruk + 2 more

A comprehensive comparison of convolutional neural network and visual transformer models on skin cancer classification.

  • Conference Article
  • Cite Count Icon 1
  • 10.1115/msec2025-155493
Comparative Performance Analysis of CNN Models for Cracking Defect Detection in 3D-Printed Polymer Parts
  • Jun 23, 2025
  • Vivek V Bhandarkar + 2 more

Additive manufacturing (AM), or 3D printing, has transformed modern manufacturing by enabling the production of complex shapes with high precision. However, AM faces challenges, such as cracking defects, particularly in parts made from thermoplastics like acrylonitrile butadiene styrene (ABS) and polylactic acid (PLA), which are valued for their strength and durability. This study evaluates five popular pre-trained convolutional neural network (CNN) models, namely DenseNet121, MobileNetV2, ResNet50, VGG16, and XceptionNet, for detecting cracking defects in 3D-printed ABS and PLA parts. The models were assessed using key metrics such as accuracy, loss, F1-score, and receiver operating characteristic (ROC) plots to identify the most effective model for real-time defect detection. A delta 3D printer equipped with a Raspberry Pi camera was used to generate a dataset of 3,705 images by printing various geometries. A design of experiments (DoE) approach was employed to capture images at different stages of the printing process, ensuring a diverse dataset. These images were pre-processed to focus on areas of interest, minimizing the computation required for the CNN models. Transfer learning was used to train the models to classify two categories: Cracking and No_Cracking. The dataset was divided into training (64%), validation (16%), and testing (20%) subsets. XceptionNet achieved the highest accuracy at 99.32%, followed by DenseNet121 at 99.05% and VGG16 at 98.92%. MobileNetV2 recorded 98.65% accuracy, while ResNet50 had the lowest at 95.00%. XceptionNet’s superior performance makes it the best option for real-time cracking defect detection, enhancing production efficiency.

  • Research Article
  • 10.1029/2024jd040961
Improving the Predictability of the US Seasonal Surface Temperature With Convolutional Neural Networks Trained on CESM2 LENS
  • Aug 8, 2024
  • Journal of Geophysical Research: Atmospheres
  • Yujay An + 1 more

To better understand and improve the prediction of the seasonal surface temperature (TS) across the United States, we employed convolutional neural network (CNN) models trained on the Community Earth System Model Version 2 Large Ensemble (CESM2 LENS). We used lagged sea surface temperatures (SST) over the tropical Pacific region, containing the information of the El Niño Southern Oscillation (ENSO), as input for the CNN models. ENSO is the principal driver of variability in seasonal US surface temperatures (TSUS) and employing CNN models allows for spatiotemporal aspects of ENSO to be analyzed to make seasonal TSUS predictions. For predicting TSUS, the CNN models exhibited significantly improved skill over standard statistical multilinear regression (MLR) models and dynamical forecasts across most regions in the US, for lead times ranging from 1 to 6 months. Furthermore, we employed the CNN models to predict seasonal TSUS during extreme ENSO events. For these events, the CNN models outperformed the MLR models in predicting the effects on seasonal TSUS, suggesting that the CNN models are able to capture the ENSO‐TSUS teleconnection more effectively. Results from a heatmap analysis demonstrate that the CNN models utilize spatial features of ENSO rather than solely the magnitude of the ENSO events, indicating that the improved skill of seasonal TSUS is due to analyzing spatial variation in ENSO events. The proposed CNN model demonstrates a promising improvement in prediction skill compared to existing methods, suggesting a potential path forward for enhancing TSUS forecast skill from subseasonal to seasonal timescales.

  • Research Article
  • 10.55041/ijsrem26057
CLASSIFICATION OF FRESH AND ROTTEN FRUITS USING DIFFERENT CNN MODELS
  • Oct 1, 2023
  • INTERANTIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT
  • Ms Suseela D + 3 more

Fruit freshness automated classification is crucial to the agricultural sector. In the traditional procedure, a human being grades the fruit. Additionally, this process is labor-intensive, time-consuming, and ineffective. Additionally, it raises production costs. Therefore, a quick, precise, and automated system that may lessen human effort, enhance production, and decrease manufacturing time and cost is needed for industrial applications. The deep learning- based model for classifying fruit freshness is used in the current work. Various Convolution Neural Network (CNN) models are proposed, and they are implemented using the publicly available "fruit fresh and rotten for classification" kaggle dataset. Three fresh fruit varieties (Apple, Banana, and Oranges) and their rotting category are employed in an experiment using the dataset. From the given fruit photos, traits or attributes are extracted using a CNN model based on deep learning. The input photos are then divided into fresh and rotting categories by a softmax method. The classification of fresh and rotten fruits uses a variety of CNN models, including Resnet50 (50 Layers), InceptionV3 (48 Layers), and VGG16 (16 Layers). The proposed various CNN models accurately and efficiently evaluate the dataset. Later, the accuracy of the proposed CNN models is compared and the highest accuracy among the three CNN models is identified. In this way, the best accuracy CNN models will be identified for classifying the fresh and rotten fruits. KEYWORDS: Deep learning, CNN model, Inception V3, Resnet50, VGG16.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 149
  • 10.3390/electronics12040955
Classifying Brain Tumors on Magnetic Resonance Imaging by Using Convolutional Neural Networks
  • Feb 14, 2023
  • Electronics
  • Marco Antonio Gómez-Guzmán + 7 more

The study of neuroimaging is a very important tool in the diagnosis of central nervous system tumors. This paper presents the evaluation of seven deep convolutional neural network (CNN) models for the task of brain tumor classification. A generic CNN model is implemented and six pre-trained CNN models are studied. For this proposal, the dataset utilized in this paper is Msoud, which includes Fighshare, SARTAJ, and Br35H datasets, containing 7023 MRI images. The magnetic resonance imaging (MRI) in the dataset belongs to four classes, three brain tumors, including Glioma, Meningioma, and Pituitary, and one class of healthy brains. The models are trained with input MRI images with several preprocessing strategies applied in this paper. The CNN models evaluated are Generic CNN, ResNet50, InceptionV3, InceptionResNetV2, Xception, MobileNetV2, and EfficientNetB0. In the comparison of all CNN models, including a generic CNN and six pre-trained models, the best CNN model for this dataset was InceptionV3, which obtained an average Accuracy of 97.12%. The development of these techniques could help clinicians specializing in the early detection of brain tumors.

  • Research Article
  • Cite Count Icon 43
  • 10.1007/s00330-020-07418-z
Automatic detection and classification of rib fractures based on patients' CT images and clinical information via convolutional neural network.
  • Nov 17, 2020
  • European radiology
  • Qing-Qing Zhou + 10 more

To develop a convolutional neural network (CNN) model for the automatic detection and classification of rib fractures in actual clinical practice based on cross-modal data (clinical information and CT images). In this retrospective study, CT images and clinical information (age, sex and medical history) from 1020 participants were collected and divided into a single-centre training set (n = 760; age: 55.8 ± 13.4years; men: 500), a single-centre testing set (n = 134; age: 53.1 ± 14.3years; men: 90), and two independent multicentre testing sets from two different hospitals (n = 62, age: 57.97 ± 11.88, men: 41; n = 64, age: 57.40 ± 13.36, men: 35). A Faster Region-based CNN (Faster R-CNN) model was applied to integrate CT images and clinical information. Then, a result merging technique was used to convert 2D inferences into 3D lesion results. The diagnostic performance was assessed on the basis of the receiver operating characteristic (ROC) curve, free-response ROC (fROC) curve, precision, recall (sensitivity), F1-score, and diagnosis time. The classification performance was evaluated in terms of the area under the ROC curve (AUC), sensitivity, and specificity. The CNN model showed improved performance on fresh, healing, and old fractures and yielded good classification performance for all three categories when both clinical information and CT images were used compared to the use of CT images alone. Compared with experienced radiologists, the CNN model achieved higher sensitivity (mean sensitivity: 0.95 > 0.77, 0.89 > 0.61 and 0.80 > 0.55), comparable precision (mean precision: 0.91 > 0.87, 0.84 > 0.77, and 0.95 > 0.70), and a shorter diagnosis time (average reduction of 126.15s). A CNN model combining CT images and clinical information can automatically detect and classify rib fractures with good performance and feasibility in actual clinical practice. • The developed convolutional neural network (CNN) performed better in fresh, healing, and old fractures and yielded a good classification performance in three categories, if both (clinical information and CT images) were used compared to CT images alone. • The CNN model had a higher sensitivity and matched precision in three categoriesthan experienced radiologists with a shorter diagnosis time in actual clinical practice.

  • Research Article
  • Cite Count Icon 40
  • 10.1016/j.foodcont.2022.109291
Identification of slightly sprouted wheat kernels using hyperspectral imaging technology and different deep convolutional neural networks
  • Aug 4, 2022
  • Food Control
  • Jingwu Zhu + 3 more

Identification of slightly sprouted wheat kernels using hyperspectral imaging technology and different deep convolutional neural networks

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.