Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
222
result(s) for
"MobileNetV2"
Sort by:
COVID-19 Detection Using Deep Learning Algorithm on Chest X-ray Images
2021
COVID-19, regarded as the deadliest virus of the 21st century, has claimed the lives of millions of people around the globe in less than two years. Since the virus initially affects the lungs of patients, X-ray imaging of the chest is helpful for effective diagnosis. Any method for automatic, reliable, and accurate screening of COVID-19 infection would be beneficial for rapid detection and reducing medical or healthcare professional exposure to the virus. In the past, Convolutional Neural Networks (CNNs) proved to be quite successful in the classification of medical images. In this study, an automatic deep learning classification method for detecting COVID-19 from chest X-ray images is suggested using a CNN. A dataset consisting of 3616 COVID-19 chest X-ray images and 10,192 healthy chest X-ray images was used. The original data were then augmented to increase the data sample to 26,000 COVID-19 and 26,000 healthy X-ray images. The dataset was enhanced using histogram equalization, spectrum, grays, cyan and normalized with NCLAHE before being applied to CNN models. Initially using the dataset, the symptoms of COVID-19 were detected by employing eleven existing CNN models; VGG16, VGG19, MobileNetV2, InceptionV3, NFNet, ResNet50, ResNet101, DenseNet, EfficientNetB7, AlexNet, and GoogLeNet. From the models, MobileNetV2 was selected for further modification to obtain a higher accuracy of COVID-19 detection. Performance evaluation of the models was demonstrated using a confusion matrix. It was observed that the modified MobileNetV2 model proposed in the study gave the highest accuracy of 98% in classifying COVID-19 and healthy chest X-rays among all the implemented CNN models. The second-best performance was achieved from the pre-trained MobileNetV2 with an accuracy of 97%, followed by VGG19 and ResNet101 with 95% accuracy for both the models. The study compares the compilation time of the models. The proposed model required the least compilation time with 2 h, 50 min and 21 s. Finally, the Wilcoxon signed-rank test was performed to test the statistical significance. The results suggest that the proposed method can efficiently identify the symptoms of infection from chest X-ray images better than existing methods.
Journal Article
Fruit Image Classification Model Based on MobileNetV2 with Deep Transfer Learning Technique
2023
Due to the rapid emergence and evolution of AI applications, the utilization of smart imaging devices has increased significantly. Researchers have started using deep learning models, such as CNN, for image classification. Unlike the traditional models, which require a lot of features to perform well, CNN does not require any handcrafted features to perform well. It uses numerous filters, which extract required features from images automatically for classification. One of the issues in the horticulture industry is fruit classification, which requires an expert with a lot of experience. To overcome this issue an automated system is required which can classify different types of fruits without the need for any human effort. In this study, a dataset of a total of 26,149 images of 40 different types of fruits was used for experimentation. The training and test set were randomly recreated and divided into the ratio of 3:1. The experiment introduces a customized head of five different layers into MobileNetV2 architecture. The classification layer of the MobileNetV2 model is replaced by the customized head, which produced the modified version of MobileNetV2 called TL-MobileNetV2. In addition, transfer learning is used to retain the pre-trained model. TL-MobileNetV2 achieves an accuracy of 99%, which is 3% higher than MobileNetV2, and the equal error rate of TL-MobileNetV2 is just 1%. Compared to AlexNet, VGG16, InceptionV3, and ResNet, the accuracy is better by 8, 11, 6, and 10%, respectively. Furthermore, the TL-MobileNetV2 model obtained 99% precision, 99% for recall, and a 99% F1-score. It can be concluded that transfer learning plays a big part in achieving better results, and the dropout technique helps to reduce the overfitting in transfer learning.
Journal Article
Attention 3D U-Net with Multiple Skip Connections for Segmentation of Brain Tumor Images
by
Nodirov, Jakhongir
,
Whangbo, Taeg Keun
,
Abdusalomov, Akmalbek Bobomirzaevich
in
Accuracy
,
Algorithms
,
attention modules
2022
Among researchers using traditional and new machine learning and deep learning techniques, 2D medical image segmentation models are popular. Additionally, 3D volumetric data recently became more accessible, as a result of the high number of studies conducted in recent years regarding the creation of 3D volumes. Using these 3D data, researchers have begun conducting research on creating 3D segmentation models, such as brain tumor segmentation and classification. Since a higher number of crucial features can be extracted using 3D data than 2D data, 3D brain tumor detection models have increased in popularity among researchers. Until now, various significant research works have focused on the 3D version of the U-Net and other popular models, such as 3D U-Net and V-Net, while doing superior research works. In this study, we used 3D brain image data and created a new architecture based on a 3D U-Net model that uses multiple skip connections with cost-efficient pretrained 3D MobileNetV2 blocks and attention modules. These pretrained MobileNetV2 blocks assist our architecture by providing smaller parameters to maintain operable model size in terms of our computational capability and help the model to converge faster. We added additional skip connections between the encoder and decoder blocks to ease the exchange of extracted features between the two blocks, which resulted in the maximum use of the features. We also used attention modules to filter out irrelevant features coming through the skip connections and, thus, preserved more computational power while achieving improved accuracy.
Journal Article
A study on expression recognition based on improved mobilenetV2 network
2024
This paper proposes an improved strategy for the MobileNetV2 neural network(I-MobileNetV2) in response to problems such as large parameter quantities in existing deep convolutional neural networks and the shortcomings of the lightweight neural network MobileNetV2 such as easy loss of feature information, poor real-time performance, and low accuracy rate in facial emotion recognition tasks. The network inherits the characteristics of MobilenetV2 depthwise separated convolution, signifying a reduction in computational load while maintaining a lightweight profile. It utilizes a reverse fusion mechanism to retain negative features, which makes the information less likely to be lost. The SELU activation function is used to replace the RELU6 activation function to avoid gradient vanishing. Meanwhile, to improve the feature recognition capability, the channel attention mechanism (Squeeze-and-Excitation Networks (SE-Net)) is integrated into the MobilenetV2 network. Experiments conducted on the facial expression datasets FER2013 and CK + showed that the proposed network model achieved facial expression recognition accuracies of 68.62% and 95.96%, improving upon the MobileNetV2 model by 0.72% and 6.14% respectively, and the parameter count decreased by 83.8%. These results empirically verify the effectiveness of the improvements made to the network model.
Journal Article
Classification of tea leaf disease using convolutional neural network approach
2024
Leaf diseases on tea plants affect the quality of tea. This issue must be overcome since preparing tea drinks requires high-quality tea leaves. Various automatic models for identifying disease in tea leaves have been developed; however, their performance is typically low since the extracted features are not selective enough. This work presents a classification model for tea leaf disease that distinguishes six leaf classes: algal spot, brown, blight, grey blight, helopeltis, red spot, and healthy. Deep learning using a convolutional neural network (CNN) builds an effective model for detecting tea leaf illness. The Kaggle public dataset contains 5,980 tea leaf images on a white background. Pre-processing was performed to reduce computing time, which involved shrinking and normalizing the image prior to augmentation. Augmentation techniques included rotation, shear, flip horizontal, and flip vertical. The CNN model was used to classify tea leaf disease using the MobileNetV2 backbone, Adam optimizer, and rectified linear unit (ReLU) activation function with 224×224 input data. The proposed model attained the highest performance, as evidenced by the accuracy value 0.9455.
Journal Article
Dimensionality Reduction and Classification of Dermatological Images using PCA and Machine Learning
2025
Skin diseases pose grave diagnosis issues since they are highly similar among classes and have varied patterns over the various skin colors, particularly in Indian subjects. The current work proposes a mixed strategy using transfer learning-based feature extraction, dimensionality reduction, and traditional machine learning classification to effectively detect skin diseases. In an experiment conducted on a database of 9478 images for five dermatological classes, features were extracted from a pre-trained MobileNetV2 network. The statistical technique, Principal Component Analysis (PCA) was used to diminish feature dimensionality to facilitate effective visualization (3D PCA plots) and computational performance. Support Vector Machine (SVM) classifiers that used PCA-reduced features were highly accurate, with evident class separability illustrated in confusion matrices and performance metrics. The suggested framework emphasizes the promise of explainable PCA-based pipelines for skin disease analysis and presents a scalable solution for dermatological AI systems in resource-limited clinical environments.
Journal Article
High-precision multiclass classification of lung disease through customized MobileNetV2 from chest X-ray images
2023
In this study, multiple lung diseases are diagnosed with the help of the Neural Network algorithm. Specifically, Emphysema, Infiltration, Mass, Pleural Thickening, Pneumonia, Pneumothorax, Atelectasis, Edema, Effusion, Hernia, Cardiomegaly, Pulmonary Fibrosis, Nodule, and Consolidation, are studied from the ChestX-ray14 dataset. A proposed fine-tuned MobileLungNetV2 model is employed for analysis. Initially, pre-processing is done on the X-ray images from the dataset using CLAHE to increase image contrast. Additionally, a Gaussian Filter, to denoise images, and data augmentation methods are used. The pre-processed images are fed into several transfer learning models; such as InceptionV3, AlexNet, DenseNet121, VGG19, and MobileNetV2. Among these models, MobileNetV2 performed with the highest accuracy of 91.6% in overall classifying lesions on Chest X-ray Images. This model is then fine-tuned to optimise the MobileLungNetV2 model. On the pre-processed data, the fine-tuned model, MobileLungNetV2, achieves an extraordinary classification accuracy of 96.97%. Using a confusion matrix for all the classes, it is determined that the model has an overall high precision, recall, and specificity scores of 96.71%, 96.83% and 99.78% respectively. The study employs the Grad-cam output to determine the heatmap of disease detection. The proposed model shows promising results in classifying multiple lesions on Chest X-ray images.
•This paper proposes a MobileNetV2-based fine-tuned model. The highlights are:•To classify lung diseases from X-ray image data.•The proposed model classifies data with higher accuracy than pre-trained models.•The model performs image feature extraction while identifying abnormalities.•Literature comparison shows the model outperforms existing studies.
Journal Article
Comparative evaluation of left ventricle segmentation using improved pyramid scene parsing network in echocardiography
2025
Automatic segmentation of the left ventricle is a challenging task due to the presence of artifacts and speckle noise in echocardiography. This paper studies the ability of a fully supervised network based on pyramid scene parsing network (PSPNet) to implement echocardiographic left ventricular segmentation. First, the lightweight MobileNetv2 was selected to replace ResNet to adjust the coding structure of the neural network, reduce the computational complexity, and integrate the pyramid scene analysis module to construct the PSPNet; secondly, introduce dilated convolution and feature fusion to propose an improved PSPNet model, and study the impact of pre-training and transfer learning on model segmentation performance; finally, the public data set challenge on endocardial three-dimensional ultrasound segmentation (CETUS) was used to train and test different backbone and initialized PSPNet models. The results demonstrate that the improved PSPNet model has strong segmentation advantages in terms of accuracy and running speed. Compared with the two classic algorithms VGG and Unet, the dice similarity coefficient (DSC) index is increased by an average of 7.6%, Hausdorff distance (HD) is reduced by 2.9%, and the mean intersection over union (mIoU) is improved by 8.8%. Additionally, the running time is greatly shortened, indicating good clinical application potential.
Journal Article
A Lightweight CNN Based on Transfer Learning for COVID-19 Diagnosis
by
Sun, Wei
,
Zhou, Jie
,
Kumar Jha, Sunil
in
Artificial neural networks
,
Central processing units
,
Classification
2022
The key to preventing the COVID-19 is to diagnose patients quickly and accurately. Studies have shown that using Convolutional Neural Networks (CNN) to analyze chest Computed Tomography (CT) images is helpful for timely COVID-19 diagnosis. However, personal privacy issues, public chest CT data sets are relatively few, which has limited CNN's application to COVID-19 diagnosis. Also, many CNNs have complex structures and massive parameters. Even if equipped with the dedicated Graphics Processing Unit (GPU) for acceleration, it still takes a long time, which is not conductive to widespread application. To solve above problems, this paper proposes a lightweight CNN classification model based on transfer learning. Use the lightweight CNN MobileNetV2 as the backbone of the model to solve the shortage of hardware resources and computing power. In order to alleviate the problem of model overfitting caused by insufficient data set, transfer learning is used to train the model. The study first exploits the weight parameters trained on the ImageNet database to initialize the MobileNetV2 network, and then retrain the model based on the CT image data set provided by Kaggle. Experimental results on a computer equipped only with the Central Processing Unit (CPU) show that it consumes only 1.06 s on average to diagnose a chest CT image. Compared to other lightweight models, the proposed model has a higher classification accuracy and reliability while having a lightweight architecture and few parameters, which can be easily applied to computers without GPU acceleration. Code:github.com/ZhouJie-520/paper-codes.
Journal Article
Pneumonia detection on x-ray image using improved depthwise separable convolutional neural networks
A single neural network model cannot capture intricate and diverse features due to its ability to learn only a finite set of patterns from the data. Additionally, training and utilising a single model can be computationally demanding. Experts propose incorporating multiple neural network models to address these constraints to extract complementary attributes. Previous research has highlighted challenges network models face, including difficulties in effectively capturing highly detailed spatial features, redundancy in network structure parameters, and restricted generalisation capabilities. This study introduces an innovative neural network architecture that combines the Xception module with the inverse residue structure to tackle these issues. Considering this, the paper presents a model for detecting pneumonia in X-ray images employing an improved depthwise separable convolutional network. This network architecture integrates the inverse residual structure from the MobileNetV2 model, using the rectified linear unit (ReLU) non-linear activation function throughout the entire network. The experimental results show an impressive recognition rate with a test accuracy of 97.24% on the chest x-ray dataset. This method can extract more profound and abstract image features while mitigating overfitting issues and enhancing the network's generalisation capacity.
Journal Article