Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
      More Filters
      Clear All
      More Filters
      Source
    • Language
166 result(s) for "convolutional block attention module"
Sort by:
RAANet: A Residual ASPP with Attention Framework for Semantic Segmentation of High-Resolution Remote Sensing Images
Classification of land use and land cover from remote sensing images has been widely used in natural resources and urban information management. The variability and complex background of land use in high-resolution imagery poses greater challenges for remote sensing semantic segmentation. To obtain multi-scale semantic information and improve the classification accuracy of land-use types in remote sensing images, the deep learning models have been wildly focused on. Inspired by the idea of the atrous-spatial pyramid pooling (ASPP) framework, an improved deep learning model named RAANet (Residual ASPP with Attention Net) is constructed in this paper, which constructed a new residual ASPP by embedding the attention module and residual structure into the ASPP. There are 5 dilated attention convolution units and a residual unit in its encoder. The former is used to obtain important semantic information at more scales, and residual units are used to reduce the complexity of the network to prevent the disappearance of gradients. In practical applications, according to the characteristics of the data set, the attention unit can select different attention modules such as the convolutional block attention model (CBAM). The experimental results obtained from the land-cover domain adaptive semantic segmentation (LoveDA) and ISPRS Vaihingen datasets showed that this model can enhance the classification accuracy of semantic segmentation compared to the current deep learning models.
A CNN-Transformer Network Combining CBAM for Change Detection in High-Resolution Remote Sensing Images
Current deep learning-based change detection approaches mostly produce convincing results by introducing attention mechanisms to traditional convolutional networks. However, given the limitation of the receptive field, convolution-based methods fall short of fully modelling global context and capturing long-range dependencies, thus insufficient in discriminating pseudo changes. Transformers have an efficient global spatio-temporal modelling capability, which is beneficial for the feature representation of changes of interest. However, the lack of detailed information may cause the transformer to locate the boundaries of changed regions inaccurately. Therefore, in this article, a hybrid CNN-transformer architecture named CTCANet, combining the strengths of convolutional networks, transformer, and attention mechanisms, is proposed for high-resolution bi-temporal remote sensing image change detection. To obtain high-level feature representations that reveal changes of interest, CTCANet utilizes tokenizer to embed the features of each image extracted by convolutional network into a sequence of tokens, and the transformer module to model global spatio-temporal context in token space. The optimal bi-temporal information fusion approach is explored here. Subsequently, the reconstructed features carrying deep abstract information are fed to the cascaded decoder to aggregate with features containing shallow fine-grained information, through skip connections. Such an aggregation empowers our model to maintain the completeness of changes and accurately locate small targets. Moreover, the integration of the convolutional block attention module enables the smoothing of semantic gaps between heterogeneous features and the accentuation of relevant changes in both the channel and spatial domains, resulting in more impressive outcomes. The performance of the proposed CTCANet surpasses that of recent certain state-of-the-art methods, as evidenced by experimental results on two publicly accessible datasets, LEVIR-CD and SYSU-CD.
A Light-Weight Network for Small Insulator and Defect Detection Using UAV Imaging Based on Improved YOLOv5
Insulator defect detection is of great significance to compromise the stability of the power transmission line. The state-of-the-art object detection network, YOLOv5, has been widely used in insulator and defect detection. However, the YOLOv5 network has limitations such as poor detection rate and high computational loads in detecting small insulator defects. To solve these problems, we proposed a light-weight network for insulator and defect detection. In this network, we introduced the Ghost module into the YOLOv5 backbone and neck to reduce the parameters and model size to enhance the performance of unmanned aerial vehicles (UAVs). Besides, we added small object detection anchors and layers for small defect detection. In addition, we optimized the backbone of YOLOv5 by applying convolutional block attention modules (CBAM) to focus on critical information for insulator and defect detection and suppress uncritical information. The experiment result shows the mean average precision (mAP) is set to 0.5, and the mAP is set from 0.5 to 0.95 of our model and can reach 99.4% and 91.7%; the parameters and model size were reduced to 3,807,372 and 8.79 M, which can be easily deployed to embedded devices such as UAVs. Moreover, the speed of detection can reach 10.9 ms/image, which can meet the real-time detection requirement.
ANC: Attention Network for COVID-19 Explainable Diagnosis Based on Convolutional Block Attention Module
Aim: To diagnose COVID-19 more efficiently and more correctly, this study proposed a novel attention network for COVID-19 (ANC). Methods: Two datasets were used in this study. An 18-way data augmentation was proposed to avoid over tting. Then, convolutional block attention module (CBAM) was integrated to our model, the structure of which is fine-tuned. Finally, Grad-CAM was used to provide an explainable diagnosis. Results: The accuracy of our ANC methods on two datasets are 96.32% ± 1.06%, and 96.00% ± 1.03%, respectively. Conclusions: This proposed ANC method is superior to 9 state-of-the-art approaches.
Remaining Useful Life Prediction of Rolling Bearings Based on CBAM-CNN-LSTM
Predicting the Remaining Useful Life (RUL) is vital for ensuring the reliability and safety of equipment and components. This study introduces a novel method for predicting RUL that utilizes the Convolutional Block Attention Module (CBAM) to address the problem that Convolutional Neural Networks (CNNs) do not effectively leverage data channel features and spatial features in residual life prediction. Firstly, Fast Fourier Transform (FFT) is applied to convert the data into the frequency domain. The resulting frequency domain data is then used as input to the convolutional neural network for feature extraction; Then, the weights of channel features and spatial features are assigned to the extracted features by CBAM, and the weighted features are then input into the Long Short-Term Memory (LSTM) network to learn temporal features. Finally, the effectiveness of the proposed model is verified using the PHM2012 bearing dataset. Compared to several existing RUL prediction methods, the mean squared error, mean absolute error, and root mean squared error of the proposed method in this paper are reduced by 53%, 16.87%, and 31.68%, respectively, which verifies the superiority of the method. Meanwhile, the experimental results demonstrate that the proposed method achieves good RUL prediction accuracy across various failure modes.
Rolling Bearing Fault Diagnosis Method Based on Attention CNN and BiLSTM Network
To solve the problems that existing bearing fault diagnosis methods cannot adaptively select features and are difficult to deal with noise interference, an end-to-end fault diagnosis method is proposed based on attention CNN and BiLSTM (ACNN-BiLSTM). In the proposed method, the raw vibration acceleration signal of the bearing is taken as the input, the short-term spatial features are extracted through a one-dimensional wide convolutional neural network, and the batch normalization algorithm is used to improve the stability of the data distribution. Following, a convolutional block attention module is introduced to redistribute the weights between different feature dimensions, enhancing the model's attention to important features. Finally, the attention-weighted features are sent to BiLSTM for further feature extraction, and the softmax classifier is used for fault diagnosis. The proposed method is compared with advanced algorithms such as WCNN-BiGRU on the CWRU public dataset. The experimental results show that ACNN-BiLSTM has the highest accuracy, recall, and F1-Measure. Even under the extreme noise interference condition of SNR = 10 dB, ACNN-BiLSTM can achieve a diagnostic accuracy of 96.58%. In addition, the proposed method is also used for fault diagnosis of bearing measured data of the VALENIAN-PT500 test bench. The results show that the average diagnostic accuracy of ACNN-BiLSTM is up to 99.79%, which has strong generality and is superior to other advanced comparison methods.
ADVIAN: Alzheimer's Disease VGG-Inspired Attention Network Based on Convolutional Block Attention Module and Multiple Way Data Augmentation
Alzheimer's disease is a neurodegenerative disease that causes 60-70% of all cases of dementia. This study is to provide a novel method that can identify AD more accurately. We first propose a VGG-inspired network (VIN) as the backbone network and investigate the use of attention mechanisms. We proposed an Alzheimer's Disease VGG-Inspired Attention Network (ADVIAN), where we integrate convolutional block attention modules on a VIN backbone. Also, 18-way data augmentation is proposed to avoid overfitting. Ten runs of 10-fold cross-validation are carried out to report the unbiased performance. The sensitivity and specificity reach 97.65 ± 1.36 and 97.86 ± 1.55, respectively. Its precision and accuracy are 97.87 ± 1.53 and 97.76 ± 1.13, respectively. The F1 score, MCC, and FMI are obtained as 97.75 ± 1.13, 95.53 ± 2.27, and 97.76 ± 1.13, respectively. The AUC is 0.9852. The proposed ADVIAN gives better results than 11 state-of-the-art methods. Besides, experimental results demonstrate the effectiveness of 18-way data augmentation.
Gearbox Fault Diagnosis Based on MSCNN-LSTM-CBAM-SE
Ensuring the safety of mechanical equipment, gearbox fault diagnosis is crucial for the stable operation of the whole system. However, existing diagnostic methods still have limitations, such as the analysis of single-scale features and insufficient recognition of global temporal dependencies. To address these issues, this article proposes a new method for gearbox fault diagnosis based on MSCNN-LSTM-CBAM-SE. The output of the CBAM-SE module is deeply integrated with the multi-scale features from MSCNN and the temporal features from LSTM, constructing a comprehensive feature representation that provides richer and more precise information for fault diagnosis. The effectiveness of this method has been validated with two sets of gearbox datasets and through ablation studies on this model. Experimental results show that the proposed model achieves excellent performance in terms of accuracy and F1 score, among other metrics. Finally, a comparison with other relevant fault diagnosis methods further verifies the advantages of the proposed model. This research offers a new solution for accurate fault diagnosis of gearboxes.
Rail State Recognition Method Based on a Convolutional Block Attention Module
In order to solve the problems of high subjectivity and serious lag in traditional rail surface state recognition based on manual experience, a rail surface state recognition model based on convolution attention is proposed. First, the transfer learning method is used to pre-train the ImageNet data set to obtain the model parameters. Secondly, the 3×3 convolution in the ResNet-50 residual block is replaced with a convolutional block attention module (CBAM) to obtain a new CBAM-ResNet. Finally, the model identification results of the rail surface status are obtained through the Softmax classifier. The results show that this method can effectively identify the rail surface status, and the identification accuracy can reach 99.68%, which is better than other deep neural network models.
Utilizing active learning and attention-CNN to classify vegetation based on UAV multispectral data
This paper presents a deep learning model based on an active learning strategy. The model achieves accurate identification of vegetation types in the study area by utilizing multispectral data obtained from preprocessing of unmanned aerial vehicle (UAV) remote sensing equipment. This approach offers advantages such as high data accuracy, mobility, and easy data collection. In active learning, the minimum confidence scoring method and a sampling technique based on a data pool are employed to reduce labeling costs. The deep learning model incorporates a semantic segmentation gated full fusion module that integrates a dual attention mechanism. This module enhances the capture of detailed texture information, optimally allocates spectral weights, and improves the model’s ability to distinguish between similar categories. At a labeling cost of 20%, the average accuracy of the model is 93.2%. Compared with other models, the proposed model achieved the highest classification accuracy in the case of limited training samples. At full annotation cost, the average accuracy is 95.32%, with only a difference of about 2%, but saving 80% of annotation cost. Therefore, active learning strategies can filter out high-value samples that are beneficial for model training, greatly reducing the annotation cost of samples Finally, the recognition results of surface vegetation cover types in the study area are presented, and the model’s accuracy is verified through field investigation.