Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
      More Filters
      Clear All
      More Filters
      Source
    • Language
972 result(s) for "image pyramid"
Sort by:
Multi-Modality Medical Image Fusion Using Convolutional Neural Network and Contrast Pyramid
Medical image fusion techniques can fuse medical images from different morphologies to make the medical diagnosis more reliable and accurate, which play an increasingly important role in many clinical applications. To obtain a fused image with high visual quality and clear structure details, this paper proposes a convolutional neural network (CNN) based medical image fusion algorithm. The proposed algorithm uses the trained Siamese convolutional network to fuse the pixel activity information of source images to realize the generation of weight map. Meanwhile, a contrast pyramid is implemented to decompose the source image. According to different spatial frequency bands and a weighted fusion operator, source images are integrated. The results of comparative experiments show that the proposed fusion algorithm can effectively preserve the detailed structure information of source images and achieve good human visual effects.
Towards Efficient Detection for Small Objects via Attention-Guided Detection Network and Data Augmentation
Small object detection has always been a difficult direction in the field of object detection, especially the detection of small objects in UAV aerial images. The images captured by UAVs have the characteristics of small objects and dense objects. In order to solve these two problems, this paper improves the performance of object detection from the aspects of data and network structure. In terms of data, the data augmentation strategy and image pyramid mechanism are mainly used. The data augmentation strategy adopts the method of image division, which can greatly increase the number of small objects, making it easier for the algorithm to be fully trained during the training process. Since the object is denser, the image pyramid mechanism is used. During the training process, the divided images are up-sampled into three different sizes, and then sent to three different detectors respectively. Finally, the detection results of the three detectors are fused to obtain the final detection results. The small object itself has few pixels and few features. In order to improve the detection performance, it is necessary to use context. This paper adds attention mechanism to the yolov5 network structure, while adding a detection head to the underlying feature map to make the network structure pay more attention to small objects. By using data augmentation and improved network structure, the detection performance of small objects can be significantly improved. The experiment in this paper is carried out on the Visdrone2019 dataset and DOTA dataset. Through experimental verification, our proposed method can significantly improve the performance of small object detection.
Infrared and Visible Image Fusion Method Based on a Principal Component Analysis Network and Image Pyramid
The aim of infrared (IR) and visible image fusion is to generate a more informative image for human observation or some other computer vision tasks. The activity-level measurement and weight assignment are two key parts in image fusion. In this paper, we propose a novel IR and visible fusion method based on the principal component analysis network (PCANet) and an image pyramid. Firstly, we use the lightweight deep learning network, a PCANet, to obtain the activity-level measurement and weight assignment of IR and visible images. The activity-level measurement obtained by the PCANet has a stronger representation ability for focusing on IR target perception and visible detail description. Secondly, the weights and the source images are decomposed into multiple scales by the image pyramid, and the weighted-average fusion rule is applied at each scale. Finally, the fused image is obtained by reconstruction. The effectiveness of the proposed algorithm was verified by two datasets with more than eighty pairs of test images in total. Compared with nineteen representative methods, the experimental results demonstrate that the proposed method can achieve the state-of-the-art results in both visual quality and objective evaluation metrics.
Extracting Citrus-Growing Regions by Multiscale UNet Using Sentinel-2 Satellite Imagery
Citrus is an important commercial crop in many areas. The management and planning of citrus growing can be supported by timely and efficient monitoring of citrus-growing regions. Their complex planting structure and the weather are likely to cause problems for extracting citrus-growing regions from remote sensing images. To accurately extract citrus-growing regions, deep learning is employed, because it has a strong feature representation ability and can obtain rich semantic information. A novel model for extracting citrus-growing regions by UNet that incorporates an image pyramid structure is proposed on the basis of the Sentinel-2 satellite imagery. A pyramid-structured encoder, a decoder, and multiscale skip connections are the three main components of the model. Additionally, atrous spatial pyramid pooling is used to prevent information loss and improve the ability to learn spatial features. The experimental results show that the proposed model has the best performance, with the precision, the intersection over union, the recall, and the F1-score reaching 88.96%, 73.22%, 80.55%, and 84.54%, respectively. The extracted citrus-growing regions have regular boundaries and complete parcels. Furthermore, the proposed model has greater overall accuracy, kappa, producer accuracy, and user accuracy than the object-oriented random forest algorithm that is widely applied in various fields. Overall, the proposed method shows a better generalization ability, higher robustness, greater accuracy, and less fragmented extraction results. This research can support the rapid and accurate mapping of large-scale citrus-growing regions.
ZoomInNet: A Novel Small Object Detector in Drone Images with Cross-Scale Knowledge Distillation
Drone-based object detection has been widely applied in ground object surveillance, urban patrol, and some other fields. However, the dramatic scale changes and complex backgrounds of drone images usually result in weak feature representation of small objects, which makes it challenging to achieve high-precision object detection. Aiming to improve small objects detection, this paper proposes a novel cross-scale knowledge distillation (CSKD) method, which enhances the features of small objects in a manner similar to image enlargement, so it is termed as ZoomInNet. First, based on an efficient feature pyramid network structure, the teacher and student network are trained with images in different scales to introduce the cross-scale feature. Then, the proposed layer adaption (LA) and feature level alignment (FA) mechanisms are applied to align the feature size of the two models. After that, the adaptive key distillation point (AKDP) algorithm is used to get the crucial positions in feature maps that need knowledge distillation. Finally, the position-aware L2 loss is used to measure the difference between feature maps from cross-scale models, realizing the cross-scale information compression in a single model. Experiments on the challenging Visdrone2018 dataset show that the proposed method draws on the advantages of the image pyramid methods, while avoids the large calculation of them and significantly improves the detection accuracy of small objects. Simultaneously, the comparison with mainstream methods proves that our method has the best performance in small object detection.
Effective Mean Square Differences: A Matching Algorithm for Highly Similar Sheet Metal Parts
The accurate identification of highly similar sheet metal parts remains a challenging issue in sheet metal production. To solve this problem, this paper proposes an effective mean square differences (EMSD) algorithm that can effectively distinguish highly similar parts with high accuracy. First, multi-level downsampling and rotation searching are adopted to construct an image pyramid. Then, non-maximum suppression is utilised to determine the optimal rotation for each layer. In the matching, by re-evaluating the contribution of the difference between the corresponding pixels, the matching weight is determined according to the correlation between the grey value information of the matching pixels, and then the effective matching coefficient is determined. Finally, the proposed effective matching coefficient is adopted to obtain the final matching result. The results illustrate that this algorithm exhibits a strong discriminative ability for highly similar parts, with an accuracy of 97.1%, which is 11.5% higher than that of the traditional methods. It has excellent potential for application and can significantly improve sheet metal production efficiency.
Plant Disease Classification and Adversarial Attack Using SimAM-EfficientNet and GP-MI-FGSM
Plant diseases have received common attention, and deep learning has also been applied to plant diseases. Deep neural networks (DNNs) have achieved outstanding results in plant diseases. Furthermore, DNNs are very fragile, and adversarial attacks in image classification deserve much attention. It is important to detect the robustness of DNNs through adversarial attacks. The paper firstly improves the EfficientNet by adding the SimAM attention module. The SimAM-EfficientNet is proposed in this paper. The experimental results show that the accuracy of the improved model on PlantVillage reaches 99.31%. The accuracy of ResNet50 is 98.33%. The accuracy of ResNet18 is 98.31%. The accuracy of DenseNet is 98.90%. In addition, the GP-MI-FGSM adversarial attack algorithm improved by gamma correction and image pyramid in this paper can increase the success rate of attack. The model proposed in this paper has an error rate of 87.6% whenattacked by the GP-MI-FGSM adversarial attack algorithm. The success rate of GP-MI-FGSM proposed in this paper is higher than other adversarial attack algorithms, including FGSM, I-FGSM, and MI-FGSM.
MBRARN: multibranch residual attention reconstruction network for medical image fusion
Medical image fusion aims to integrate complementary information from multimodal medical images and has been widely applied in the field of medicine, such as clinical diagnosis, pathology analysis, and healing examinations. For the fusion task, feature extraction is a crucial step. To obtain significant information embedded in medical images, many deep learning-based algorithms have been proposed recently and achieved good fusion results. However, most of them can hardly capture the independent and underlying features, which leads to unsatisfactory fusion results. To address these issues, a multibranch residual attention reconstruction network (MBRARN) is proposed for the medical image fusion task. The proposed network mainly consists of three parts: feature extraction, feature fusion, and feature reconstruction. Firstly, the input medical images are converted into three scales by image pyramid operation and then are input into three branches of the proposed network respectively. The purpose of this procedure is to capture the local detailed information and the global structural information. Then, convolutions with residual attention modules are designed, which can not only enhance the captured outstanding features, but also make the network converge fast and stably. Finally, feature fusion is performed with the designed fusion strategy. In this step, a new more effective fusion strategy is correspondently designed for MRI-SPECT based on the Euclidean norm, called feature distance ratio (FDR). The experimental results conducted on Harvard whole brain atlas dataset demonstrate that the proposed network can achieve better results in terms of both subjective and objective evaluation, compared with some state-of-the-art medical image fusion algorithms.
RIPF-Unet for regional landslides detection: a novel deep learning model boosted by reversed image pyramid features
Rapid detection of landslides using remote sensing images plays a key role in hazard assessment and mitigation. Many deep convolutional neural network-based models have been proposed for this purpose; however, for small-scale landslide detection, excessive convolution and pooling process may cause potential texture information loss, which can lead to misclassification of landslide target. In this paper, we present a novel UNet model for the automatic detection of landslides, wherein the reversed image pyramid features (RIPFs) are adapted to mitigate the information loss caused by a succession of convolution and pooling. The proposed RIPF-Unet model is trained and validated using the open-source landslides dataset of the Bijie area, Guizhou Province, China, wherein the precision of the proposed model is observed to increase by 3.5% and 4.0%, compared to the conventional UNet and UNet + + model, respectively. The proposed RIPF-Unet model is further applied to the case of the Longtoushan region after the 2014 Ms.6.5 Ludian earthquake. Results show that the proposed model achieves a 96.63% accuracy for detecting landslides using remote sensing images. And the RIPF-Unet model is also advanced in its compact parameter size; notably, it is 31% lighter compared to the UNet + + model.
Rich Structural Index for Stereoscopic Image Quality Assessment
The human visual system (HVS), affected by viewing distance when perceiving the stereo image information, is of great significance to study of stereoscopic image quality assessment. Many methods of stereoscopic image quality assessment do not have comprehensive consideration for human visual perception characteristics. In accordance with this, we propose a Rich Structural Index (RSI) for Stereoscopic Image objective Quality Assessment (SIQA) method based on multi-scale perception characteristics. To begin with, we put the stereo pair into the image pyramid based on Contrast Sensitivity Function (CSF) to obtain sensitive images of different resolution. Then, we obtain local Luminance and Structural Index (LSI) in a locally adaptive manner on gradient maps which consider the luminance masking and contrast masking. At the same time we use Singular Value Decomposition (SVD) to obtain the Sharpness and Intrinsic Structural Index (SISI) to effectively capture the changes introduced in the image (due to distortion). Meanwhile, considering the disparity edge structures, we use gradient cross-mapping algorithm to obtain Depth Texture Structural Index (DTSI). After that, we apply the standard deviation method for the above results to obtain contrast index of reference and distortion components. Finally, for the loss caused by the randomness of the parameters, we use Support Vector Machine Regression based on Genetic Algorithm (GA-SVR) training to obtain the final quality score. We conducted a comprehensive evaluation with state-of-the-art methods on four open databases. The experimental results show that the proposed method has stable performance and strong competitive advantage.