Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
      More Filters
      Clear All
      More Filters
      Source
    • Language
550 result(s) for "Convolution neural network (CNN)"
Sort by:
Depthwise Separable Convolution Neural Network for High-Speed SAR Ship Detection
As an active microwave imaging sensor for the high-resolution earth observation, synthetic aperture radar (SAR) has been extensively applied in military, agriculture, geology, ecology, oceanography, etc., due to its prominent advantages of all-weather and all-time working capacity. Especially, in the marine field, SAR can provide numerous high-quality services for fishery management, traffic control, sea-ice monitoring, marine environmental protection, etc. Among them, ship detection in SAR images has attracted more and more attention on account of the urgent requirements of maritime rescue and military strategy formulation. Nowadays, most researches are focusing on improving the ship detection accuracy, while the detection speed is frequently neglected, regardless of traditional feature extraction methods or modern deep learning (DL) methods. However, the high-speed SAR ship detection is of great practical value, because it can provide real-time maritime disaster rescue and emergency military planning. Therefore, in order to address this problem, we proposed a novel high-speed SAR ship detection approach by mainly using depthwise separable convolution neural network (DS-CNN). In this approach, we integrated multi-scale detection mechanism, concatenation mechanism and anchor box mechanism to establish a brand-new light-weight network architecture for the high-speed SAR ship detection. We used DS-CNN, which consists of a depthwise convolution (D-Conv2D) and a pointwise convolution (P-Conv2D), to substitute for the conventional convolution neural network (C-CNN). In this way, the number of network parameters gets obviously decreased, and the ship detection speed gets dramatically improved. We experimented on an open SAR ship detection dataset (SSDD) to validate the correctness and feasibility of the proposed method. To verify the strong migration capacity of our method, we also carried out actual ship detection on a wide-region large-size Sentinel-1 SAR image. Ultimately, under the same hardware platform with NVIDIA RTX2080Ti GPU, the experimental results indicated that the ship detection speed of our proposed method is faster than other methods, meanwhile the detection accuracy is only lightly sacrificed compared with the state-of-art object detectors. Our method has great application value in real-time maritime disaster rescue and emergency military planning.
Tomato Leaf Disease Detection using Deep Learning Techniques
Plant diseases cause low agricultural productivity. Plant diseases are challenging to control and identify by the majority of farmers. In order to reduce future losses, early disease diagnosis is necessary. This study looks at how to identify tomato plant leaf disease using machine learning techniques, including the Fuzzy Support Vector Machine (Fuzzy-SVM), Convolution Neural Network (CNN), and Region-based Convolution Neural Network (R-CNN). The findings were confirmed using images of tomato leaves with six diseases and healthy samples. Image scaling, color thresholding, flood filling approaches for segmentation, gradient local ternary pattern, and Zernike moments’ features are used to train the pictures. R-CNN classifiers are used to classify the illness kind. The classification methods of Fuzzy SVM and CNN are analyzed and compared with R-CNN to determine the most accurate model for plant disease prediction. The R-CNN-based Classifier has the most remarkable accuracy of 96.735 percent compared to the other classification approaches.
Application of Enhanced CPC for Load Identification, Preventive Maintenance and Grid Interpretation
Currents’ Physical Components (CPC) theory with spectral component representation is proposed as a generic grid interpretation method for detecting variations and structures. It is shown theoretically and validated experimentally that scattered and reactive CPC currents are highly suited for anomaly detection. CPC are enhanced by recursively disassembling the currents into 6 scattered subcomponents and 22 subcomponents overall, where additional anomalies dominate the subcurrents. Further disassembly is useful for anomaly detection and for grid deciphering. It is shown that the newly introduced syntax is highly effective for identifying variations even when the detected signals are in the order of 10−3 compared to conventional methods. The admittance physical components’ transfer functions, Yi(ω), have been shown to improve the physical sensory function. The approach is exemplified in two scenarios demonstrating much higher sensitivity than classical electrical measurements. The proposed module may be located at a data center remote from the sensor. The CPC preprocessor, by means of a deep learning CNN, is compared to the current FFT and the current input raw data, which demonstrates 18% improved accuracy over FFT and 45% improved accuracy over raw current i(t). It is shown that the new preprocessor/detector enables highly accurate anomaly detection with the CNN classification core.
Review of deep learning: concepts, CNN architectures, challenges, applications, future directions
In the last few years, the deep learning (DL) computing paradigm has been deemed the Gold Standard in the machine learning (ML) community. Moreover, it has gradually become the most widely used computational approach in the field of ML, thus achieving outstanding results on several complex cognitive tasks, matching or even beating those provided by human performance. One of the benefits of DL is the ability to learn massive amounts of data. The DL field has grown fast in the last few years and it has been extensively used to successfully address a wide range of traditional applications. More importantly, DL has outperformed well-known ML techniques in many domains, e.g., cybersecurity, natural language processing, bioinformatics, robotics and control, and medical information processing, among many others. Despite it has been contributed several works reviewing the State-of-the-Art on DL, all of them only tackled one aspect of the DL, which leads to an overall lack of knowledge about it. Therefore, in this contribution, we propose using a more holistic approach in order to provide a more suitable starting point from which to develop a full understanding of DL. Specifically, this review attempts to provide a more comprehensive survey of the most important aspects of DL and including those enhancements recently added to the field. In particular, this paper outlines the importance of DL, presents the types of DL techniques and networks. It then presents convolutional neural networks (CNNs) which the most utilized DL network type and describes the development of CNNs architectures together with their main features, e.g., starting with the AlexNet network and closing with the High-Resolution network (HR.Net). Finally, we further present the challenges and suggested solutions to help researchers understand the existing research gaps. It is followed by a list of the major DL applications. Computational tools including FPGA, GPU, and CPU are summarized along with a description of their influence on DL. The paper ends with the evolution matrix, benchmark datasets, and summary and conclusion.
Theoretical Understanding of Convolutional Neural Network: Concepts, Architectures, Applications, Future Directions
Convolutional neural networks (CNNs) are one of the main types of neural networks used for image recognition and classification. CNNs have several uses, some of which are object recognition, image processing, computer vision, and face recognition. Input for convolutional neural networks is provided through images. Convolutional neural networks are used to automatically learn a hierarchy of features that can then be utilized for classification, as opposed to manually creating features. In achieving this, a hierarchy of feature maps is constructed by iteratively convolving the input image with learned filters. Because of the hierarchical method, higher layers can learn more intricate features that are also distortion and translation invariant. The main goals of this study are to help academics understand where there are research gaps and to talk in-depth about CNN’s building blocks, their roles, and other vital issues.
TextConvoNet: a convolutional neural network based architecture for text classification
This paper presents, TextConvoNet, a novel Convolutional Neural Network (CNN) based architecture for binary and multi-class text classification problems. Most of the existing CNN-based models use one-dimensional convolving filters, where each filter specializes in extracting n-grams features of a particular input word embeddings (Sentence Matrix). These features can be termed as intra-sentence n-gram features. To the best of our knowledge, all the existing CNN models for text classification are based on the aforementioned concept. The presented TextConvoNet not only extracts the intra-sentence n-gram features but also captures the inter-sentence n-gram features in input text data. It uses an alternative approach for input matrix representation and applies a two-dimensional multi-scale convolutional operation on the input. We perform an experimental study on five binary and multi-class classification datasets and evaluate the performance of the TextConvoNet for text classification. The results are evaluated using eight performance measures, accuracy, precision, recall, f1-score, specificity, gmean1, gmean2, and Mathews correlation coefficient (MCC). Furthermore, we extensively compared presented TextConvoNet with machine learning, deep learning, and attention-based models. The experimental results evidenced that the presented TextConvoNet outperformed and yielded better performance than the other used models for text classification purposes.
Novel Transfer Learning Approach for Medical Imaging with Limited Labeled Data
Deep learning requires a large amount of data to perform well. However, the field of medical image analysis suffers from a lack of sufficient data for training deep learning models. Moreover, medical images require manual labeling, usually provided by human annotators coming from various backgrounds. More importantly, the annotation process is time-consuming, expensive, and prone to errors. Transfer learning was introduced to reduce the need for the annotation process by transferring the deep learning models with knowledge from a previous task and then by fine-tuning them on a relatively small dataset of the current task. Most of the methods of medical image classification employ transfer learning from pretrained models, e.g., ImageNet, which has been proven to be ineffective. This is due to the mismatch in learned features between the natural image, e.g., ImageNet, and medical images. Additionally, it results in the utilization of deeply elaborated models. In this paper, we propose a novel transfer learning approach to overcome the previous drawbacks by means of training the deep learning model on large unlabeled medical image datasets and by next transferring the knowledge to train the deep learning model on the small amount of labeled medical images. Additionally, we propose a new deep convolutional neural network (DCNN) model that combines recent advancements in the field. We conducted several experiments on two challenging medical imaging scenarios dealing with skin and breast cancer classification tasks. According to the reported results, it has been empirically proven that the proposed approach can significantly improve the performance of both classification scenarios. In terms of skin cancer, the proposed model achieved an F1-score value of 89.09% when trained from scratch and 98.53% with the proposed approach. Secondly, it achieved an accuracy value of 85.29% and 97.51%, respectively, when trained from scratch and using the proposed approach in the case of the breast cancer scenario. Finally, we concluded that our method can possibly be applied to many medical imaging problems in which a substantial amount of unlabeled image data is available and the labeled image data is limited. Moreover, it can be utilized to improve the performance of medical imaging tasks in the same domain. To do so, we used the pretrained skin cancer model to train on feet skin to classify them into two classes—either normal or abnormal (diabetic foot ulcer (DFU)). It achieved an F1-score value of 86.0% when trained from scratch, 96.25% using transfer learning, and 99.25% using double-transfer learning.
Deep Learning Based Lung Cancer Detection and Classification
Lung diseases are indeed the lung-affecting diseases which impair the respiratory mechanism. Lung cancer has been one of the leading causes of mortality in humans worldwide. Early detection can enhance survival chances amid humans. If the condition is diagnosed in time, the average survival rates for people with lung cancer rise from 14 to 49 percent. While computed tomography (CT) is far more effective than X-ray, a thorough diagnosis includes multiple imaging approaches to support each other. A deep neural network for detecting lung cancer from CT images is developed and evaluated. For the classification of the lung image as normal or malignant, a densely connected convolution neural network (DenseNet) and adaptive boosting algorithm wasused. A dataset of 201 lung images is used in which 85% of the images are used for training and 15% of the images are used for testing and classification. Experimental results showed that the proposed method achieved an accuracy of 90.85%. Keywords: DenseNet, Image Processing, Deep Learning, Convolution Neural Networks (CNN).
Detection of myocardial infarction based on novel deep transfer learning methods for urban healthcare in smart cities
One of the common cardiac disorders is a cardiac attack called Myocardial infarction (MI), which occurs due to the blockage of one or more coronary arteries. Timely treatment of MI is important and slight delay results in severe consequences. Electrocardiogram (ECG) is the main diagnostic tool to monitor and reveal the MI signals. The complex nature of MI signals along with noise poses challenges to doctors for accurate and quick diagnosis. Manually studying large amounts of ECG data can be tedious and time-consuming. Therefore, there is a need for methods to automatically analyze the ECG data and make diagnosis. Number of studies has been presented to address MI detection, but most of these methods are computationally expensive and faces the problem of overfitting while dealing real data. In this paper, an effective computer-aided diagnosis (CAD) system is presented to detect MI signals using the convolution neural network (CNN) for urban healthcare in smart cities. Two types of transfer learning techniques are employed to retrain the pre-trained VGG-Net (Fine-tuning and VGG-Net as fixed feature extractor) and obtained two new networks VGG-MI1 and VGG-MI2. In the VGG-MI1 model, the last layer of the VGG-Net model is replaced with a specific layer according to our requirements and various functions are optimized to reduce overfitting. In the VGG-MI2 model, one layer of the VGG-Net model is selected as a feature descriptor of the ECG images to describe it with informative features. Considering the limited availability of dataset, ECG data is augmented which has increased the classification performance. A standard well-known database Physikalisch-Technische Bundesanstalt (PTB) Diagnostic ECG is used for the validation of the proposed framework. It is evident from experimental results that the proposed framework achieves a high accuracy surpasses the existing methods. In terms of accuracy, sensitivity, and specificity; VGG-MI1 achieved 99.02%, 98.76%, and 99.17%, respectively, while VGG-MI2 models achieved an accuracy of 99.22%, a sensitivity of 99.15%, and a specificity of 99.49%.
Object Detection in Very High-Resolution Aerial Images Using One-Stage Densely Connected Feature Pyramid Network
Object detection in very high-resolution (VHR) aerial images is an essential step for a wide range of applications such as military applications, urban planning, and environmental management. Still, it is a challenging task due to the different scales and appearances of the objects. On the other hand, object detection task in VHR aerial images has improved remarkably in recent years due to the achieved advances in convolution neural networks (CNN). Most of the proposed methods depend on a two-stage approach, namely: a region proposal stage and a classification stage such as Faster R-CNN. Even though two-stage approaches outperform the traditional methods, their optimization is not easy and they are not suitable for real-time applications. In this paper, a uniform one-stage model for object detection in VHR aerial images has been proposed. In order to tackle the challenge of different scales, a densely connected feature pyramid network has been proposed by which high-level multi-scale semantic feature maps with high-quality information are prepared for object detection. This work has been evaluated on two publicly available datasets and outperformed the current state-of-the-art results on both in terms of mean average precision (mAP) and computation time.