Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
15 result(s) for "Radiuk, Pavlo"
Sort by:
Applying 3D U-Net Architecture to the Task of Multi-Organ Segmentation in Computed Tomography
The achievement of high-precision segmentation in medical image analysis has been an active direction of research over the past decade. Significant success in medical imaging tasks has been feasible due to the employment of deep learning methods, including convolutional neural networks (CNNs). Convolutional architectures have been mostly applied to homogeneous medical datasets with separate organs. Nevertheless, the segmentation of volumetric medical images of several organs remains an open question. In this paper, we investigate fully convolutional neural networks (FCNs) and propose a modified 3D U-Net architecture devoted to the processing of computed tomography (CT) volumetric images in the automatic semantic segmentation tasks. To benchmark the architecture, we utilised the differentiable Sørensen-Dice similarity coefficient (SDSC) as a validation metric and optimised it on the training data by minimising the loss function. Our hand-crafted architecture was trained and tested on the manually compiled dataset of CT scans. The improved 3D UNet architecture achieved the average SDSC score of 84.8 % on testing subset among multiple abdominal organs. We also compared our architecture with recognised state-of-the-art results and demonstrated that 3D U-Net based architectures could achieve competitive performance and efficiency in the multi-organ segmentation task.
Intelligent Integrated System for Fruit Detection Using Multi-UAV Imaging and Deep Learning
In the context of Industry 4.0, one of the most significant challenges is enhancing efficiency in sectors like agriculture by using intelligent sensors and advanced computing. Specifically, the task of fruit detection and counting in orchards represents a complex issue that is crucial for efficient orchard management and harvest preparation. Traditional techniques often fail to provide the timely and precise data necessary for these tasks. With the agricultural sector increasingly relying on technological advancements, the integration of innovative solutions is essential. This study presents a novel approach that combines artificial intelligence (AI), deep learning (DL), and unmanned aerial vehicles (UAVs). The proposed approach demonstrates superior real-time capabilities in fruit detection and counting, utilizing a combination of AI techniques and multi-UAV systems. The core innovation of this approach is its ability to simultaneously capture and synchronize video frames from multiple UAV cameras, converting them into a cohesive data structure and, ultimately, a continuous image. This integration is further enhanced by image quality optimization techniques, ensuring the high-resolution and accurate detection of targeted objects during UAV operations. Its effectiveness is proven by experiments, achieving a high mean average precision rate of 86.8% in fruit detection and counting, which surpasses existing technologies. Additionally, it maintains low average error rates, with a false positive rate at 14.7% and a false negative rate at 18.3%, even under challenging weather conditions like cloudiness. Overall, the practical implications of this multi-UAV imaging and DL-based approach are vast, particularly for real-time fruit recognition in orchards, marking a significant stride forward in the realm of digital agriculture that aligns with the objectives of Industry 4.0.
Facial Emotion Recognition for Photo and Video Surveillance Based on Machine Learning and Visual Analytics
Modern video surveillance systems mainly rely on human operators to monitor and interpret the behavior of individuals in real time, which may lead to severe delays in responding to an emergency. Therefore, there is a need for continued research into the designing of interpretable and more transparent emotion recognition models that can effectively detect emotions in safety video surveillance systems. This study proposes a novel technique incorporating a straightforward model for detecting sudden changes in a person’s emotional state using low-resolution photos and video frames from surveillance cameras. The proposed technique includes a method of the geometric interpretation of facial areas to extract features of facial expression, the method of hyperplane classification for identifying emotional states in the feature vector space, and the principles of visual analytics and “human in the loop” to obtain transparent and interpretable classifiers. The experimental testing using the developed software prototype validates the scientific claims of the proposed technique. Its implementation improves the reliability of abnormal behavior detection via facial expressions by 0.91–2.20%, depending on different emotions and environmental conditions. Moreover, it decreases the error probability in identifying sudden emotional shifts by 0.23–2.21% compared to existing counterparts. Future research will aim to improve the approach quantitatively and address the limitations discussed in this paper.
Explainable Deep Learning: A Visual Analytics Approach with Transition Matrices
The non-transparency of artificial intelligence (AI) systems, particularly in deep learning (DL), poses significant challenges to their comprehensibility and trustworthiness. This study aims to enhance the explainability of DL models through visual analytics (VA) and human-in-the-loop (HITL) principles, making these systems more transparent and understandable to end users. In this work, we propose a novel approach that utilizes a transition matrix to interpret results from DL models through more comprehensible machine learning (ML) models. The methodology involves constructing a transition matrix between the feature spaces of DL and ML models as formal and mental models, respectively, improving the explainability for classification tasks. We validated our approach with computational experiments on the MNIST, FNC-1, and Iris datasets using a qualitative and quantitative comparison criterion, that is, how different the results obtained by our approach are from the ground truth of the training and testing samples. The proposed approach significantly enhanced model clarity and understanding in the MNIST dataset, with SSIM and PSNR values of 0.697 and 17.94, respectively, showcasing high-fidelity reconstructions. Moreover, achieving an F1m score of 77.76% and a weighted accuracy of 89.38%, our approach proved its effectiveness in stance detection with the FNC-1 dataset, complemented by its ability to explain key textual nuances. For the Iris dataset, the separating hyperplane constructed based on the proposed approach allowed for enhancing classification accuracy. Overall, using VA, HITL principles, and a transition matrix, our approach significantly improves the explainability of DL models without compromising their performance, marking a step forward in developing more transparent and trustworthy AI systems.
Criticality Assessment of Wind Turbine Defects via Multispectral UAV Fusion and Fuzzy Logic
Ensuring the structural integrity of wind turbines is crucial for the sustainability of wind energy. A significant challenge remains in transitioning from mere defect detection to objective, scalable criticality assessment for prioritizing maintenance. In this work, we propose a novel comprehensive framework that leverages multispectral unmanned aerial vehicle (UAV) imagery and a novel standards-aligned Fuzzy Inference System to automate this task. Our contribution is validated on two open research-oriented datasets representing small on- and offshore machines: the public AQUADA-GO and Thermal WTB Inspection datasets. An ensemble of YOLOv8n models trained on fused RGB-thermal data achieves a mean Average Precision (mAP@.5) of 92.8% for detecting cracks, erosion, and thermal anomalies. The core novelty, a 27-rule Fuzzy Inference System derived from the IEC 61400-5 standard, translates quantitative defect parameters into a five-level criticality score. The system’s output demonstrates exceptional fidelity to expert assessments, achieving a mean absolute error of 0.14 and a Pearson correlation of 0.97. This work provides a transparent, repeatable, and engineering-grounded proof of concept, demonstrating a promising pathway toward predictive, condition-based maintenance strategies and supporting the economic viability of wind energy.
Hierarchical Deep Learning Model for Identifying Similar Targets in UAV Imagery
Accurate object detection in UAV imagery is critical for situational awareness, yet conventional deep learning models often struggle to distinguish between visually similar targets. To address this challenge, this study introduces a hierarchical deep learning architecture that decomposes the multi-class detection task into a structured, multi-level classification cascade. Our approach combines a high-recall Faster R-CNN for initial object proposal, specialized YOLO models for granular feature extraction, and a dedicated FT-Transformer for fine-grained classification. Experimental evaluation on a complex dataset demonstrated the effectiveness of this strategy. The hierarchical model achieved an aggregate F1-score of 93.9%, representing a 1.41% improvement over the 92.46% F1-score from a traditional, non-hierarchical baseline model. These results indicate that a modular, coarse-to-fine cascade can effectively reduce inter-class ambiguity, offering a scalable approach to improving object recognition in complex UAV-based monitoring environments. This work contributes a promising approach to developing more accurate and reliable situational awareness systems.
Unsupervised Knowledge Extraction of Distinctive Landmarks from Earth Imagery Using Deep Feature Outliers for Robust UAV Geo-Localization
Vision-based navigation is a common solution for the critical challenge of GPS-denied Unmanned Aerial Vehicle (UAV) operation, but a research gap remains in the autonomous discovery of robust landmarks from aerial survey imagery needed for such systems. In this work, we propose a framework to fill this gap by identifying visually distinctive urban buildings from aerial survey imagery and curating them into a landmark database for GPS-free UAV localization. The proposed framework constructs semantically rich embeddings using intermediate layers from a pre-trained YOLOv11n-seg segmentation network. This novel technique requires no additional training. An unsupervised landmark selection strategy, based on the Isolation Forest algorithm, then identifies objects with statistically unique embeddings. Experimental validation on the VPAIR aerial-to-aerial benchmark shows that the proposed max-pooled embeddings, assembled from selected layers, significantly improve retrieval performance. The top-1 retrieval accuracy for landmarks more than doubled compared to typical buildings (0.53 vs. 0.31), and a Recall@5 of 0.70 is achieved for landmarks. Overall, this study demonstrates that unsupervised outlier selection in a carefully constructed embedding space yields a highly discriminative, computation-friendly set of landmarks suitable for real-time, robust UAV navigation.
DyTAM: Accelerating Wind Turbine Inspections with Dynamic UAV Trajectory Adaptation
Wind energy’s crucial role in global sustainability necessitates efficient wind turbine maintenance, traditionally hindered by labor-intensive, risky manual inspections. UAV-based inspections offer improvements yet often lack adaptability to dynamic conditions like blade pitch and wind. To overcome these limitations and enhance inspection efficacy, we introduce the Dynamic Trajectory Adaptation Method (DyTAM), a novel approach for automated wind turbine inspections using UAVs. Within the proposed DyTAM, real-time image segmentation identifies key turbine components—blades, tower, and nacelle—from the initial viewpoint. Subsequently, the system dynamically computes blade pitch angles, classifying them into acute, vertical, and horizontal tilts. Based on this classification, DyTAM employs specialized, parameterized trajectory models—spiral, helical, and offset-line paths—tailored for each component and blade orientation. DyTAM allows for cutting total inspection time by 78% over manual approaches, decreasing path length by 17%, and boosting blade coverage by 6%. Field trials at a commercial site under challenging wind conditions show that deviations from planned trajectories are lowered by 68%. By integrating advanced path models (spiral, helical, and offset-line) with robust optical sensing, the DyTAM-based system streamlines the inspection process and ensures high-quality data capture. The dynamic adaptation is achieved through a closed-loop control system where real-time visual data from the UAV’s camera is continuously processed to update the flight trajectory on the fly, ensuring optimal inspection angles and distances are maintained regardless of blade position or external disturbances. The proposed method is scalable and can be extended to multi-UAV scenarios, laying a foundation for future efforts in real-time, large-scale wind infrastructure monitoring.
An Adaptive Machine Learning Approach to Sustainable Traffic Planning: High-Fidelity Pattern Recognition in Smart Transportation Systems
Effective and sustainable planning for future smart transportation systems is hindered by outdated traffic management models that fail to capture real-world dynamics, leading to congestion and significant environmental impact. To address this, advanced machine learning models are required to provide high-fidelity insights into urban mobility. In this work, we propose an adaptive machine learning approach to traffic pattern recognition that synergizes the HDBSCAN and k-means clustering algorithms. By employing a data-driven weighted voting mechanism, our solution provides a robust analytical foundation for sustainable planning, integrating structural analysis with precise cluster refinement. The crafted model was validated using a high-fidelity simulation of the Khmelnytskyi, Ukraine, transport network, where it demonstrated a superior ability to identify distinct traffic modes, achieving a V-measure of 0.79–0.82 and improving cluster compactness by 10–14% over standalone algorithms. It also attained a scenario identification accuracy of 92.8–95.0% with a temporal coherence of 0.94. These findings confirm that our adaptive approach is a foundational technology for intelligent transport systems, enabling the planning and deployment of more responsive, efficient, and sustainable urban mobility solutions.
Towards Transparent AI in Medicine: ECG-Based Arrhythmia Detection with Explainable Deep Learning
Cardiovascular diseases are the leading cause of death globally, highlighting the need for accurate diagnostic tools. To address this issue, we introduce a novel approach for arrhythmia detection based on electrocardiogram (ECG) that incorporates explainable artificial intelligence through three key methods. First, we developed an enhanced R peak detection method that integrates domain-specific knowledge into the ECG, improving peak identification accuracy by accounting for the characteristic features of R peaks. Second, we proposed an arrhythmia classification method utilizing a modified convolutional neural network (CNN) architecture with additional convolutional and batch normalization layers. This model processes a triad of cardio cycles—the preceding, current, and following cycles—to capture temporal dependencies and hidden features related to arrhythmias. Third, we implemented an interpretation method that explains CNN’s decisions using clinically relevant features, making the results understandable to clinicians. Using the MIT-BIH database, our approach achieved an accuracy of 99.43%, with F1-scores approaching 100% for major arrhythmia classes. The integration of these methods enhances both the performance and transparency of arrhythmia detection systems.