Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
1,690
result(s) for
"3D segmentation"
Sort by:
A Point-Wise LiDAR and Image Multimodal Fusion Network (PMNet) for Aerial Point Cloud 3D Semantic Segmentation
by
Ryosuke Nakamura
,
Vinayaraj Poliyapram
,
Weimin Wang
in
Colleges & universities
,
Datasets
,
Deep learning
2019
3D semantic segmentation of point cloud aims at assigning semantic labels to each point by utilizing and respecting the 3D representation of the data. Detailed 3D semantic segmentation of urban areas can assist policymakers, insurance companies, governmental agencies for applications such as urban growth assessment, disaster management, and traffic supervision. The recent proliferation of remote sensing techniques has led to producing high resolution multimodal geospatial data. Nonetheless, currently, only limited technologies are available to fuse the multimodal dataset effectively. Therefore, this paper proposes a novel deep learning-based end-to-end Point-wise LiDAR and Image Multimodal Fusion Network (PMNet) for 3D segmentation of aerial point cloud by fusing aerial image features. PMNet respects basic characteristics of point cloud such as unordered, irregular format and permutation invariance. Notably, multi-view 3D scanned data can also be trained using PMNet since it considers aerial point cloud as a fully 3D representation. The proposed method was applied on two datasets (1) collected from the urban area of Osaka, Japan and (2) from the University of Houston campus, USA and its neighborhood. The quantitative and qualitative evaluation shows that PMNet outperforms other models which use non-fusion and multimodal fusion (observational-level fusion and feature-level fusion) strategies. In addition, the paper demonstrates the improved performance of the proposed model (PMNet) by over-sampling/augmenting the medium and minor classes in order to address the class-imbalance issues.
Journal Article
A point cloud segmentation network with hybrid convolution and differential channels
2025
In recent years, point-based segmentation methods have made significant progress in improving segmentation accuracy. However, existing approaches still suffer from several key limitations. Traditional convolution operations make it difficult to effectively model the irregular geometry in point cloud data, resulting in insufficient sensitivity to spatial details. Additionally, current methods have limitations in collaboratively modeling and integrating global and local information. For this reason, we propose a 3D segmentation network based on hybrid convolution and differential channels. Specifically, we design a hybrid convolutional feature extraction (HCFE) module for processing 3D semantic information and spatial information independently, using different convolution kernels to obtain the subtle geometric structure differences between points. Then, we propose a Differential Channel Feature Interaction (DCFI) Module to enhance the local details and global channel information through Differential Convolution (DCU) and a Simplified Channel Attention Mechanism (S_ECA), respectively, and adaptively fuse the two types of information by Dynamic Interaction Mechanism (DIM), achieving their cooperative optimization. Compared to existing methods, HDC_Net has clear advantages in detail capturing and the integration of local and global information. A large number of experiments demonstrate the effectiveness and superiority of the model proposed in this study.
Journal Article
A Deep Learning-Based Workflow for Dendritic Spine Segmentation
by
Martin-Abadal, Miguel
,
García-Lorenzo, Marcos
,
Cosmin-Toader, Nicusor
in
Algorithms
,
Alzheimer's disease
,
Anatomy
2022
The morphological analysis of dendritic spines is an important challenge for the neuroscientific community. Most state-of-the-art techniques rely on user-supervised algorithms to segment the spine surface, especially those designed for light microscopy images. Therefore, processing large dendritic branches is costly and time-consuming. Although deep learning (DL) models have become one of the most commonly used tools in image segmentation, they have not yet been successfully applied to this problem. In this paper, we study the feasibility of using DL models to automatize spine segmentation from confocal microscopy images. Supervised learning is the most frequently used method for training DL models. This approach requires large data sets of high-quality segmented images (ground truth). As mentioned above, the segmentation of microscopy images is time-consuming and, therefore, in most cases, neuroanatomists only reconstruct relevant branches of the stack. Additionally, some parts of the dendritic shaft and spines are not segmented due to dyeing problems. In the context of this research, we tested the most successful architectures in the DL biomedical segmentation field. To build the ground truth, we used a large and high-quality data set, according to standards in the field. Nevertheless, this data set is not sufficient to train convolutional neural networks for accurate reconstructions. Therefore, we implemented an automatic preprocessing step and several training strategies to deal with the problems mentioned above. As shown by our results, our system produces a high-quality segmentation in most cases. Finally, we integrated several postprocessing user-supervised algorithms in a graphical user interface application to correct any possible artifacts.
Journal Article
DEHA-Net: A Dual-Encoder-Based Hard Attention Network with an Adaptive ROI Mechanism for Lung Nodule Segmentation
2023
Measuring pulmonary nodules accurately can help the early diagnosis of lung cancer, which can increase the survival rate among patients. Numerous techniques for lung nodule segmentation have been developed; however, most of them either rely on the 3D volumetric region of interest (VOI) input by radiologists or use the 2D fixed region of interest (ROI) for all the slices of computed tomography (CT) scan. These methods only consider the presence of nodules within the given VOI, which limits the networks’ ability to detect nodules outside the VOI and can also encompass unnecessary structures in the VOI, leading to potentially inaccurate segmentation. In this work, we propose a novel approach for 3D lung nodule segmentation that utilizes the 2D region of interest (ROI) inputted from a radiologist or computer-aided detection (CADe) system. Concretely, we developed a two-stage lung nodule segmentation technique. Firstly, we designed a dual-encoder-based hard attention network (DEHA-Net) in which the full axial slice of thoracic computed tomography (CT) scan, along with an ROI mask, were considered as input to segment the lung nodule in the given slice. The output of DEHA-Net, the segmentation mask of the lung nodule, was inputted to the adaptive region of interest (A-ROI) algorithm to automatically generate the ROI masks for the surrounding slices, which eliminated the need for any further inputs from radiologists. After extracting the segmentation along the axial axis, at the second stage, we further investigated the lung nodule along sagittal and coronal views by employing DEHA-Net. All the estimated masks were inputted into the consensus module to obtain the final volumetric segmentation of the nodule. The proposed scheme was rigorously evaluated on the lung image database consortium and image database resource initiative (LIDC/IDRI) dataset, and an extensive analysis of the results was performed. The quantitative analysis showed that the proposed method not only improved the existing state-of-the-art methods in terms of dice score but also showed significant robustness against different types, shapes, and dimensions of the lung nodules. The proposed framework achieved the average dice score, sensitivity, and positive predictive value of 87.91%, 90.84%, and 89.56%, respectively.
Journal Article
MVTN: Learning Multi-view Transformations for 3D Understanding
by
AlZahrani, Faisal
,
Hamdi, Abdullah
,
Giancola, Silvio
in
Artificial Intelligence
,
Classification
,
Computer Imaging
2025
Multi-view projection techniques have shown themselves to be highly effective in achieving top-performing results in the recognition of 3D shapes. These methods involve learning how to combine information from multiple view-points. However, the camera view-points from which these views are obtained are often fixed for all shapes. To overcome the static nature of current multi-view techniques, we propose learning these view-points. Specifically, we introduce the Multi-View Transformation Network (MVTN), which uses differentiable rendering to determine optimal view-points for 3D shape recognition. As a result, MVTN can be trained end-to-end with any multi-view network for 3D shape classification. We integrate MVTN into a novel adaptive multi-view pipeline that is capable of rendering both 3D meshes and point clouds. Our approach demonstrates state-of-the-art performance in 3D classification and shape retrieval on several benchmarks (ModelNet40, ScanObjectNN, ShapeNet Core55). Further analysis indicates that our approach exhibits improved robustness to occlusion compared to other methods. We also investigate additional aspects of MVTN, such as 2D pretraining and its use for segmentation. To support further research in this area, we have released MVTorch, a PyTorch library for 3D understanding and generation using multi-view projections.
Journal Article
Improved Point Cloud Representation via a Learnable Sort–Mix–Attend Mechanism
2026
Recent years have seen remarkable progress in deep learning on 3D point clouds, with hierarchical architectures becoming standard. Most work has focused on developing increasingly complex operators, such as self-attention, while enhancing the representational capacity of efficient point-wise MLP-based backbones has received less attention. We address this issue by proposing a differentiable module that learns to impose a task-driven canonical structure on local point sets. Our proposed SMA (Sort–Mix–Attend) layer dynamically serializes a neighborhood by generating a geometric basis and using a differentiable sorting mechanism. This enables an efficient MLP-based network to model rich feature interactions, adaptively modulating features prior to the final symmetric aggregation function. We demonstrate that SMA effectively enhances standard backbones for 3D classification and segmentation. Specifically, integrating SMA into PointNeXt-S achieves an Overall Accuracy (OA) of 88.3% on the challenging ScanObjectNN dataset, an improvement of 0.6% over the baseline. Furthermore, it boosts the classic PointNet++ architecture by a significant 5.2% in OA. We also introduce a highly efficient SMA-Tiny variant that achieves 86.0% OA with only 0.3 M parameters, proving the structural superiority, computational cost-effectiveness, and practical significance of our method for real-world 3D perception tasks.
Journal Article
LimeSeg: a coarse-grained lipid membrane simulation for 3D image segmentation
by
Machado, Sarah
,
Mercier, Vincent
,
Chiaruttini, Nicolas
in
3D segmentation
,
Algorithms
,
Bioinformatics
2019
Background
3D segmentation is often a prerequisite for 3D object display and quantitative measurements. Yet existing voxel-based methods do not directly give information on the object surface or topology. As for spatially continuous approaches such as level-set, active contours and meshes, although providing surfaces and concise shape description, they are generally not suitable for multiple object segmentation and/or for objects with an irregular shape, which can hamper their adoption by bioimage analysts.
Results
We developed LimeSeg, a computationally efficient and spatially continuous 3D segmentation method. LimeSeg is easy-to-use and can process many and/or highly convoluted objects. Based on the concept of SURFace ELements (“Surfels”), LimeSeg resembles a highly coarse-grained simulation of a lipid membrane in which a set of particles, analogous to lipid molecules, are attracted to local image maxima. The particles are self-generating and self-destructing thus providing the ability for the membrane to evolve towards the contour of the objects of interest.
The capabilities of LimeSeg: simultaneous segmentation of numerous non overlapping objects, segmentation of highly convoluted objects and robustness for big datasets are demonstrated on experimental use cases (epithelial cells, brain MRI and FIB-SEM dataset of cellular membrane system respectively).
Conclusion
In conclusion, we implemented a new and efficient 3D surface reconstruction plugin adapted for various sources of images, which is deployed in the user-friendly and well-known ImageJ environment.
Journal Article
Transformers for Neuroimage Segmentation: Scoping Review
by
Rustamov, Zahiriddin
,
Damseh, Rafat
,
Iratni, Maya
in
Brain
,
Brain - diagnostic imaging
,
Brain cancer
2025
Neuroimaging segmentation is increasingly important for diagnosing and planning treatments for neurological diseases. Manual segmentation is time-consuming, apart from being prone to human error and variability. Transformers are a promising deep learning approach for automated medical image segmentation.
This scoping review will synthesize current literature and assess the use of various transformer models for neuroimaging segmentation.
A systematic search in major databases, including Scopus, IEEE Xplore, PubMed, and ACM Digital Library, was carried out for studies applying transformers to neuroimaging segmentation problems from 2019 through 2023. The inclusion criteria allow only for peer-reviewed journal papers and conference papers focused on transformer-based segmentation of human brain imaging data. Excluded are the studies dealing with nonneuroimaging data or raw brain signals and electroencephalogram data. Data extraction was performed to identify key study details, including image modalities, datasets, neurological conditions, transformer models, and evaluation metrics. Results were synthesized using a narrative approach.
Of the 1246 publications identified, 67 (5.38%) met the inclusion criteria. Half of all included studies were published in 2022, and more than two-thirds used transformers for segmenting brain tumors. The most common imaging modality was magnetic resonance imaging (n=59, 88.06%), while the most frequently used dataset was brain tumor segmentation dataset (n=39, 58.21%). 3D transformer models (n=42, 62.69%) were more prevalent than their 2D counterparts. The most developed were those of hybrid convolutional neural network-transformer architectures (n=57, 85.07%), where the vision transformer is the most frequently used type of transformer (n=37, 55.22%). The most frequent evaluation metric was the Dice score (n=63, 94.03%). Studies generally reported increased segmentation accuracy and the ability to model both local and global features in brain images.
This review represents the recent increase in the adoption of transformers for neuroimaging segmentation, particularly for brain tumor detection. Currently, hybrid convolutional neural network-transformer architectures achieve state-of-the-art performances on benchmark datasets over standalone models. Nevertheless, their applicability remains highly limited by high computational costs and potential overfitting on small datasets. The heavy reliance of the field on the brain tumor segmentation dataset hints at the use of a more diverse set of datasets to validate the performances of models on a variety of neurological diseases. Further research is needed to define the optimal transformer architectures and training methods for clinical applications. Continuing development may make transformers the state-of-the-art for fast, accurate, and reliable brain magnetic resonance imaging segmentation, which could lead to improved clinical tools for diagnosing and evaluating neurological disorders.
Journal Article
An Experimental HBIM Processing: Innovative Tool for 3D Model Reconstruction of Morpho-Typological Phases for the Cultural Heritage
by
Barrile, Vincenzo
,
Bilotta, Giuliana
,
Bernardo, Ernesto
in
3D segmentation
,
Accuracy
,
Aerial photography
2022
In this paper, we want to propose an investigation and a re-reading of the “Conventazzo” of San Pietro di Deca in Torrenova (ME), through the use of geomatics techniques (laser scanner, UAV—Unmanned Aerial Vehicle-photogrammetry and BIM—Building Information Modeling) and a reconstruction and representation of different morpho-typological phases that highlight the numerous changes that this structure has undergone over the years. Particular attention was given to the BIM/HBIM (Heritage BIM) construction, bearing in mind that, in particular, the use of HBIM software for cultural heritage cannot perfectly represent old buildings with complex notable and particularly detailed architecture. Specifically, a new methodology is presented in order to replicate the complex details found in antique buildings, through the direct insertion of various 3D model parts (.obj) (point cloud segmentation from laser scanner and UAV/photogrammetry survey) into a BIM environment that includes intelligent objects linked to form the smart model. By having a huge amount of information available in a single digital model (HBIM), and by including all the information acquired during the survey campaign, it is possible to study the morphotypological evolutions of the building without the need to carry out subsequent survey campaigns. The limit of the proposed methodology, compared to the most used methodologies (despite the good results obtained), is that it requires the use of many types of software and is very slow. The proposed methodology was put to the test on the reconstruction of the “Conventazzo” in San Pietro di Deca, Torrenova (Messina).
Journal Article
Advancing Wound Filling Extraction on 3D Faces: An Auto-Segmentation and Wound Face Regeneration Approach
by
Le, Thinh D.
,
Nguyen-Xuan, H.
,
Nguyen, Phuong D.
in
Accuracy
,
Artificial neural networks
,
Automation
2024
Facial wound segmentation plays a crucial role in preoperative planning and optimizing patient outcomes in various medical applications. In this paper, we propose an efficient approach for automating 3D facial wound segmentation using a two-stream graph convolutional network. Our method leverages the Cir3D-FaIR dataset and addresses the challenge of data imbalance through extensive experimentation with different loss functions. To achieve accurate segmentation, we conducted thorough experiments and selected a high-performing model from the trained models. The selected model demonstrates exceptional segmentation performance for complex 3D facial wounds. Furthermore, based on the segmentation model, we propose an improved approach for extracting 3D facial wound fillers and compare it to the results of the previous study. Our method achieved a remarkable accuracy of 0.9999993% on the test suite, surpassing the performance of the previous method. From this result, we use 3D printing technology to illustrate the shape of the wound filling. The outcomes of this study have significant implications for physicians involved in preoperative planning and intervention design. By automating facial wound segmentation and improving the accuracy of wound-filling extraction, our approach can assist in carefully assessing and optimizing interventions, leading to enhanced patient outcomes. Additionally, it contributes to advancing facial reconstruction techniques by utilizing machine learning and 3D bioprinting for printing skin tissue implants. Our source code is available at https://github.com/SIMOGroup/WoundFilling3D.
Journal Article