Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
17
result(s) for
"Peng, Daifeng"
Sort by:
End-to-End Change Detection for High Resolution Satellite Images Using Improved UNet
by
Zhang, Yongjun
,
Peng, Daifeng
,
Guan, Haiyan
in
Architectural engineering
,
Change detection
,
Coders
2019
Change detection (CD) is essential to the accurate understanding of land surface changes using available Earth observation data. Due to the great advantages in deep feature representation and nonlinear problem modeling, deep learning is becoming increasingly popular to solve CD tasks in remote-sensing community. However, most existing deep learning-based CD methods are implemented by either generating difference images using deep features or learning change relations between pixel patches, which leads to error accumulation problems since many intermediate processing steps are needed to obtain final change maps. To address the above-mentioned issues, a novel end-to-end CD method is proposed based on an effective encoder-decoder architecture for semantic segmentation named UNet++, where change maps could be learned from scratch using available annotated datasets. Firstly, co-registered image pairs are concatenated as an input for the improved UNet++ network, where both global and fine-grained information can be utilized to generate feature maps with high spatial accuracy. Then, the fusion strategy of multiple side outputs is adopted to combine change maps from different semantic levels, thereby generating a final change map with high accuracy. The effectiveness and reliability of our proposed CD method are verified on very-high-resolution (VHR) satellite image datasets. Extensive experimental results have shown that our proposed approach outperforms the other state-of-the-art CD methods.
Journal Article
RSICDNet: A Novel Regional Scribble-Based Interactive Change Detection Network for Remote Sensing Images
2026
To address the issues of inadequate performance and excessive interaction costs when handling large-scale and complex-shaped change areas with existing interaction forms, this paper proposes RSICDNet, an interactive change detection (ICD) model with regional scribble interaction. In this framework, regional scribble interaction is introduced for the first time to provide rich spatial prior information for accurate ICD. Specifically, RSICDNet first employs an interaction processing network to extract interactive features, and subsequently utilizes the High-Resolution Network (HRNet) backbone to extract features from bi-temporal remote sensing images concatenated along the channel dimension. To effectively integrate these two information streams, an Interaction Fusion and Refinement Module (IFRM) is proposed, which injects the spatial priors from the interactive features into the high-level semantic features. Finally, an Object Contextual Representation (OCR) module is applied to further refine feature representations, and a lightweight segmentation head is used to generate final change map. Furthermore, a human–computer ICD application has been developed based on RSICDNet, significantly enhancing its potential for practical deployment. To validate the effectiveness of the proposed RSICDNet, extensive experiments are conducted against mainstream interactive deep learning models on the WHU-CD, LEVIR-CD, and CLCD datasets. The quantitative results demonstrate that RSICDNet achieves optimal Number of Interactions (NoI) metrics across all three datasets. Specifically, its NoI80 values reach 1.15, 1.45, and 3.42 on the WHU-CD, LEVIR-CD, and CLCD datasets, respectively. The qualitative results confirm a clear advantage for RSICDNet, which consistently delivers visually superior outcomes using the same or often fewer interactions.
Journal Article
RACDNet: Resolution- and Alignment-Aware Change Detection Network for Optical Remote Sensing Imagery
2022
Change detection (CD) methods work on the basis of co-registered multi-temporal images with equivalent resolutions. Due to the limitation of sensor imaging conditions and revisit period, it is difficult to acquire the desired images, especially in emergency situations. In addition, accurate multi-temporal images co-registration is largely limited by vast object changes and matching algorithms. To this end, a resolution- and alignment-aware change detection network (RACDNet) is proposed for multi-resolution optical remote-sensing imagery CD. In the first stage, to generate high-quality bi-temporal images, a light-weighted super-resolution network is proposed by fully considering the construction difficulty of different regions, which facilitates to detailed information recovery. Adversarial loss and perceptual loss are further adopted to improve the visual quality. In the second stage, deformable convolution units are embedded in a novel Siamese–UNet architecture for bi-temporal deep features alignment; thus, robust difference features can be generated for change information extraction. We further use an atrous convolution module to enlarge the receptive field, and an attention module to bridge the semantic gap between the encoder and decoder. To verify the effectiveness of our RACDNet, a novel multi-resolution change detection dataset (MRCDD) is created by using Google Earth. The quantitative and qualitative experimental results demonstrate that our RACDNet is capable of enhancing the details of the reconstructed images significantly, and the performance of CD surpasses other state-of-the-art methods by a large margin.
Journal Article
Perturbation Matters: A Novel Approach for Semi-Supervised Remote Sensing Imagery Change Detection
2025
Due to the challenge of acquiring abundant labeled samples, semi-supervised change detection (SSCD) approaches are becoming increasingly popular in tackling CD tasks with limited labeled data. Despite their success, these methods tend to come with complex network architectures or cumbersome training procedures, which also ignore the domain gap between the labeled data and unlabeled data. Differently, we hypothesize that diverse perturbations are more favorable to exploit the potential of unlabeled data. In light of this spirit, we propose a novel SSCD approach based on Weak–strong Augmentation and Class-balanced Sampling (WACS-SemiCD). Specifically, we adopt a simple mean-teacher architecture to deal with labeled branch and unlabeled branch separately, where supervised learning is conducted on the labeled branch, while weak–strong consistency learning (e.g., sample perturbations’ consistency and feature perturbations’ consistency) is imposed for the unlabeled. To improve domain generalization capacity, an adaptive CutMix augmentation is proposed to inject the knowledge from the labeled data into the unlabeled data. A class-balanced sampling strategy is further introduced to mitigate class imbalance issues in CD. Particularly, our proposed WACS-SemiCD achieves competitive SSCD performance on three publicly available CD datasets under different labeled settings. Comprehensive experimental results and systematic analysis underscore the advantages and effectiveness of our proposed WACS-SemiCD.
Journal Article
Robust Feature Matching with Spatial Smoothness Constraints
2020
Feature matching is to detect and match corresponding feature points in stereo pairs, which is one of the key techniques in accurate camera orientations. However, several factors limit the feature matching accuracy, e.g., image textures, viewing angles of stereo cameras, and resolutions of stereo pairs. To improve the feature matching accuracy against these limiting factors, this paper imposes spatial smoothness constraints over the whole feature point sets with the underlying assumption that feature points should have similar matching results with their surrounding high-confidence points and proposes a robust feature matching method with the spatial smoothness constraints (RMSS). The core algorithm constructs a graph structure from the feature point sets and then formulates the feature matching problem as the optimization of a global energy function with first-order, spatial smoothness constraints based on the graph. For computational purposes, the global optimization of the energy function is then broken into sub-optimizations of each feature point, and an approximate solution of the energy function is iteratively derived as the matching results of the whole feature point sets. Experiments on close-range datasets with some above limiting factors show that the proposed method was capable of greatly improving the matching robustness and matching accuracy of some feature descriptors (e.g., scale-invariant feature transform (SIFT) and Speeded Up Robust Features (SURF)). After the optimization of the proposed method, the inlier number of SIFT and SURF was increased by average 131.9% and 113.5%, the inlier percentages between the inlier number and the total matches number of SIFT and SURF were increased by average 259.0% and 307.2%, and the absolute matching accuracy of SIFT and SURF was improved by average 80.6% and 70.2%.
Journal Article
Object-Oriented Change Detection Method Based on Spectral–Spatial–Saliency Change Information and Fuzzy Integral Decision Fusion for HR Remote Sensing Images
2022
Spectral features in remote sensing images are extensively utilized to detect land cover changes. However, detection noise appearing in the changing maps due to the abundant spatial details in the high-resolution images makes it difficult to acquire an accurate interpretation result. In this paper, an object-oriented change detection approach is proposed which integrates spectral–spatial–saliency change information and fuzzy integral decision fusion for high-resolution remote sensing images with the purpose of eliminating the impact of detection noise. First, to reduce the influence of feature uncertainty, spectral feature change is generated by three independent methods, and spatial change information is obtained by spatial feature set construction and the optimal feature selection strategy. Secondly, the saliency change map of bi-temporal images is obtained with the co-saliency detection method to complement the insufficiency of image features. Then, the image objects are acquired by multi-scale segmentation based on the staking images. Finally, different pixel-level image change information and the segmentation result are fused using the fuzzy integral decision theory to determine the object change probability. Three high-resolution remote sensing image datasets and three comparative experiments were carried out to evaluate the performance of the proposed algorithm. Spectral–spatial–saliency change information was found to play a major role in the change detection of high-resolution remote sensing images, and the fuzzy integral decision strategy was found to effectively obtain reliable changed objects to improve the accuracy and robustness of change detection.
Journal Article
Affine-Function Transformation-Based Object Matching for Vehicle Detection from Unmanned Aerial Vehicle Imagery
by
Guan, Haiyan
,
Peng, Daifeng
,
Yu, Yongtao
in
Classification
,
Comparative studies
,
comparative study
2019
Vehicle detection from remote sensing images plays a significant role in transportation related applications. However, the scale variations, orientation variations, illumination variations, and partial occlusions of vehicles, as well as the image qualities, bring great challenges for accurate vehicle detection. In this paper, we present an affine-function transformation-based object matching framework for vehicle detection from unmanned aerial vehicle (UAV) images. First, meaningful and non-redundant patches are generated through a superpixel segmentation strategy. Then, the affine-function transformation-based object matching framework is applied to a vehicle template and each of the patches for vehicle existence estimation. Finally, vehicles are detected and located after matching cost thresholding, vehicle location estimation, and multiple response elimination. Quantitative evaluations on two UAV image datasets show that the proposed method achieves an average completeness, correctness, quality, and F1-measure of 0.909, 0.969, 0.883, and 0.938, respectively. Comparative studies also demonstrate that the proposed method achieves compatible performance with the Faster R-CNN and outperforms the other eight existing methods in accurately detecting vehicles of various conditions.
Journal Article
BUILDING CHANGE DETECTION BY COMBINING LiDAR DATA AND ORTHO IMAGE
2016
The elevation information is not considered in the traditional building change detection methods. This paper presents an algorithm of combining LiDAR data and ortho image for 3D building change detection. The advantages of the proposed approach lie in the fusion of the height and spectral information by thematic segmentation. Furthermore, the proposed method also combines the advantages of pixel-level and object-level change detection by image differencing and object analysis. Firstly, two periods of LiDAR data are filtered and interpolated to generate their corresponding DSMs. Secondly, a binary image of the changed areas is generated by means of differencing and filtering the two DSMs, and then thematic layer is generated and projected onto the DSMs and DOMs. Thirdly, geometric and spectral features of the changed area are calculated, which is followed by decision tree classification for the purpose of extracting the changed building areas. Finally, the statistics of the elevation and area change information as well as the change type of the changed buildings are done for building change analysis. Experimental results show that the completeness and correctness of building change detection are close to 81.8% and 85.7% respectively when the building area is larger than 80 m2, which are increased about 10% when compared with using ortho image alone.
Journal Article
BUILDING CHANGE DETECTION BY COMBINING LiDAR DATA AND ORTHO IMAGE
2016
The elevation information is not considered in the traditional building change detection methods. This paper presents an algorithm of combining LiDAR data and ortho image for 3D building change detection. The advantages of the proposed approach lie in the fusion of the height and spectral information by thematic segmentation. Furthermore, the proposed method also combines the advantages of pixel-level and object-level change detection by image differencing and object analysis. Firstly, two periods of LiDAR data are filtered and interpolated to generate their corresponding DSMs. Secondly, a binary image of the changed areas is generated by means of differencing and filtering the two DSMs, and then thematic layer is generated and projected onto the DSMs and DOMs. Thirdly, geometric and spectral features of the changed area are calculated, which is followed by decision tree classification for the purpose of extracting the changed building areas. Finally, the statistics of the elevation and area change information as well as the change type of the changed buildings are done for building change analysis. Experimental results show that the completeness and correctness of building change detection are close to 81.8% and 85.7% respectively when the building area is larger than 80 m2, which are increased about 10% when compared with using ortho image alone.
Journal Article
Adapting Segment Anything Model for Change Detection in HR Remote Sensing Images
2024
Vision Foundation Models (VFMs) such as the Segment Anything Model (SAM) allow zero-shot or interactive segmentation of visual contents, thus they are quickly applied in a variety of visual scenes. However, their direct use in many Remote Sensing (RS) applications is often unsatisfactory due to the special imaging characteristics of RS images. In this work, we aim to utilize the strong visual recognition capabilities of VFMs to improve the change detection of high-resolution Remote Sensing Images (RSIs). We employ the visual encoder of FastSAM, an efficient variant of the SAM, to extract visual representations in RS scenes. To adapt FastSAM to focus on some specific ground objects in the RS scenes, we propose a convolutional adaptor to aggregate the task-oriented change information. Moreover, to utilize the semantic representations that are inherent to SAM features, we introduce a task-agnostic semantic learning branch to model the semantic latent in bi-temporal RSIs. The resulting method, SAMCD, obtains superior accuracy compared to the SOTA methods and exhibits a sample-efficient learning ability that is comparable to semi-supervised CD methods. To the best of our knowledge, this is the first work that adapts VFMs for the CD of HR RSIs.