Query:
学者姓名:李蒙蒙
Refining:
Year
Type
Indexed by
Source
Complex
Former Name
Co-
Language
Clean All
Abstract :
Change detection is a fundamental yet challenging task in remote sensing, crucial for monitoring urban expansion, land use changes, and environmental dynamics. However, compared with common color images, objects in remote sensing images exhibit minimal interclass variation and significant intraclass variation in the spectral dimension, with obvious scale inconsistency in the spatial dimension. Change detection complexity presents significant challenges, including differentiating similar objects, accounting for scale variations, and identifying pseudo changes. This research introduces a dual fine-grained network with a frequency Transformer (named as FTransDF-Net) to address the above issues. Specifically, for small-scale and approximate spectral ground objects, the network employs an encoder-decoder architecture consisting of dual fine-grained gated (DFG) modules. This enables the extraction and fusion of fine-grained level information in dual dimensions of features, facilitating a comprehensive analysis of their differences and correlations. As a result, a dynamic fusion representation of salient information is achieved. Additionally, we develop a lightweight frequency transformer (LFT) with minimal parameters for detecting large-scale ground objects that undergo significant changes over time. This is achieved by incorporating a frequency attention (FA) module, which utilizes Fourier transform to model long-range dependencies and combines global adaptive attentive features with multi-level fine-grained features. Our comparative experiments across four publicly available datasets demonstrate that FTransDF-Net reaches advanced results. Importantly, it outperforms the leading comparison method by 1.23% and 2.46% regarding IoU metrics concerning CDD and DSIFN, respectively. Furthermore, efficacy for each module is substantiated through ablation experiments. The code is accessible on https://github.com/LeeThrzz/FTrans-DF-Net.
Keyword :
Change detection Change detection Dual fine-grained Dual fine-grained Frequency transformer Frequency transformer Remote sensing Remote sensing
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Li, Zhen , Zhang, Zhenxin , Li, Mengmeng et al. Dual Fine-Grained network with frequency Transformer for change detection on remote sensing images [J]. | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION , 2025 , 136 . |
MLA | Li, Zhen et al. "Dual Fine-Grained network with frequency Transformer for change detection on remote sensing images" . | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION 136 (2025) . |
APA | Li, Zhen , Zhang, Zhenxin , Li, Mengmeng , Zhang, Liqiang , Peng, Xueli , He, Rixing et al. Dual Fine-Grained network with frequency Transformer for change detection on remote sensing images . | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION , 2025 , 136 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Building type information indicates the functional properties of buildings and plays a crucial role in smart city development and urban socioeconomic activities. Existing methods for classifying building types often face challenges in accurately distinguishing buildings between types while maintaining well-delineated boundaries, especially in complex urban environments. This study introduces a novel framework, i.e., CNN-Transformer cross-attention feature fusion network (CTCFNet), for building type classification from very high resolution remote sensing images. CTCFNet integrates convolutional neural networks (CNNs) and Transformers using an interactive cross-encoder fusion module that enhances semantic feature learning and improves classification accuracy in complex scenarios. We develop an adaptive collaboration optimization module that applies human visual attention mechanisms to enhance the feature representation of building types and boundaries simultaneously. To address the scarcity of datasets in building type classification, we create two new datasets, i.e., the urban building type (UBT) dataset and the town building type (TBT) dataset, for model evaluation. Extensive experiments on these datasets demonstrate that CTCFNet outperforms popular CNNs, Transformers, and dual-encoder methods in identifying building types across various regions, achieving the highest mean intersection over union of 78.20% and 77.11%, F1 scores of 86.83% and 88.22%, and overall accuracy of 95.07% and 95.73% on the UBT and TBT datasets, respectively. We conclude that CTCFNet effectively addresses the challenges of high interclass similarity and intraclass inconsistency in complex scenes, yielding results with well-delineated building boundaries and accurate building types.
Keyword :
Accuracy Accuracy Architecture Architecture Buildings Buildings Building type classification Building type classification CNN-transformer networks CNN-transformer networks cross-encoder cross-encoder Earth Earth Feature extraction Feature extraction feature interaction feature interaction Optimization Optimization Remote sensing Remote sensing Semantics Semantics Transformers Transformers very high resolution remote sensing very high resolution remote sensing Visualization Visualization
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Zhang, Shaofeng , Li, Mengmeng , Zhao, Wufan et al. Building Type Classification Using CNN-Transformer Cross-Encoder Adaptive Learning From Very High Resolution Satellite Images [J]. | IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING , 2025 , 18 : 976-994 . |
MLA | Zhang, Shaofeng et al. "Building Type Classification Using CNN-Transformer Cross-Encoder Adaptive Learning From Very High Resolution Satellite Images" . | IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 18 (2025) : 976-994 . |
APA | Zhang, Shaofeng , Li, Mengmeng , Zhao, Wufan , Wang, Xiaoqin , Wu, Qunyong . Building Type Classification Using CNN-Transformer Cross-Encoder Adaptive Learning From Very High Resolution Satellite Images . | IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING , 2025 , 18 , 976-994 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Building extraction from very high-resolution remote-sensing images still faces two main issues: (1) small buildings are severely omitted and the extracted building shapes have a low consistency with ground truths. (2) supervised deep-learning methods have poor performance in few-shot scenarios, limiting the practical application of these methods. To address the first issue, we propose an asymmetric Siamese multitask network integrating adversarial edge learning called ASMBR-Net for building extraction. It contains an efficient asymmetric Siamese feature extractor comprising pre-trained backbones of convolutional neural networks and Transformers under pre-training and fine-tuning paradigms. This extractor balances the local and global feature representation and reduces training costs. Adversarial edge-learning technology automatically integrates edge constraints and strengthens the modeling ability of small and complex building-shaped patterns. Aiming to overcome the second issue, we introduce a self-training framework and design an instance transfer strategy to generate reliable pseudo-samples. We examined the proposed method on the WHU and Massachusetts (MA) datasets and a self-constructed Dongying (DY) dataset, comparing it with state-of-the-art methods. The experimental results show that our method achieves the highest F1-score of 96.06%, 86.90%, and 84.98% on the WHU, MA, and DY datasets, respectively. Ablation experiments further verify the effectiveness of the proposed method. The code is available at: https://github.com/liuxuanguang/ASMBR-Net
Keyword :
Adversarial learning Adversarial learning Building extraction Building extraction Multitask learning Multitask learning Self-learning Self-learning VHR remote-sensing image VHR remote-sensing image
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Liu, Xuanguang , Li, Yujie , Dai, Chenguang et al. Extraction buildings from very high-resolution images with asymmetric siamese multitask networks and adversarial edge learning [J]. | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION , 2025 , 136 . |
MLA | Liu, Xuanguang et al. "Extraction buildings from very high-resolution images with asymmetric siamese multitask networks and adversarial edge learning" . | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION 136 (2025) . |
APA | Liu, Xuanguang , Li, Yujie , Dai, Chenguang , Zhang, Zhenchao , Ding, Lei , Li, Mengmeng et al. Extraction buildings from very high-resolution images with asymmetric siamese multitask networks and adversarial edge learning . | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION , 2025 , 136 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Heterogeneous change detection is a task of considerable practical importance and significant challenge in remote sensing. Heterogeneous change detection involves identifying change areas using remote sensing images obtained from different sensors or imaging conditions. Recently, research has focused on feature space translation methods based on deep learning technology for heterogeneous images. However, these types of methods often lead to the loss of original image information, and the translated features cannot be efficiently compared, further limiting the accuracy of change detection. For these issues, we propose a cross-modal feature interaction network (CMFINet). Specifically, CMFINet introduces a cross-modal interaction module (CMIM), which facilitates the interaction between heterogeneous features through attention exchange. This approach promotes consistent representation of heterogeneous features while preserving image characteristics. Additionally, we design a differential feature extraction module (DFEM) to enhance the extraction of true change features from spatial and channel dimensions, facilitating efficient comparison after feature interaction. Extensive experiments conducted on the California, Toulouse, and Wuhan datasets demonstrate that CMFINet outperforms eight existing methods in identifying change areas in different scenes from multimodal images. Compared to the existing methods applied to the three datasets, CMFINet achieved the highest F1 scores of 83.93%, 75.65%, and 95.42%, and the highest mIoU values of 85.38%, 78.34%, and 94.87%, respectively. The results demonstrate the effectiveness and applicability of CMFINet in heterogeneous change detection.
Keyword :
attention mechanisms attention mechanisms Change detection Change detection CNN CNN feature interaction feature interaction heterogeneous remote sensing images heterogeneous remote sensing images
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Yang, Zhiwei , Wang, Xiaoqin , Lin, Haihan et al. Cross-modal feature interaction network for heterogeneous change detection [J]. | GEO-SPATIAL INFORMATION SCIENCE , 2025 . |
MLA | Yang, Zhiwei et al. "Cross-modal feature interaction network for heterogeneous change detection" . | GEO-SPATIAL INFORMATION SCIENCE (2025) . |
APA | Yang, Zhiwei , Wang, Xiaoqin , Lin, Haihan , Li, Mengmeng , Lin, Mengjing . Cross-modal feature interaction network for heterogeneous change detection . | GEO-SPATIAL INFORMATION SCIENCE , 2025 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Precise information on agricultural parcels is crucial for effective farm management, crop mapping, and monitoring. Current techniques often encounter difficulties in automatically delineating vectorized parcels from remote sensing images, especially in irregular-shaped areas, making it challenging to derive closed and vectorized boundaries. To address this, we treat parcel delineation as identifying valid parcel vertices from remote sensing images to generate parcel polygons. We introduce a Point-Line-Region interactive multitask network (PLR-Net) that jointly learns semantic features of parcel vertices, boundaries, and regions through point-, line-, and region-related subtasks within a multitask learning framework. We derived an attraction field map (AFM) to enhance the feature representation of parcel boundaries and improve the detection of parcel regions while maintaining high geometric accuracy. The point-related subtask focuses on learning features of parcel vertices to obtain preliminary vertices, which are then refined based on detected boundary pixels to derive valid parcel vertices for polygon generation. We designed a spatial and channel excitation module for feature interaction to enhance interactions between points, lines, and regions. Finally, the generated parcel polygons are refined using the Douglas-Peucker algorithm to regularize polygon shapes. We evaluated PLR-Net using high-resolution GF-2 satellite images from the Shandong, Xinjiang, and Sichuan provinces of China and medium-resolution Sentinel-2 images from The Netherlands. Results showed that our method outperformed existing state-of-the-art techniques (e.g., BsiNet, SEANet, and Hisup) in pixel- and object-based geometric accuracy across all datasets, achieving the highest IoU and polygonal average precision on GF2 datasets (e.g., 90.84% and 82.00% in Xinjiang) and on the Sentinel-2 dataset (75.86% and 47.1%). Moreover, when trained on the Xinjiang dataset, the model successfully transferred to the Shandong dataset, achieving an IoU score of 83.98%. These results demonstrate that PLR-Net is an accurate, robust, and transferable method suitable for extracting vectorized parcels from diverse regions and types of remote sensing images. The source codes of our model are available at https://github.com/mengmengli01/PLR-Net-demo/tree/main.
Keyword :
Agricultural parcel delineation Agricultural parcel delineation Multitask neural networks Multitask neural networks PLR-Net PLR-Net Point-line-region interactive Point-line-region interactive Vectorized parcels Vectorized parcels
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Li, Mengmeng , Lu, Chengwen , Lin, Mengjing et al. Extracting vectorized agricultural parcels from high-resolution satellite images using a Point-Line-Region interactive multitask model [J]. | COMPUTERS AND ELECTRONICS IN AGRICULTURE , 2025 , 231 . |
MLA | Li, Mengmeng et al. "Extracting vectorized agricultural parcels from high-resolution satellite images using a Point-Line-Region interactive multitask model" . | COMPUTERS AND ELECTRONICS IN AGRICULTURE 231 (2025) . |
APA | Li, Mengmeng , Lu, Chengwen , Lin, Mengjing , Xiu, Xiaolong , Long, Jiang , Wang, Xiaoqin . Extracting vectorized agricultural parcels from high-resolution satellite images using a Point-Line-Region interactive multitask model . | COMPUTERS AND ELECTRONICS IN AGRICULTURE , 2025 , 231 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Phenological information on crop growth aids in identifying crop types from remote sensing images, but its incorporation into classification models is insufficiently exploited, especially in deep learning frameworks. This study presents a new model, Phenological-Temporal-Spatial LSTM (PST-LSTM), for mapping tobacco planting areas in smallholder farmlands using time-series Sentinel-1 Synthetic Aperture Radar (SAR) images. The PSTLSTM model is built on a multi-modal learning framework that fuses phenological information with deep spatial-temporal features. We applied the model to extract tobacco planting areas in Ninghua, Pucheng, and Shanghang Counties, in Fujian Province, and Luoping County in Yunnan Province, China. We compared PSTLSTM with existing methods based on phenological rules and Dynamic Time Warping (DTW) methods, and analyzed its strength in feature fusion. Results showed that our model outperformed these methods, achieving an overall accuracy (OA) of 93.16% compared to 86.69% and 85.93% for the phenological rules and DTW methods, respectively, in the Ninghua area. PST-LSTM effectively integrated time-series data with phenological information derived from different strategies at the feature level and performed better than existing feature fusion methods (based upon fuzzy sets) at the decision level. It also demonstrated a better spatial transferability than other methods when applied to different study areas, achieving an OA of 90.95%, 91.41%, and 80.75% for the Pucheng, Shanghang, and Luoping areas, respectively, using training samples from Ninghua. We conclude that PST-LSTM can effectively extract tobacco planting areas in smallholder farming from time-series SAR images and has the potential for mapping other crop types.
Keyword :
LSTM) LSTM) Phenological knowledge Phenological knowledge Phenological-Temporal-Spatial LSTM (PST Phenological-Temporal-Spatial LSTM (PST Smallholder farming Smallholder farming Time -series SAR images Time -series SAR images Tobacco mapping Tobacco mapping
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Li, Mengmeng , Feng, Xiaomin , Belgiu, Mariana . Mapping tobacco planting areas in smallholder farmlands using Phenological-Spatial-Temporal LSTM from time-series Sentinel-1 SAR images [J]. | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION , 2024 , 129 . |
MLA | Li, Mengmeng et al. "Mapping tobacco planting areas in smallholder farmlands using Phenological-Spatial-Temporal LSTM from time-series Sentinel-1 SAR images" . | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION 129 (2024) . |
APA | Li, Mengmeng , Feng, Xiaomin , Belgiu, Mariana . Mapping tobacco planting areas in smallholder farmlands using Phenological-Spatial-Temporal LSTM from time-series Sentinel-1 SAR images . | INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION , 2024 , 129 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Semantic change detection (SCD) from very high-resolution (VHR) images involves two key challenges: 1) the global features of bitemporal images tend to be extracted insufficiently, leading to imprecise land cover semantic classification results; and 2) the detected changed objects exhibit ambiguous boundaries, resulting in low geometric accuracy. To address these two issues, we propose an SCD method called TBSCD-Net based on a multitask learning framework to simultaneously identify different types of semantic changes and regularize change boundaries. First, we construct a hybrid encoder combining transformer and convolutional neural network (CNN) (TCEncoder) to enhance the extraction of global context information. A bitemporal semantic linkage module (Bi-SLM) is embedded into the TCEncoder to enhance the semantic correlations between bitemporal images. Second, we introduce a boundary-region joint extractor based on Laplacian operators (LOBRE) to regularize the changed objects. We evaluated the effectiveness of the proposed method using the SECOND dataset and a Fuzhou GF-2 SCD dataset (FZ-SCD) and compared it with seven existing methods. The proposed method performed better than the other evaluated methods as it achieved 24.42% separation kappa (Sek) and 20.18% global total classification error (GTC) on the SECOND dataset and 23.10% Sek and 23.15% GTC on the FZ-SCD dataset. The results of ablation studies on the FZ-SCD dataset also verified the effectiveness of the developed modules for SCD.
Keyword :
Boundary regularization Boundary regularization Decoding Decoding Feature extraction Feature extraction Land surface Land surface Laplace equations Laplace equations multitask learning multitask learning semantic change detection (SCD) semantic change detection (SCD) Semantics Semantics Siamese neural network Siamese neural network Task analysis Task analysis Transformers Transformers very high-resolution (VHR) satellite images very high-resolution (VHR) satellite images
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Liu, Xuanguang , Dai, Chenguang , Zhang, Zhenchao et al. TBSCD-Net: A Siamese Multitask Network Integrating Transformers and Boundary Regularization for Semantic Change Detection From VHR Satellite Images [J]. | IEEE GEOSCIENCE AND REMOTE SENSING LETTERS , 2024 , 21 . |
MLA | Liu, Xuanguang et al. "TBSCD-Net: A Siamese Multitask Network Integrating Transformers and Boundary Regularization for Semantic Change Detection From VHR Satellite Images" . | IEEE GEOSCIENCE AND REMOTE SENSING LETTERS 21 (2024) . |
APA | Liu, Xuanguang , Dai, Chenguang , Zhang, Zhenchao , Li, Mengmeng , Wang, Hanyun , Ji, Hongliang et al. TBSCD-Net: A Siamese Multitask Network Integrating Transformers and Boundary Regularization for Semantic Change Detection From VHR Satellite Images . | IEEE GEOSCIENCE AND REMOTE SENSING LETTERS , 2024 , 21 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Current land use classification models based on very high-resolution (VHR) remote sensing images often suffer from high sample dependence and poor transferability. To address these challenges, we propose an unsupervised multisource domain adaptation framework for cross-domain land use classification that eliminates the need for repeatedly using source domain data. Our method uses the Swin Transformer as the backbone of the source domain model to extract features from multiple source domain samples. The model is trained on source domain samples, and unlabeled target domain samples are then used for target domain model training. To minimize the feature discrepancies between the source and target domains, we use a weighted information maximization loss and self-supervised pseudolabels to alleviate cross-domain classification noise. We conducted experiments on four public scene datasets and four new land use scene datasets created from different VHR images in four Chinese cities. Results show that our method outperformed three existing single-source cross-domain methods (i.e., DANN, DeepCORAL, and DSAN) and four multisource cross-domain methods (i.e., M3SDA, PTMDA, MFSAN, and SHOT), achieving the highest average classification accuracy and strong stability. We conclude that our method has high potential for practical applications in cross-domain land use classification using VHR images.
Keyword :
Adaptation models Adaptation models Computational modeling Computational modeling Cross-domain classification Cross-domain classification Data models Data models Feature extraction Feature extraction land use classification land use classification multisource domain adaptation multisource domain adaptation Remote sensing Remote sensing Swin transformer Swin transformer Transformers Transformers Urban areas Urban areas very high resolution (VHR) remote sensing images very high resolution (VHR) remote sensing images
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Li, Mengmeng , Zhang, Congcong , Zhao, Wufan et al. Cross-Domain Urban Land Use Classification via Scenewise Unsupervised Multisource Domain Adaptation With Transformer [J]. | IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING , 2024 , 17 : 10051-10066 . |
MLA | Li, Mengmeng et al. "Cross-Domain Urban Land Use Classification via Scenewise Unsupervised Multisource Domain Adaptation With Transformer" . | IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 17 (2024) : 10051-10066 . |
APA | Li, Mengmeng , Zhang, Congcong , Zhao, Wufan , Zhou, Wen . Cross-Domain Urban Land Use Classification via Scenewise Unsupervised Multisource Domain Adaptation With Transformer . | IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING , 2024 , 17 , 10051-10066 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
遥感技术已成为快速有效获取农业大棚覆盖信息的重要途径,但遥感影像空间分辨率大小对提取精度的影响具有双重性,选择适宜分辨率影像具有重要意义。以南方农业塑料大棚为研究对象,利用GF-1、GF-2和Sentinel-2形成1~16 m间6个不同空间分辨率影像数据集,基于面向对象影像分析方法(Object-Based Image Analysis,OBIA),分别利用面向对象卷积神经网络(Convolutional Neural Network,CNN)方法和随机森林(Random forest,RF)方法开展大棚提取,分析提取精度和不同方法下的差异性。结果表明:(1)CNN和RF方法下,农业大棚的提取精度随着影像分辨率降低总体呈下降趋势,在1~16 m的影像上均能检测到农业大棚;(2)相对于RF方法,CNN方法对影像空间分辨率要求更高,在1~2 m分辨率下,CNN方法有更少的漏提和误提,但在4m及更低分辨率下,RF方法的适用性更高;(3)2 m分辨率影像是大棚信息提取的最佳空间分辨率,可经济有效地实现大棚监测。
Keyword :
农业大棚提取 农业大棚提取 空间分辨率 空间分辨率 随机森林 随机森林 面向对象CNN方法 面向对象CNN方法 高分辨率遥感数据 高分辨率遥感数据
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | 林欣怡 , 汪小钦 , 汤紫霞 et al. 基于面向对象CNN和RF的不同空间分辨率遥感影像农业大棚提取研究 [J]. | 遥感技术与应用 , 2024 , 39 (02) : 315-327 . |
MLA | 林欣怡 et al. "基于面向对象CNN和RF的不同空间分辨率遥感影像农业大棚提取研究" . | 遥感技术与应用 39 . 02 (2024) : 315-327 . |
APA | 林欣怡 , 汪小钦 , 汤紫霞 , 李蒙蒙 , 吴瑞姣 , 黄德华 . 基于面向对象CNN和RF的不同空间分辨率遥感影像农业大棚提取研究 . | 遥感技术与应用 , 2024 , 39 (02) , 315-327 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
建筑物变化检测在城市环境监测、土地规划管理和违章违规建筑识别等应用中具有重要作用。针对传统孪生神经网络在影像变化检测中存在的检测边界与实际边界吻合度低的问题,本文结合面向对象图像分析技术,提出一种基于面向对象孪生神经网络(Obj-SiamNet)的高分辨率遥感影像变化检测方法,利用模糊集理论自动融合多尺度变化检测结果,并通过生成对抗网络实现训练样本迁移。该方法应用在高分二号和高分七号高分辨率卫星影像中,并与基于时空自注意力的变化检测模型(STANet)、视觉变化检测网络(ChangeNet)和孪生UNet神经网络模型(Siam-NestedUNet)进行比较。结果表明:(1)融合面向对象多尺度分割的检测结果较单一尺度分割的检测结果,召回率最高提升32%,F1指数最高提升25%,全局总体误差(GTC)最高降低7%;(2)在样本数量有限的情况下,通过生成对抗网络进行样本迁移,与未使用样本迁移前的检测结果相比,召回率最高提升16%,F1指数最高提升14%,GTC降低了9%;(3) Obj-SiamNet方法较其他变化检测方法,整体检测精度得到提升,F1指数最高提升23%,GTC最高降低9%。该方法有效提高了建筑物变化检测在几何和属性方面的精度,并能有效利用开放地理数据集,降低了模型训练样本制作成本,提升了检测效率和适用性。
Keyword :
孪生神经网络 孪生神经网络 模糊集融合 模糊集融合 生成对抗网络 生成对抗网络 遥感变化检测 遥感变化检测 面向对象多尺度分析 面向对象多尺度分析 高分辨率遥感影像 高分辨率遥感影像
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | 刘宣广 , 李蒙蒙 , 汪小钦 et al. 基于面向对象孪生神经网络的高分辨率遥感影像建筑物变化检测 [J]. | 遥感学报 , 2024 , 28 (02) : 437-454 . |
MLA | 刘宣广 et al. "基于面向对象孪生神经网络的高分辨率遥感影像建筑物变化检测" . | 遥感学报 28 . 02 (2024) : 437-454 . |
APA | 刘宣广 , 李蒙蒙 , 汪小钦 , 张振超 . 基于面向对象孪生神经网络的高分辨率遥感影像建筑物变化检测 . | 遥感学报 , 2024 , 28 (02) , 437-454 . |
Export to | NoteExpress RIS BibTex |
Version :
Export
Results: |
Selected to |
Format: |