• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
成果搜索

author:

Wu, X. (Wu, X..) [1] | Huang, Z. (Huang, Z..) [2] | Wang, L. (Wang, L..) [3] | Chanussot, J. (Chanussot, J..) [4] | Tian, J. (Tian, J..) [5]

Indexed by:

Scopus

Abstract:

In large-scale disaster events, the planning of optimal rescue routes depends on the object detection ability at the disaster scene, with one of the main challenges being the presence of dense and occluded objects. Existing methods, which are typically based on the RGB modality, struggle to distinguish targets with similar colors and textures in crowded environments and are unable to identify obscured objects. To this end, we first construct two multimodal dense and occlusion vehicle detection datasets for large-scale events, utilizing RGB and height map modalities. Based on these datasets, we propose a multimodal collaboration network (MuDet) for dense and occluded vehicle detection, MuDet for short. MuDet hierarchically enhances the completeness of discriminable information within and across modalities and differentiates between simple and complex samples. MuDet includes three main modules: Unimodal Feature Hierarchical Enhancement (Uni-Enh), Multimodal Cross Learning (Mul-Lea), and Hard-easy Discriminative (He-Dis) Pattern. Uni-Enh and Mul-Lea enhance the features within each modality and facilitate the cross-integration of features from two heterogeneous modalities. He-Dis effectively separates densely occluded vehicle targets with significant intra-class differences and minimal inter-class differences by defining and thresholding confidence values, thereby suppressing the complex background. Experimental results on two re-labeled multimodal benchmark datasets, the 4K Stereo Aerial Imagery of a Large Camping Site (4K-SAI-LCS) dataset, and the ISPRS Potsdam dataset, demonstrate the robustness and generalization of the MuDet.  © 1980-2012 IEEE.

Keyword:

Convolutional neural networks (CNNs) dense and occluded hard-easy balanced attention large-scale disaster events multimodal vehicle detection (MVD) remote Sensing (RS)

Community:

  • [ 1 ] [Wu X.]Beijing University of Posts and Telecommunications, School of Computer Science, National Pilot Software Engineering School, Beijing, 100876, China
  • [ 2 ] [Huang Z.]Fuzhou University, Key Laboratory of Spatial Data Mining and Information Sharing of the Ministry of Education, The Academy of Digital China, The Natl. and Local Joint Engineering Research Center of Satellite Geospatial Information Technology, Fuzhou, 350108, China
  • [ 3 ] [Wang L.]Beijing University of Posts and Telecommunications, School of Computer Science, National Pilot Software Engineering School, Beijing, 100876, China
  • [ 4 ] [Chanussot J.]University of Grenoble Alpes, Cnrs, Grenoble Inp, GIPSA-Lab, Grenoble, 38000, France
  • [ 5 ] [Chanussot J.]Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, 100094, China
  • [ 6 ] [Tian J.]German Aerospace Center, Remote Sensing Technology Institute, Weßling, 82205, Germany

Reprint 's Address:

Email:

Show more details

Related Keywords:

Source :

IEEE Transactions on Geoscience and Remote Sensing

ISSN: 0196-2892

Year: 2024

Volume: 62

Page: 1-12

7 . 5 0 0

JCR@2023

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 0

Affiliated Colleges:

Online/Total:83/10133110
Address:FZU Library(No.2 Xuyuan Road, Fuzhou, Fujian, PRC Post Code:350116) Contact Us:0591-22865326
Copyright:FZU Library Technical Support:Beijing Aegean Software Co., Ltd. 闽ICP备05005463号-1