Query:
学者姓名:庄一新
Refining:
Year
Type
Indexed by
Source
Complex
Co-
Language
Clean All
Abstract :
Neural implicit representations are highly effective for single-view 3D reconstruction (SVR). It represents 3D shapes as neural fields and conditions shape prediction on input image features. Image features can be less effective when significant variations of occlusions, views, and appearances exist from the image. To learn more robust features, we design a new feature encoding scheme that works in both image and shape space. Specifically, we present a geometry-aware 2D convolutional kernel to learn image appearance and view information along with geometric relations. The convolutional kernel operates at the 2D projections of a point-based 3D geometric structure, called spatial pattern. Furthermore, to enable the network to discover adaptive spatial patterns that capture non-local contexts, the kernel is devised to be deformable and exploited by a spatial pattern generator. Experimental results on both synthetic and real datasets demonstrate the superiority of the proposed method. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023.
Keyword :
Convolution Convolution Geometry Geometry Image reconstruction Image reconstruction
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Zhuang, Yixin , Wang, Yujie , Liu, Yunzhe et al. Neural Implicit 3D Shapes from Single Images with Spatial Patterns [C] . 2023 : 210-227 . |
MLA | Zhuang, Yixin et al. "Neural Implicit 3D Shapes from Single Images with Spatial Patterns" . (2023) : 210-227 . |
APA | Zhuang, Yixin , Wang, Yujie , Liu, Yunzhe , Chen, Baoquan . Neural Implicit 3D Shapes from Single Images with Spatial Patterns . (2023) : 210-227 . |
Export to | NoteExpress RIS BibTex |
Version :
Abstract :
Visual (re)localization addresses the problem of estimating the 6-DoF (Degree of Freedom) camera pose of a query image captured in a known scene, which is a key building block of many computer vision and robotics applications. Recent advances in structure-based localization solve this problem by memorizing the mapping from image pixels to scene coordinates with neural networks to build 2D-3D correspondences for camera pose optimization. However, such memorization requires training by amounts of posed images in each scene, which is heavy and inefficient. On the contrary, few-shot images are usually sufficient to cover the main regions of a scene for a human operator to perform visual localization. In this paper, we propose a scene region classification approach to achieve fast and effective scene memorization with few-shot images. Our insight is leveraging a) pre-learned feature extractor, b) scene region classifier, and c) meta-learning strategy to accelerate training while mitigating overfitting. We evaluate our method on both indoor and outdoor benchmarks. The experiments validate the effectiveness of our method in the few-shot setting, and the training time is significantly reduced to only a few minutes. 11Code available at: https://github.com/siyandong/SRC © 2022 IEEE.
Keyword :
Cameras Cameras Computer vision Computer vision Degrees of freedom (mechanics) Degrees of freedom (mechanics) Learning systems Learning systems
Cite:
Copy from the list or Export to your reference management。
GB/T 7714 | Dong, Siyan , Wang, Shuzhe , Zhuang, Yixin et al. Visual Localization via Few-Shot Scene Region Classification [C] . 2022 : 393-402 . |
MLA | Dong, Siyan et al. "Visual Localization via Few-Shot Scene Region Classification" . (2022) : 393-402 . |
APA | Dong, Siyan , Wang, Shuzhe , Zhuang, Yixin , Kannala, Juho , Pollefeys, Marc , Chen, Baoquan . Visual Localization via Few-Shot Scene Region Classification . (2022) : 393-402 . |
Export to | NoteExpress RIS BibTex |
Version :
Export
Results: |
Selected to |
Format: |