Image depth estimation assisted by multi-view projection

Na minha lista:
Detalhes bibliográficos
Publicado no:Complex & Intelligent Systems vol. 11, no. 1 (Jan 2025), p. 66
Publicado em:
Springer Nature B.V.
Assuntos:
Acesso em linha:Citation/Abstract
Full Text - PDF
Tags: Adicionar Tag
Sem tags, seja o primeiro a adicionar uma tag!

MARC

LEADER 00000nab a2200000uu 4500
001 3141262394
003 UK-CbPIL
022 |a 2199-4536 
022 |a 2198-6053 
024 7 |a 10.1007/s40747-024-01688-6  |2 doi 
035 |a 3141262394 
045 2 |b d20250101  |b d20250131 
245 1 |a Image depth estimation assisted by multi-view projection 
260 |b Springer Nature B.V.  |c Jan 2025 
513 |a Journal Article 
520 3 |a In recent years, deep learning has significantly advanced the development of image depth estimation algorithms. The depth estimation network with single-view input can only extract features from a single 2D image, often neglecting the information contained in neighboring views, resulting in learned features that lack real geometrical information in the 3D world and stricter constraints on the 3D structure, leading to limitations in the performance of image depth estimation. In the absence of accurate camera information, the multi-view geometric cues obtained by some methods may not accurately reflect the real 3D structure, resulting in a lack of multi-view geometric constraints in image depth estimation algorithms. To address this problem, a multi-view projection-assisted image depth estimation network is proposed, which integrates multi-view stereo vision into a deep learning-based encoding-decoding image depth estimation framework without pre-estimation of view bitmap. The network estimates optical flow for pixel-level matching across views, thereby projecting the features of neighboring views to the reference viewpoints for self-attentive feature aggregation, compensating for the lack of stereo geometry information in the image depth estimation framework. Additionally, a multi-view reprojection error is designed for supervised optical flow estimation to effectively constrain the optical flow estimation process. In addition, a long-distance attention decoding module is proposed to achieve effective extraction and aggregation of features in distant areas of the scene, which enhances the perception capability for outdoor long-distance. Experimental results on the KITTI dataset, vKITTI dataset, and SeasonDepth dataset demonstrate that our method achieves significant improvements compared to other state-of-the-art depth estimation techniques. This confirms its superior performance in image depth estimation. 
653 |a Feature extraction 
653 |a Geometric constraints 
653 |a Algorithms 
653 |a Datasets 
653 |a Images 
653 |a Encoding-Decoding 
653 |a Deep learning 
653 |a Machine learning 
653 |a Optical flow (image analysis) 
773 0 |t Complex & Intelligent Systems  |g vol. 11, no. 1 (Jan 2025), p. 66 
786 0 |d ProQuest  |t Advanced Technologies & Aerospace Database 
856 4 1 |3 Citation/Abstract  |u https://www.proquest.com/docview/3141262394/abstract/embedded/6A8EOT78XXH2IG52?source=fedsrch 
856 4 0 |3 Full Text - PDF  |u https://www.proquest.com/docview/3141262394/fulltextPDF/embedded/6A8EOT78XXH2IG52?source=fedsrch