完整後設資料紀錄
DC 欄位語言
dc.contributor.authorChen, Po-Hengen_US
dc.contributor.authorYang, Hsiao-Chienen_US
dc.contributor.authorChen, Kuan-Wenen_US
dc.contributor.authorChen, Yong-Shengen_US
dc.date.accessioned2020-10-05T02:01:10Z-
dc.date.available2020-10-05T02:01:10Z-
dc.date.issued2020-01-01en_US
dc.identifier.issn1057-7149en_US
dc.identifier.urihttp://dx.doi.org/10.1109/TIP.2020.3000611en_US
dc.identifier.urihttp://hdl.handle.net/11536/155205-
dc.description.abstractThe goal of Multi-View Stereo (MVS) is to reconstruct 3D point-cloud model from multiple views. On the basis of the considerable progress of deep learning, an increasing amount of research has moved from traditional MVS methods to learning-based ones. However, two issues remain unsolved in the existing state-of-the-art methods: (1) only high-level information is considered for depth estimation. This may reduce the localization accuracy of 3D points as the learned model lacks spatial information; and (2) most of the methods require additional post-processing or network refinement to generate a smooth 3D model. This significantly increases the number of model parameters or the computational complexity. To this end, we propose MVSNet++, an end-to-end trainable network for dense depth estimation. Such an estimated depth map can further be applied to 3D model reconstruction. Different from previous methods, in the proposed method, we first adopt feature pyramid structures for both feature extraction and cost volume regularization. This can lead to accurate 3D point localization by fusing multi-level information. To generate smooth depth map, we then carefully integrate instance normalization into MVSNet++ without increasing model parameters and computational burden. Furthermore, we additionally design three loss functions and integrate Curriculum Learning framework into the training process, which can lead to an accurate reconstruction of 3D model. MVSNet++ is evaluated on DTU and Tanks & Temples benchmarks with comprehensive ablation studies. Experimental results demonstrate that our proposed method performs favorably against previous state-of-the-art methods, showing the accuracy and effectiveness of the proposed MVSNet++.en_US
dc.language.isoen_USen_US
dc.subjectFeature extractionen_US
dc.subjectThree-dimensional displaysen_US
dc.subjectSolid modelingen_US
dc.subjectComputational modelingen_US
dc.subjectImage reconstructionen_US
dc.subjectEstimationen_US
dc.subjectTrainingen_US
dc.subjectMulti-view stereoen_US
dc.subjectdeep learningen_US
dc.subject3D model reconstructionen_US
dc.subjectfeature aggregationen_US
dc.subjectplane sweep algorithmen_US
dc.titleMVSNet plus plus : Learning Depth-Based Attention Pyramid Features for Multi-View Stereoen_US
dc.typeArticleen_US
dc.identifier.doi10.1109/TIP.2020.3000611en_US
dc.identifier.journalIEEE TRANSACTIONS ON IMAGE PROCESSINGen_US
dc.citation.volume29en_US
dc.citation.spage7261en_US
dc.citation.epage7273en_US
dc.contributor.department資訊工程學系zh_TW
dc.contributor.department電子工程學系及電子研究所zh_TW
dc.contributor.departmentDepartment of Computer Scienceen_US
dc.contributor.departmentDepartment of Electronics Engineering and Institute of Electronicsen_US
dc.identifier.wosnumberWOS:000553851400002en_US
dc.citation.woscount0en_US
顯示於類別:期刊論文