标题: 基于虚拟视角的立体影片合成
Virtual-view-based Stereo Video Composition
作者: 张钧凯
Chang, Chun-Kai
杭学鸣
Hang, Hsueh-Ming
电子工程学系 电子研究所
关键字: 影像合成;视角合成;分割;不匹配;深度竞争;摄影机轨迹;image composition;view synthesis;segmentation;mismatch;depth competition;camera motion
公开日期: 2012
摘要: 立体数位内容日益被重视,新型态的技术包含了自由视点视讯(FTV, Free Viewpoint Television)与扩增实境(AR, Augmented Reality),这些应用以任意视点合成技术为最主要的关键议题。有许多的任意视点合成演算法被提出,通常都是利用多重影像以及其对应的深度资讯图来产生虚拟试点的影像以达到任意视点的效果。我们利用这种基于影像与深度的影像合成渲染(DIBR, Depth Image-based Rendering)来产生背景置换后的立体视讯内容。

输入为两组多个摄影机分别拍摄的两组视讯,我们希望结合这些输入来产生新的立体场景。此立体场景由其中一组输入的前景物体,与另外一组的背景场景共同组成。为了这个目的,我们将以多组视讯间场景间不匹配(mismatch)的角度来观察,在此论文中主要将讨论包括摄影机参数以及摄影机定位的不匹配。当使用者在背景影像中选取了定位点(landing point),我们需要经由改变摄影机相关参数来合成出背景场景的对应虚拟视角(配合前景摄影机),以达成背景置换。这样的方式可以大幅增加创作的自由度。

相较于传统的影像创作(Image Composition),上述的过程需要利用到深度几何的资讯。欲被合成的背景场景需要经由虚拟摄影机参数的计算。此外,为了保持场景物体间互相遮蔽的关系,在背景置换时深度竞争(Depth Competition)是另外一个被探讨的议题。当我们将静态影像延伸至视讯时,我们需要摄影机移动行为的资讯来补偿不同场景间摄影机的移动不匹配问题。实验结果显示我们可以达成令人满意的视觉观感。
3D video is gaining its popularity recently. In addition to the conventional left-right view 3D pictures, new forms of 3D video such as free viewpoint TV (FTV) and augmented reality (AR) are introduced. The Depth Image-based Rendering (DIBR) technique is one enabling rendering technique behind these applications. Typically, it uses multiple views with depth information to generate the intermediate view at any arbitrary viewpoint. We can use the DIBR techniques to produce new stereo videos with background substitution.
Given two sets of videos captured by two sets of multiple cameras, we like to combine them to create a new stereo scene with the foreground objects from one set of video and the background from the other set. We will study a few mismatch issues between two scenes such as camera parameter mismatch and camera orientation mismatch problems in this thesis. We propose a floor model to adjust the camera orientation. Once we pick up the landing point (of foreground) in the background scene, we need to adjust the background camera parameters (position etc.) to match the foreground object, which enriches the freedom of composition.
In contrast to the conventional 2D composition methods, the depth information is used in the above calculation. Thus, the new background scenes may have to be synthesized based on the calculated virtual camera parameters and the given background pictures. The depth competition problem is another issue to maintain the inter-occlusion relationship in the composite new scene. If we extend this 3D composition form still pictures to motion pictures, we need the camera movement information too. The camera motion is estimated for individual scene to solve the mismatch of camera motion of two scenes. Plausible results are demonstrated using the proposed algorithms.
URI: http://140.113.39.130/cdrfb3/record/nctu/#GT070050206
http://hdl.handle.net/11536/72284
显示于类别:Thesis


文件中的档案:

  1. 020601.pdf
  2. 020602.pdf

If it is a zip file, please download the file and unzip it, then open index.html in a browser to view the full text content.