透過您的圖書館登入
IP:13.58.121.131
  • 學位論文

Spatio-Temporally Consistent Multi-View Video Synthesis for Autostereoscopic Display

適用於自動立體顯示器且具時空一致性的多視角視訊合成

指導教授 : 賴尚宏
若您是本文的作者,可授權文章由華藝線上圖書館中協助推廣。

摘要


In this paper, we propose a novel algorithm to generate multiple virtual views from a video-plus-depth sequence for modern autostereoscopic displays. To synthesize realistic content in the disocclusion regions from the virtual views is the main challenging problem in this task. In order to produce perceptually satisfactory images, our proposed algorithm takes advantage of spatial coherence and temporal consistency to handle the uncertain pixels in the disocclusion regions. On the one hand, regarding the spatial coherence, we incorporate the intensity gradient strength with the depth information to determine the filling priority for inpainting the disocclusion regions, so that the continuity of image structures can be preserved. On the other hand, the temporal consistency is enforced by considering the intensities in the disocclusion regions across the adjacent frames through an optimization process. We propose an iterative re-weighted framework to jointly consider intensity and depth consistency in the adjacent frames, which not only imposes temporal consistency but also reduces noise disturbance. Finally, for accelerating the multi-view synthesis process, we apply the proposed view synthesis algorithm to generate the images plus depth at the leftmost and rightmost viewpoints, so that the intermediate views are efficiently interpolated through image warping according to the associated depth maps between the two views.

並列摘要


在本文中,我們提出了一個新演算法可以從一個視訊加深度的影像序列去產生多個虛擬視角的視訊,並可應用於現代的自動立體顯示器上。在這個任務中,要合成原本被遮蔽的區域裡的實際內容是最主要的挑戰。為了產生令人感覺滿意的影像,我們提出的方法利用了空間的連貫性和時間上的一致性去處理在原本被遮蔽的區域裡那些不確定的點。關於空間的連貫性,我們會整合色彩的強度梯度與深度資訊來決定填補的優先順序以維持影像裡的連續性結構。至於時間一致性方面,我們會將時間上相鄰影像的色彩強度加入考慮。我們提出了一個反覆進行的方法,這個方法不僅有加入了時間一致性也降低了雜訊干擾。最後,為了加速多視角的合成過程,我們提出的影像合成演算法會先產生最左端視角點的視訊加深度和最右端視角點的視訊加深度,然後就可以透過影像轉換有效率地內插出中間的視角點影像。

參考文獻


[1] C. Tomasi and R. Manduchi, “Bilateral filtering for gray and color images,” IEEE International Conference on Computer Vision, pp. 839-846, 1998.
[3] A. Criminisi, P. P´erez and K. Toyama, “Object removal by exemplar-based inpainting,” IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol. 2, No. 2, pp.721-728, 2003.
[6] R-P.M. Berretty, F.J. Peters and G..T.G. Volleberg, “Real time rendering for multiview autostereoscopic,” SPIE on Stereoscopic Displays and Virtual Reality Systems XIII, Vol. 6055, pp.208-219, 2006.
[8] F. Christoph, “A 3D-TV system based on video plus depth information,” Asilomar Conference on Signal, System and Computers, vol. 2, pp. 1529-1533, 2003.
[9] F. Christoph, “Depth-image-based rendering (DIBR), compression, and transmission for a new approach on 3D-TV,” SPIE on Stereoscopic Displays and Virtual Reality Systems XI, Vol.5291, pp.93-104, 2004.

延伸閱讀