透過您的圖書館登入
IP:18.118.148.168
  • 學位論文

類神經輻射場之可解釋隱變量於三維物件操控

Interpreting Latent Representation in Neural Radiance Fields for Manipulating Object Semantics

指導教授 : 王鈺強
若您是本文的作者,可授權文章由華藝線上圖書館中協助推廣。

摘要


控制三維物體的形變一直以來都是三維視覺領域中備受討論的領域。近年來,全新的三維表示法「類神經輻射場(NeRF)」被提出且蓬勃發展,在對場景的建模上取得很大的成功。使用此種表示法來進行物體合成或控制建模內容逐漸為人們所關注。此篇論文中,我們採用了一個能夠感知語義的生成型類神經輻射場,透過探索與詮釋為特定類別建模的生成型類神經輻射場所學習到的隱變量,得以對該類別之特定區域進行編輯。以預先訓練的生成型類神經輻射場為基礎,我們加入一個語義分割器,用來對每種物體作內部的區域分割,使得此類神經輻射場能夠同時渲染出所選視角的二維圖像與相對應的語義分割結果。我們提出的架構能對生成型類神經輻射所學到的隱變量做操縱,除了可以針對特定部位做編輯,編輯效果也能有不同程度的變化。我們將此架構以不同的生成型類神經輻射場和不同物體類別的資料集做實驗,結果成功地驗證此方法的有效性和實用性。

並列摘要


Manipulating 3D objects has been among the active research topic for 3D vision. With the development and success of neural radiance field (NeRF) on scene modeling, synthesizing and manipulating 3D objects using such a representation becomes desirable. In this thesis, we introduce a semantic-aware generative NeRF, which is able to interpret the latent representation learned by category-specific generative NeRFs and to achieve editing of particular part attributes. With pre-trained generative NeRF, we propose to deploy a semantic segmentor for performing part segmentation on the object category. This allows the rendering of the 2D image and prediction of the corresponding segmentation mask. Our proposed scheme learns to manipulate the resulting latent representation and is optimized to edit the object part of interest with varying degrees. We conduct experiments on various object categories on benchmark datasets, and the results successfully verify the effectiveness and practicality of our proposed model.

參考文獻


B. Mildenhall, P. P. Srinivasan, M. Tancik, J. T. Barron, R. Ramamoorthi,and R. Ng, “Nerf: Representing scenes as neural radiance fields for view synthesis,” in European conference on computer vision. Springer, 2020, pp.405–421.
J. L. Schonberger and J.-M. Frahm, “Structure-from-motion revisited,” in Pro-ceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 4104–4113.
Y.-J. Yuan, Y.-K. Lai, T. Wu, L. Gao, and L. Liu, “A revisit of shape editing techniques: From the geometric to the neural viewpoint,” Journal of Computer Science and Technology, vol. 36, no. 3, pp. 520–554, 2021.
J. J. Park, P. Florence, J. Straub, R. Newcombe, and S. Lovegrove, “Deepsdf: Learning continuous signed distance functions for shape representation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 165–174.
Z. Chen and H. Zhang, “Learning implicit fields for generative shape modeling,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 5939–5948.

延伸閱讀