房間空間佈局預測是輸入一張房間圖片而輸出對應面標籤。之前的方法是預測一組房間的角落點和其對應的房間類型,然後再根據房間類型將事先定義好的坊間角落順序依序連線產生其平面標籤圖。這個方法要求要事先定義房間的類型。 在本篇論文中,我們提出一個基於平面切割之全卷積神經網路自動編碼器來預測房間空間佈局。以一張圖片當作輸入資料,直接輸出一個平面切割的端到端可訓練自動編碼器。我們使用了省略與連接的技巧增強了輸出結果的準確率,還使用了多目標訓練方式訓練了平面分割及平面邊線熱度圖,平面邊線熱度圖的目的是為了學習到直線的特徵和細節。相較於之前的方法,我們的方法並不需要事先定義房間類型,只需要標簽好對應平面的資料集,而且可以處理擁有非直線牆角的房間架構類型。
Room layout estimation from the monocular RGB image is the purpose of the paper. The state-of-the-art method first estimates an ordered set of room layout key points and room types and then generates the corresponding segmentation image of room layout based on them. However, the method requires predefining 11 room types which are actually a limitation. Therefore, we want to directly estimate the segmentation image of room layout without doing the predefinition and would not be restricted to room types. In this paper, an end-to-end trainable Autoencoder is built to achieve the goal. Moreover, we adopt two ideas while building Autoencoder. One is skip connected architecture called modified U-net, and the other is multi-task which aims to generate both the boundaries contour map and the segmentation image. The experiment results show that ours can lead to better room layout estimations which are not limited to the room types.