透過您的圖書館登入
IP:18.224.33.107
  • 學位論文

RA-GAN: 多重領域圖像轉換使用相對屬性值

RA-GAN: Multi-domain Image-to-Image Translation via Relative Attributes

指導教授 : 廖世偉

摘要


多域圖像到圖像的翻譯最近越來越受到關注。以前的方法將圖像和一些目標屬性作為輸入,並生成具有所需屬性的輸出圖像。但是,這有一個局限性。它們需要指定整個屬性集,即使它們中的大多數都不會被更改。為了解決這一局限性,我們提出了一種新的實用的多域圖像到圖像轉換公式RA-GAN。關鍵的想法是使用相對屬性,它描述了所選屬性的所需變化。為此,我們提出了一個對抗框架,它學習單個生成器來翻譯不僅與相關屬性相匹配,而且表現出更好質量的圖像。此外,我們的發生器能夠通過連續地改變感興趣的特定屬性來修改圖像,同時保留其他特徵。實驗結果證明了我們的方法在面部屬性轉移和插值任務中的定性和定量的有效性。

並列摘要


Multi-domain image-to-image translation has gained increasing attention recently. Previous methods take an image and some target attributes as inputs and generate an output image that has the desired attributes. However, this has one limitation. They require specifying the entire set of attributes even most of them would not be changed. To address this limitation, we propose RA-GAN, a novel and practical formulation to multi-domain image-to-image translation. The key idea is the use of relative attributes, which describes the desired change on selected attributes. To this end, we propose an adversarial framework that learns a single generator to translate images that not only match the relative attributes but also exhibit better quality. Moreover, Our generator is capable of modifying images by changing particular attributes of interest in a continuous manner while preserving the other ones. Experimental results demonstrate the effectiveness of our approach both qualitatively and quantitatively to the tasks of facial attribute transfer and interpolation.

參考文獻


[1] D. Berthelot, C. Raffel, A. Roy, and I. Goodfellow. Understanding and improving interpolation in autoencoders via an adversarial regularizer. arXiv preprint arXiv:1807.07543, 2018.
[2] A. Brock, J. Donahue, and K. Simonyan. Large scale gan training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096, 2018.
[3] Y. Choi, M. Choi, M. Kim, J.-W. Ha, S. Kim, and J. Choo. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In CVPR, 2018.
[4] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. In NIPS, 2014.
[5] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR, 2016.

延伸閱讀