透過您的圖書館登入
IP:18.217.8.82
  • 學位論文

基於全域卷積網路與串連特徵圖的語意分割

Semantic Segmentation via Global Convolutional Network and Concatenated Feature Maps

指導教授 : 張隆紋

摘要


自從Long 等人發表了基於經訓練的分類卷積類神經網路(CNN)的完全卷積網路(FCN),更多的方法開始用分類CNN來建構他們的語意分割CNN並在許多具挑戰性的資料集上取得一流的表現。由於ResNet在2015年取得了ImageNet分類競賽的冠軍,大部分的分割CNN選擇將他們的網路建構在 ResNet 之上。 去年,黃等人提出了一個新的分類CNN名叫DenseNet。在這之後Jégou等人僅僅用一連串的DenseNet建構模塊來建造他們的整個分割CNN,名叫FC-DenseNet,就在CamVid資料集上取得了一流的成果。 在這篇論文中,我們希望能證明直接把DenseNet建構模塊實做到分割CNN並不是應用它最好的方法。因此,我們用DenseNet的設計概念來修改一個基於ResNet架構名為GCN的分割CNN,並建構我們自己的類神經網路。我們的網路用更少的訓練參數便在CamVid資料集上取得了69.34%的mean-IoU成績,超越了FC-DenseNet在論文中取得的66.9%。

並列摘要


After Long et al. proposed the fully convolutional network (FCN) based on a pre-trained classification convolutional neural network (CNN), more methods started using classification CNN to build their semantic segmentation CNN and achieved state-of-the-art performances on challenging datasets. Since ResNet won the first place of ImageNet classification task in 2015. Most of the segmentation CNN chose to build their network based on the ResNet. Recently, Huang et al. introduced a new classification CNN called DenseNet. Then Jégou et al. simply used a sequence of building blocks of DenseNet to build their entire segmentation CNN, called FC-DenseNet, and achieved state-of-the-art result on CamVid dataset. In this thesis, we want to prove that implement DenseNet building block directly into a segmentation CNN is not the best way of using it. Therefore, we implement the design concept of DenseNet into a ResNet-based segmentation CNN called GCN and build our own network. Our network uses less computation resource to obtain a mean-IoU score of 69.34% on CamVid dataset, surpass the 66.9% obtained in the paper of FC-DenseNet.

並列關鍵字

Semantic Segmentation Deep Learning CNN

參考文獻


[1] A. Krizhevsky, I. Sutskever and G.E. Hinton, “ImageNet classification with deep convolutional neural networks,” in Conference on Neural Information Processing Systems, 2012, pp. 1097-1105.
[2] K. Simonyan, and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in CoRR, abs/1409.1556, 2014.
[3] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke and A. Rabinovich, “Going deeper with convolutions,” in IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 1-9.
[4] K. He, X. Zhang, S. Ren and J. Sun, “Deep residual learning for image recognition,” in IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770-778.
[5] G. Huang, Z. Liu, L. van der Maaten and K. Q. Weinberger, “Densely connected convolutional networks,” in IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 2261-2269.

延伸閱讀