透過您的圖書館登入
IP:3.16.15.149
  • 學位論文

利用圖像概念分割之影像分類器可釋性萃取

Explainability Extraction of Image Classification based on Concept Segmentation

指導教授 : 藍俊宏
若您是本文的作者,可授權文章由華藝線上圖書館中協助推廣。

摘要


自類神經網路的模型困境有所突破、重受學術、產業界青睞以來,影像辨識的技術亦突飛猛進,尤其是搭配大幅提昇的電腦硬體運算能力,常使用深度神經網路模型來進行圖像的分類或辨識。深度神經網路擅長從資料中找出錯綜複雜的規律並自動萃取隱藏特徵。因此得以攻克以前難以完成的預測任務,然而深度神經網路常被視為難以理解的黑盒子,模型訓練完成後無法知悉其內部運作機制,倘若模型運作機制與人類認知產生落差、甚至相左,在特定應用領域上恐難以協助決策、甚至造成危害,縱然有高度的預測效果,也因其不可解釋的特質而降低了實用性。 針對圖像分類器的解析,現有主流解釋性方法多聚焦在像素層級的解釋,本研究發展基於概念區塊的解釋性框架,其特色是萃取之概念能夠維持在圖像相近區域,並建立以概念作為特徵的可自釋模型來逼近黑盒子,最後綜合不同預測類別的概念重要性排序,檢測影像分類器的推論規則是否合乎人類的判斷邏輯,進而增加實務採用深度神經網路技術的信心。透過實例驗證,本研究提出的概念萃取符合直覺,並能有效解釋圖像分類結果。

並列摘要


As the model limitation of Artificial Neural Networks (ANNs) has been broken through, AI techniques are back to the center stage again for academics and industries. The capability of image classification has also advanced significantly, and many applications are realized especially thanks to the greatly improved computing power. Deep Neural Nets (DNNs) are good at finding intricate rules/patterns from data and automatically extracting hidden features. The prediction tasks which were difficult to solve can be overcome quickly now. However, DNNs are often regarded as incomprehensible black boxes which cannot be unfolded once the model is trained. If its internal inference mechanism deviates or even contradicts human cognition, it may be difficult to support decision-making in specific application fields. For the explanation decomposition of image classifiers, the mainstream methods focus on the interpretation at the pixel level. Significant pixels, which may spread sparsely, are then aggregated to explain the model. This thesis develops an explaining framework based on the image concept, which is a block of neighboring pixels once extracted. A concept-based and thus explainable model is built to approximate the black box model. Concept importance ranking across various predicting classes is then investigated and compared with the intuitive inference logic. Hopefully, the creditability of adopting DNN-based image classification can be increased. Through the proper case study, the proposed method can extract intuitive concepts as well as explain the black-box model logically.

參考文獻


Lecun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278-2324. doi:10.1109/5.726791
Achanta, R., Shaji, A., Smith, K., Lucchi, A., Fua, P., & Susstrunk, S. (2012). SLIC Superpixels Compared to State-of-the-Art Superpixel Methods. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34. doi:10.1109/TPAMI.2012.120
Adebayo, J., Gilmer, J., Muelly, M., Goodfellow, I., Hardt, M., & Kim, B. (2018). Sanity checks for saliency maps. Paper presented at the Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, Canada.
Alzubaidi, L., Zhang, J., Humaidi, A. J., Al-Dujaili, A., Duan, Y., Al-Shamma, O., . . . Farhan, L. (2021). Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. Journal of big data, 8(1), 53-53. doi:10.1186/s40537-021-00444-8
Ancona, M., Ceolini, E., Öztireli, C., & Gross, M. H. (2018). Towards better understanding of gradient-based attribution methods for Deep Neural Networks. Paper presented at the ICLR.

延伸閱讀