透過您的圖書館登入
IP:3.16.137.108
  • 學位論文

果蠅影像序列的行為標註系統及機器學習辨識

Behavior Annotation System for Drosophila Image Sequence and Behavior Recognition Using Machine Learning

指導教授 : 劉震昌

摘要


由於果蠅與人類的基因有許多相似的地方,透過觀測基因改變後果蠅的求偶行為可以了解哪些基因會影響人類的行為。對果蠅求偶動作的進行影像分析,過去是使用人工方式去測量並記錄,導致生物學家們需要花費大量時間觀察及記錄,此方法效率不高且容易出錯。為了降低觀察所耗費的時間並加快分析行為數據,本篇論文開發一個能讓生物學家輕易掌握的果蠅影像序列的行為標註系統,使生物學家在研究及分析果蠅求偶行為能更有效率。後續並利用生物學家對影片中的行為標記,再配合自動影像物件切割程式,可以切割出用於後續機器學習的果蠅行為圖像。 本論文建立機器學習是以DIGITS作為建立機器學習的平台且以Caffe當作機器學習的框架,我們使用AlexNet及GoogLeNet的網路架構來訓練模型。本論文的辨識行為種類共分為Singing、Tapping、Attempting三種,其中Singing與Tapping選用的圖像張數為7602張,Attempting選用的圖像張數為5184張。我們在DIGITS上的使用微調後的AlexNet模型驗證準確率為99.79%、GoogleNet的模型驗證準確率為99.88%。為了驗證微調後的AlexNet及GoogLeNet的模型可靠度,我們在Singing、Tapping、Attempting各選用180張與訓練資料集相同影片產生出的圖像來做測試,經過測試後我們得到AlexNet的準確率為75.9%、GoogLeNet的準確率為93.3%,接著使用Singing、Tapping、Attempting各200張與訓練資料集不同的影片產生出的圖像做測試,得到AlexNet的準確率為62.7%、GoogLeNet的準確率為60%。

並列摘要


Since the genes of Drosophila and humans have many similarities, by observing the courtship behavior of the flies that result from genetic changes, we can understand which genes affect human behavior. In the past, the image analysis of the courtship action of fruit flies used manual methods to measure and record, which caused biologists to spend numerous hours observing and recording. This method is inefficient and error-prone. In order to reduce the time spent on observation and speed up the analysis of behavioral data, this thesis has developed a behavioral labeling system for fruit fly image sequences, which can be easily mastered by biologists so can study and analyze Drosophila courtship behavior more effectively. Subsequent use of biologists’ marking of behavior in the film with the automatic image object segmentation program, can cut out the fruit fly behavior image for subsequent machine learning. In this thesis, machine learning is established by using DIGITS as the machine learning platform and Caffe as the machine learning framework. We also use the network architecture of AlexNet and GoogLeNet to train the model. There are three types of identification behaviors in this thesis: Singing, Tapping, and Attempting. The number of images labeled as Singing and Tapping is 7602 respectively, and the number of images labeled as Attempting is 5184. We used the fine-tuned AlexNet model and the training accuracy rate on DIGITS was 99.79%, on the other hand, the GoogleNet model training accuracy rate was 99.88%. In order to verify the reliability of the fine-tuned AlexNet and GoogLeNet models, we selected 180 images for Singing, Tapping, and Attempting respectively generated from the same videos as the testing data set. After testing, we got the accuracy of AlexNet to be 75.9%, and GoogLeNet's accuracy rate was 93.3%. Then, we selected 200 images for Singing, Tapping, and Attempting respectively generated by different videos as the testing data set. We got AlexNet accuracy rate of 62.7%, and GoogLeNet accuracy rate of 60%.

參考文獻


[1] M. B. Sokolowski, “Drosophila: Genetics meets behaviour,” Nature Reviews Genetics, pp. 879-890, 2001.
[2] Jenny Yuen, Bryan Russell, Ce Liu, Antonio Torralba, “LabelMe video: Building a video database with human annotations,” ICCV, 2009.
[3] Hung-Yin Tsai ,Yen-Wen Huang, “Image tracking study on courtship behavior of Drosophila,” PLOS ONE, 2012.
[4] Siyuan Lu, Zhihai Lu, Yu-Dong Zhang, “Pathological brain detection based on AlexNet and transfer learning,” Journal of computational science, pp. 41-47, 2019.
[5] Caffe, https://caffe.berkeleyvision.org/

延伸閱讀