透過您的圖書館登入
IP:18.226.166.214
  • 學位論文

多視角環境下之行為分析與辨識

Multi-View Behavior Analysis and Recognition

指導教授 : 謝君偉
若您是本文的作者,可授權文章由華藝線上圖書館中協助推廣。

摘要


近年來行為分析在多媒體影像處理的問題中,扮演著很重要的角色,同時有許多實際應用於生活之中,例如:社區監控、視訊監控、遠距醫療監控、異常行為偵測…等。一般的行為分析研究限制於固定視角,但於現實環境當中,人類行為不會只在預設的視角下被察覺。因此,本篇論文著重於多視角環境下使用單一攝影機之人類行為的分析與辨識。   本系統將多視角之行為分析轉換至單一視角下處理。先使用姿勢樣板比對找出最相似的姿勢樣板,再將樣板比對結果轉換至指定視角當中同姿勢的樣板。利用樣板之間的相關性與樣版轉換機率的輔助,使得指定視角中的樣板轉換有準確的結果。將此連續的樣板存入統計陣列之中,將此陣列與資料庫之行為陣列進行辨識,可分析出此視訊之中發生的行為。   由實驗結果得知,我們的系統可以適用於多視角環境下之行為分析系統,並且有良好的表現。

並列摘要


This paper presents a new behavior classification system that can be used to analyze human movements from any views and directly form videos. Technically, if a person is observed from other views, his appearances will change significantly. To freely recognize his behaviors, traditional methods tend to adopt 3-D data for behavior modeling and analysis. However, its inherent correspondence process is very time-consuming and will make it inappropriate for real time applications. To tackle this problem, this thesis proposes a novel human representation scheme for recognizing human behaviors from any views. In this scheme, a novel view alignment method is first proposed for mapping each action sequence (captured from any views) to a fixed view. To achieve this mapping, the spatial and temporal features should be first extracted from each action sequence. To extract the spatial feature, the central contexts of each posture are then extracted through a triangulation technique. Then, a set of key postures is selected and built for converting an action to a symbol string. To reduce the converting errors, a transitional probably table is built for recording the possibility of one posture transferring to another posture. With the table and the central context feature, each action sequence can be then aligned to a fixed view and then represented by an action matrix. After that, matching two action sequences from arbitrary views will become a single-view matrix matching problem. Then, the Viterbi algorithm is used for aligning two action sequences and then classifying them to different behavior types. Experimental results prove that the proposed method is a robust and accurate tool for human movement analysis from any views.

參考文獻


[21] C. Rao, A. Gritai and M. Shah, “View-invariant Alignment and Matching of Video Sequences,” International Conference on Computer Vision, vol. 2, pp. 939-945, Oct. 2003.
[2] R. Cucchiara, C. Grana, A. Prati, and R. Vezzani, “Probabilities posture classification for human-behavior analysis,” IEEE Transactions on Systems, Man, and Cybernetics-Part A: System and Humans, vol. 35, no. 1: pp. 42-54, Jan. 2005.
[1] T. B. Moeslund and E. Granum, “A survey of computer vision-based human motion capture,” Computer Vision and Image Understanding, vol. 81, no. 3, pp. 231-268, Mar. 2001.
[3] Haritaoglu, D. Harwood, and L.S. Davis, “W4: real-time surveillance of people and their activities,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no. 8, pp. 809-830, 2000.
[5] I. Mikic, et al., “Human body model acquisition and tracking using voxel data,” International Journal of Computer Vision, vol. 53, no. 3, pp. 199-223, 2003.

延伸閱讀