透過您的圖書館登入
IP:3.138.114.132
  • 學位論文

針對即時手部辨識及追蹤所設計以粒子濾波器為基礎之視窗旋轉及縮放演算法

Particle-based Window Rotation and Scaling Scheme for Real-time Hand Recognition and Tracking Systems

指導教授 : 方凱田

摘要


在這篇論文中,我們提出了一個使用單一攝影機就可以即時辨識出不只手的位置、還有手的大小和角度的多層級系統。這系統中包含了視窗定位層、視窗縮放層和視窗旋轉層以分別用來偵測手的位置、大小和角度。 從攝影機抓到的影像先透過預先處理以去除掉非膚色的背景。接著,每一個層級都利用方向梯度直方圖、支持向量機和粒子濾波器來偵測並追蹤手部視窗的不同特性。一個粒子在我們系統中被定義成一個可旋轉的長方形視窗,因為他可以同時表達手的位置、大小和角度這三個狀態。 不像傳統的多層級系統必須逐案地偵測出許多非手的情況然後再移除,我們的系統不只可以直接判斷出手的狀態、還可以在膚色背景之下正常運作。這是因為我們系統中每一個層級都只負責處理擁有相似特性的資料,如此有利於支持向量機訓練模型和預測資料。再者,一個層級可以透過我們所提出的「跨層級狀態傳遞」來和其他層級共享預測到的結果,以此來幫助其他層級限縮他們可能的狀態空間。「跨層級狀態傳遞」讓每個粒子可以不用在整個狀態空間中探索,如此可以大幅度地減低即時運作時的負擔。這樣的架構允許每個層級專注在自己負責的特性以至能在一個低多樣性的空間裡運作,並且可以分享預測到的狀態來幫助其他層級建立這種低多樣性的空間。最後,我們還提出了一個方法去重複利用每次計算出來的方向梯度直方圖,以至可以增加大量粒子以改善系統預測結果同時又不嚴重降低幀率。

並列摘要


In this thesis, a real-time multi-stage system that can perceive not only hand location but also hand size and hand angle using a single camera is proposed. The system is called PWRS in this thesis, and it is the abbreviation of Particle-based Window Rotation and Scaling. There are three stages, window-locating stage, window-scaling stage and window-rotating stage, recognizing the location, size and angle of a hand respectively. The frame captured by the webcam is preprocessed first to exclude those non-skin contexts. Then, each stage employs Histogram of Oriented Gradients (HOG), Support Vector Machines (SVM), and particle filter to detect and track different characteristics of the hand window. A particle defined in our system is a rotated rectangular window which can describe the location, size and angle of a hand at the same time. Unlike traditional multi-stage system which needs to detect and then remove lots of non-hand regions case by case, PWRS can not only directly recognize the hand states but also work normally in front of skin-like background. Because, each stage in PWRS is responsible for data with similar characteristic so as to benefit the SVM training and prediction. Also, a stage can share the predicted results with the others by proposed cross-stage propagation (CSP) to help them narrow down the possible region of states. With CSP, particles need not explore the entire state space, and this considerably alleviates the online loading. The architecture allows each stage to concentrate on its own target characteristic so as to work in a diversity-reduced space, and to share the results with other stages so as to help them construct the low-diversitied space. Finally, a proposed method called sub-HOG extraction (SHE) is designed to reuse HOG features so as to increase numerous particles for better prediction without collapsing frames per second (FPS).

參考文獻


[1] Hui-Shyong Yeo and Byung-Gook Lee and Hyotaek Lim, “Hand tracking and gesture recognition system for human-computer interaction using low-cost hardware,” Multimed Tools Appl Multimedia Tools and Applications, vol. 74, no. 8, pp. 2687–2715, apr 2015. [Online]. Available: http://dx.doi.og/10.1007/s11042-013-1501-1
[4] C. Liang, Y. Song, and Y. Zhang, “Hand gesture recognition using view projection from point cloud,” in 2016 IEEE International Conference on Image Processing (ICIP), Sept 2016, pp. 4413–4417.
[7] P. Mohan, S. Srivastava, G. Tiwari, and R. Kala, “Background and skin colour independent hand region extraction and static gesture recognition,” in 2015 Eighth International Conference on Contemporary Computing (IC3), Aug 2015, pp. 144–149.
[10] N. Fujishima and T. Ietsuka, “Basic construction of a natural finger outline extraction system with a color glove,” in 2016 IEEE/ACIS 15th International Conference on Computer and Information Science (ICIS), June 2016, pp. 1–6.
[11] Jiang Guo, Jun Cheng, Jianxin Pang, and Yu Guo, “Real-time Hand Detection based on Multi-stage HOG-SVM Classifier,” in 2013 IEEE International Conference on Image Processing, Sept 2013, pp. 4108–4111.

延伸閱讀