透過您的圖書館登入
IP:3.134.103.74
  • 學位論文

以SURF與HRI影像演算法整合三軸角錐形氣壓並聯式機構機械臂之研究

The Integration of The SURF and HRI Image Algorithm with A Three-Axial Pyramidal Pneumatic Parallel Manipulator

指導教授 : 江茂雄

摘要


本文旨在以SURF與HRI影像演算法整合三軸角錐形氣壓並聯式機構機械臂,使機械臂可藉由影像辨識物件形狀及位置,自動完成取放作業之工作。本文採用加速強健特徵點演算法 (Speed up robust feature, SURF)來定義目標物體的特徵點,並且藉由比對當前畫面中的特徵點來判斷目標物是否存在於影像中。為了強化特徵比對結果並算出抓物控制所需的參考點,本論文採用隨機取樣篩選演算法 (RANdom Sample Consensus, RANSAC)來估測平面轉換矩陣 (Homography matrix)以準確的標出目標物的中心點。並利用最新發展出的遊戲設備,ASUS Xtion Pro Live深度攝影機能夠直接擷取目標點3-D空間座標。此外,本文並發展出一套座標估測的校正法,提高對於目標物座標估測之準確度。 人機互動是提供非受訓練的使用者能夠更容易、更有效率與機器互動的一種方法。本研究提出一套利用手勢辨識來操控機械臂的方法,使用網路攝影機取得影像後經過多個處理程序,達到辨識之功能。其中的影像處理包含膚色偵測、雜訊消除和形態學以計算手指數量,並將資訊傳給控制器,利用以上所提出的理論規劃機械臂端點之軌跡。最後,經過實驗驗證本文提出之方法的可行性,證實此系統可成功導引三軸式機械臂之端點順利的移動並拿取所設定的目標物。

並列摘要


The objective of this study is to develop the SURF and HRI image algorithm integrated with a three-axial pneumatic parallel manipulator. The manipulator can pick and place objects automatically by the feature information of the image through the SURF algorithm with scale- and rotation-invariants. The speed up robust feature (SURF) algorithm is used to define the feature of a target object and to match features between the current image and the object database for confirming the target. To strengthen the feature matching results and calculate the necessary reference control point, we adopt the RANSAC(RANdom Sample And Consensus) algorithm to estimate the planar transformation matrix (homography matrix) in order to accurately mark the center of target. The ASUS Xtion Pro Live depth camera which can directly estimate the 3-D location of target point is used in this study. A set of coordinate estimation calibration method is developed to improve the accuracy of target location estimation. Human-Robot-Interaction (HRI) offers a way to enable untrained users to interact with robots more easily and efficiently. This study also presents a method for hand gesture recognition to command the manipulator. The stages include skin detection to effectively capture only the skin region of the hand, noise elimination, and applications of the morphology to determine the active finger count. Once the finger count is determined, the information is transmitted to the manipulator controller. The end-effector of the manipulator can then move to the desired location according to the finger count. Finally, the experiments of the three-axial manipulator end-effector integrated with the feature recognition algorithm demonstrate that the proposed methods can achieve the feature recognition and pick and place of the target object successfully.

參考文獻


[21] J.-P. Merlet, Parallel Robots: Springer, 2006.
[2] T. Lindeberg, "Feature detection with automatic scale selection," International journal of computer vision, vol. 30, pp. 79-116, 1998.
[3] K. Mikolajczyk and C. Schmid, "Indexing based on scale invariant interest points," in Computer Vision, 2001. ICCV 2001. Proceedings. Eighth IEEE International Conference on, 2001, pp. 525-531.
[4] D. G. Lowe, "Object recognition from local scale-invariant features," in Computer Vision, 1999. The proceedings of the seventh IEEE international conference on, 1999, pp. 1150-1157.
[9] X. Yin and M. Xie, "Finger identification and hand posture recognition for human–robot interaction," Image and Vision Computing, vol. 25, pp. 1291-1300, 2007.

延伸閱讀