| 研究生: |
張湘菱 Hsiang-Ling Chang |
|---|---|
| 論文名稱: |
基於智慧眼鏡之擴增實境輔助系統 A Smart-Glasses-based Augmented Reality Assisted System |
| 指導教授: |
蘇木春
Mu-Chun Su |
| 口試委員: | |
| 學位類別: |
碩士 Master |
| 系所名稱: |
資訊電機學院 - 資訊工程學系 Department of Computer Science & Information Engineering |
| 論文出版年: | 2018 |
| 畢業學年度: | 106 |
| 語文別: | 中文 |
| 論文頁數: | 105 |
| 中文關鍵詞: | 擴增實境 、手勢偵測 、手指指向分析 |
| 外文關鍵詞: | augmented reality, hand gesture recognition, finger-pointing analysis |
| 相關次數: | 點閱:23 下載:0 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
近年來,擴增實境蓬勃發展,隨著智慧眼鏡的問世,更是讓這項技術為生活帶來許多的便利性。本論文提出一套擴增實境輔助系統,將系統架設於智慧眼鏡上,讓使用者得以透過個人視角使用此系統。
本系統功能包含(1)使用一套簡易流程建立裝置物件資料集,採用Mask R-CNN方法辨識裝置上的物件類別與位置(2)從影像中擷取指向手勢,並分析指向,依照手指指向顯示物件資訊(3)使用校準物件分析虛擬輔助工具該顯示的角度。
本系統的研發目的在於提供一套可輔助技術人員訓練之系統,藉由指向手勢,即可顯示操作人員欲了解的物件資訊。根據系統的實驗顯示,物件偵測的辨識率達到95.5%;手勢偵測的Kappa值為0.93,且平均偵測到手勢的秒數為0.26秒,即使在不同光線下,指向分析的準確率也有79%,由此可證明本論文所使用的方法具有很好的可信度。
Recently, the application of augmented reality has become more and more prevalent. With the advent of smart-glasses, the related research of augmented reality has been grown vigorously. Therefore, this dissertation proposes an augmented reality assisted system which will be set up on the smart-glasses. By setting the system on the smart glasses, the user will be able to use the system in personal perspective.
There are three main features in the system.(1)With a set of simple procedure, the system will set up dataset of objects on the device. Moreover, the system can identify the object and its position on the device automatically by using the Mask R-CNN method.(2)By capturing pointing gesture from the image and analyzing the pointing direction, the system will display the object information according to the finger pointing direction.(3)Using the calibrating object to analyze the rotation angles of virtual tools.
The aim of this system is to provide a system that can assist technicians in training. With the finger pointing, the system can show the object information which the user wants to know on the smart-glasses immediately. According to the results of the experiments, the percentage of recognition of object detection is 95.5%, the Kappa value of recognition of gesture detection is 0.93, and the average time for detecting pointing gesture is 0.26 seconds. Furthermore, even under different light, the proportion of accuracy of the pointing analysis is up to 79%. Based on the results of the experiments, it was proved that the method which was applied in this dissertation is applicable.
[1] R. T. Azuma, "A survey of augmented reality," Teleoperators and Virtual Environments, vol. 6, no. 4, pp. 355-385, 1997.
[2] G. A. Lee, A. Dünser, S. Kim, and M. Billinghurst, "CityViewAR: A mobile outdoor AR application for city visualization," in IEEE International Symposium on Mixed and Augmented Reality - Arts, Media, and Humanities, 2012.
[3] 巫穎彤, “有助發展遲緩兒童學習之擴增實境互動系統之研製,” 國立中央大學軟體工程研究所碩士論文, 2017.
[4] S. J. Henderson and S. Feiner, "Evaluating the benefits of augmented reality for task localization in maintenance of an armored personnel carrier turret," in International Symposium on Mixed and Augmented Reality, 2009.
[5] V. G. Bellile, S. Bourgeois, M. Tamaazousti, S. N. Collette, and S. Knodel, "A mobile markerless Augmented Reality system for the automotive field," in IEEE ISMAR 2012 workshop on tracking methods and applications, 2012.
[6] J. Limaab, R. Robertoa, F. Simõesa, M. Almeidaa, L. Figueiredoa, J. M. Teixeiraab, and V. Teichrieba, "Markerless tracking system for augmented reality in the automotive industry," Expert Systems with Applications, vol. 82, pp. 100-114, 2017.
[7] R. Girshick, J. Donahue, T. Darrell, and J. Malik, "Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation," in IEEE Conference on Computer Vision and Pattern Recognition, 2014.
[8] R. Girshick, "Fast R-CNN," in IEEE International Conference on Computer Vision, 2015.
[9] S. Ren, K. He, R. Girshick, and J. Sun, "Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 6, pp. 1137-1149, 2016.
[10] K. He, G. Gkioxari, P. Dollár, and R. Girshick, "Mask R-CNN," in IEEE International Conference on Computer Vision , 2017.
[11] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You Only Look Once: Unified, Real-Time Object Detection," in IEEE Conference on Computer Vision and Pattern Recognition, 2016.
[12] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C. Y. Fu, and A. C. Berg, "SSD: Single Shot MultiBox Detector," in Computer Vision and Pattern Recognition, 2016.
[13] K. E. A. van de Sande, J. R. R. Uijlings, T. Gevers, and A. W. M. Smeulders, "Segmentation as selective search for object recognition," in International Conference on Computer Vision, 2011.
[14] T. Y. Lin, P. Dollár, R. Girshick, K. He, B. Hariharan, and S. Belongie, "Feature Pyramid Networks for Object Detection," in Computer Vision and Pattern Recognition, 2017.
[15] J. Long, E. Shelhamer, and T. Darrell, "Fully Convolutional Networks for Semantic Segmentation," in IEEE Conference on Computer Vision and Pattern Recognition, 2015.
[16] M. Kolsch, M. Turk, and T. Hollerer, "Vision-based interfaces for mobility," in The First Annual International Conference on Mobile and Ubiquitous Systems: Networking and Services, 2004.
[17] 羅冠中, “即時手勢辨識系統及其於戰場情資顯示平台之應用,” 國立中央大學資訊工程學系碩士論文, 2015.
[18] N. H. Dardas and N. D. Georganas, "Real-Time Hand Gesture Detection and Recognition Using Bag-of-Features and Support Vector Machine Techniques," IEEE Transactions on Instrumentation and Measurement, vol. 60, no. 11, pp. 3592-3607, 2011.
[19] H. Asano, T. Nagayasu, T. Orimo, K. Terabayashi, M. Ohta, and K. Umeda, "Recognition of finger-pointing direction using color clustering and image segmentation," in The SICE Annual Conference, 2013.
[20] D. Lee and S. Lee, "Vision‐Based Finger Action Recognition by Angle Detection and Contour Analysis," Electronics and Telecommunications Research Institute, vol. 33, no. 3, pp. 415-422, 2011.
[21] E. Tamura, Y. Yamashita, Y. Ho, E. Sato-Shimokawara, T. Nishitani, and T. Yamaguchi, "Wearable finger motion input interface system with GMM foreground segmentation," in Conference on Technologies and Applications of Artificial Intelligence, 2015.
[22] "EPSON BT-300智慧眼鏡官方網站," [Online]. Available: https://www.epson.com.tw/家用系列/智慧穿戴裝置/智慧眼鏡/BT-300/p/V11H756054. [Accessed 17 - May - 2018].
[23] "Labelme: Image Polygonal Annotation with Python," [Online]. Available: https://github.com/wkentaro/labelme. [Accessed 6 - Jun - 2018].
[24] S. Kolkur, D. Kalbande, P. Shimpi, C. Bapat, and J. Jatakia, "Human Skin Detection Using RGB, HSV and YCbCr Color Models," in Computer Vision and Pattern Recognition, 2017.
[25] "Unity Assert Store - Workplace Tools," [Online]. Available: https://assetstore.unity.com/packages/3d/workplace-tools-86242. [Accessed 19 - Jun - 2018].
[26] "OpenCV," [Online]. Available: https://opencv.org/. [Accessed 6 - Jun - 2018].