| 研究生: |
陳昌暐 Chang-Wei Chen |
|---|---|
| 論文名稱: |
多模態生物特徵神經網路分類器應用於寵物身分識別 Multi-Modal Biometric Neural Network Classifier for Pet Identity Authentication |
| 指導教授: | 陳慶瀚 |
| 口試委員: | |
| 學位類別: |
碩士 Master |
| 系所名稱: |
資訊電機學院 - 資訊工程學系 Department of Computer Science & Information Engineering |
| 論文出版年: | 2018 |
| 畢業學年度: | 106 |
| 語文別: | 中文 |
| 論文頁數: | 96 |
| 中文關鍵詞: | 多模態 |
| 相關次數: | 點閱:7 下載:0 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
目前政府對於寵物和流浪動物管理主要採用的方法是植入RFID晶片,用以識別寵物身分。由於採用侵入式手段,民眾為寵物植入晶片意願不高,導致管理漏洞。本研究針對以非侵入的影像辨識方法進行寵物身分識別,提出一個多模態生物辨識方法。我們先藉由CNN分類出犬種,再擷取寵物鼻紋、身形輪廓及臉部幾何等多模態生物特徵,最後藉由混合神經網路分類器進行犬隻身分識別。最後我們設計了一個多模態混合神經網路(MM-HNN)分類器系統來驗證寵物辨識的性能。實驗比較顯示,單純使用CNN深度學習模型從事寵物身分識別的等錯誤率為23.77%,MM-HNN寵物身分識別的等錯誤率為13.45%,而CNN犬種辨識加上MM-HNN身分識別的等錯誤率為4.65%;在三群同胞犬的實驗中,身分識別的等錯誤率為4.57%;在三種生物特徵模態的辨識實驗中,在紋理模態中辨識率為88.33%,輪廓模態中辨識率為84.83%,臉部模態中辨識率為79.83%,最後混合三種模態而成的多模態辨識率則達到95%,因此證明了多模態混合神經網路分類器具有良好的辨識性能。
At present, the government mainly uses embedded RFID chip to supervise pets and stray animals to identify the pets. As the means is invasive, common people' willingness to embed the chip in pets is not high, leading to administrative vulnerability. This study proposes a multi modal biological recognition method for using non-invasive image recognition method to identify pets. The dog species is classified by CNN (Convolutional Neural Network), and then the pet's multi modal biological features are extracted, such as muzzle pattern, body contour and facial geometry. Finally, the dog is identified by hybrid neural network classifier. A Multi Modal Hybrid Neural Network (MM-HNN) classifier system is designed to validate the performance of pet identification. The empirical comparison shows that the Equal Error Rate (EER) of simply using CNN deep learning model for pet identification is 23.77%, the EER of MM-HNN pet identification is 13.45%, and the EER of CNN dog species recognition + MM-HNN identification is 4.65%. In the experiment on three groups of fellow dog, the EER of identification is 4.57%. In the recognition experiment on three biological feature modals, the texture modal recognition rate is 88.33%, the contour modal recognition rate is 84.83%, the face mode recognition rate is 79.83%, and the multi modal recognition rate of the three modals is 95%, proving that the MM-HNN classifier has good recognition performance.
[1] J. Daugman, "How iris recognition works," in The essential guide to image processing: Elsevier, pp. 715-739, 2009.
[2] J. G. Daugman, "Biometric personal identification system based on iris analysis," ed: Google Patents, 1994.
[3] A. Ross, A. Jain, and J. Reisman, "A hybrid fingerprint matcher," Pattern Recognition, vol. 36, no. 7, pp. 1661-1673, 2003.
[4] R. P. Wildes et al., "Automated, non-invasive iris recognition system and method," ed: Google Patents, 1996.
[5] R.-L. Hsu, M. Abdel-Mottaleb, and A. K. Jain, "Face detection in color images," IEEE transactions on pattern analysis and machine intelligence, vol. 24, no. 5, pp. 696-706, 2002.
[6] L. Muda, M. Begam, and I. Elamvazuthi, "Voice recognition algorithms using mel frequency cepstral coefficient (MFCC) and dynamic time warping (DTW) techniques," arXiv preprint arXiv:1003.4083, 2010.
[7] A. S. Carpentier, C. Jean, M. Barret, A. Chassagneux, and S. Ciccione, "Stability of facial scale patterns on green sea turtles Chelonia mydas over time: A validation for the use of a photo-identification method," Journal of Experimental Marine Biology and Ecology, vol. 476, pp. 15-21, 2016.
[8] S. J. Carter, I. P. Bell, J. J. Miller, and P. P. Gash, "Automated marine turtle photograph identification using artificial neural networks, with application to green turtles," Journal of experimental marine biology and ecology, vol. 452, pp. 105-110, 2014.
[9] A. Marshall and S. Pierce, "The use and abuse of photographic identification in sharks and rays," Journal of fish biology, vol. 80, no. 5, pp. 1361-1379, 2012.
[10] C. W. Speed, M. G. Meekan, and C. J. Bradshaw, "Spot the match–wildlife photo-identification using information theory," Frontiers in zoology, vol. 4, no. 1, p. 2, 2007.
[11] R. Fergus, L. Fei-Fei, P. Perona, and A. Zisserman, "Learning object categories from internet image searches," Proceedings of the IEEE, vol. 98, no. 8, pp. 1453-1466, 2010.
[12] 周宛儀, "基於紋理分析的狗鼻孔影像切割," 中興大學資訊科學與工程學系學位論文, pp. 1-68, 2017.
[13] C. E. DeCamp, "Kinetic and kinematic gait analysis and the assessment of lameness in the dog," Veterinary Clinics: Small Animal Practice, vol. 27, no. 4, pp. 825-840, 1997.
[14] K. Delac and M. Grgic, "A survey of biometric recognition methods," in Electronics in Marine, 2004. Proceedings Elmar 2004. 46th International Symposium, pp. 184-193: IEEE, 2004.
[15] P. Sanjekar and J. Patil, "An overview of multimodal biometrics," Signal & Image Processing, vol. 4, no. 1, p. 57, 2013.
[16] M. Gudavalli, S. V. Raju, A. V. Babu, and D. S. Kumar, "Multimodal Biometrics--Sources, Architecture and Fusion Techniques: An Overview," in Biometrics and Security Technologies (ISBAST), 2012 International Symposium on, pp. 27-34: IEEE, 2012.
[17] D. Silver et al., "Mastering the game of Go with deep neural networks and tree search," nature, vol. 529, no. 7587, pp. 484-489, 2016.
[18] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, "Gradient-based learning applied to document recognition," Proceedings of the IEEE, vol. 86, no. 11, pp. 2278-2324, 1998.
[19] S. Ji, W. Xu, M. Yang, and K. Yu, "3D convolutional neural networks for human action recognition," IEEE transactions on pattern analysis and machine intelligence, vol. 35, no. 1, pp. 221-231, 2013.
[20] A. Karpathy, G. Toderici, S. Shetty, T. Leung, R. Sukthankar, and L. Fei-Fei, "Large-scale video classification with convolutional neural networks," in Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 1725-1732, 2014.
[21] O. Abdel-Hamid, A.-r. Mohamed, H. Jiang, L. Deng, G. Penn, and D. Yu, "Convolutional neural networks for speech recognition," IEEE/ACM Transactions on audio, speech, and language processing, vol. 22, no. 10, pp. 1533-1545, 2014.
[22] S. Lawrence, C. L. Giles, A. C. Tsoi, and A. D. Back, "Face recognition: A convolutional neural-network approach," IEEE transactions on neural networks, vol. 8, no. 1, pp. 98-113, 1997.
[23] J. K. Hawkins, "Textural properties for pattern recognition," Picture processing and psychopictorics, pp. 347-370, 1970.
[24] R. M. Haralick and K. Shanmugam, "Textural features for image classification," IEEE Transactions on systems, man, and cybernetics, no. 6, pp. 610-621, 1973.
[25] T. Ojala, M. Pietikainen, and T. Maenpaa, "Multiresolution gray-scale and rotation invariant texture classification with local binary patterns," IEEE Transactions on pattern analysis and machine intelligence, vol. 24, no. 7, pp. 971-987, 2002.
[26] H. Tamura, S. Mori, and T. Yamawaki, "Textural features corresponding to visual perception," IEEE Transactions on Systems, man, and cybernetics, vol. 8, no. 6, pp. 460-473, 1978.
[27] D. Zhang and G. Lu, "Shape-based image retrieval using generic Fourier descriptor," Signal Processing: Image Communication, vol. 17, no. 10, pp. 825-848, 2002.
[28] D. F. Specht, "Probabilistic neural networks," Neural networks, vol. 3, no. 1, pp. 109-118, 1990.
[29] A. Alemi, "Improving inception and image classification in tensorflow," G oogle R esearch B log (https://research. googleblog. com/2016/08/improving-inception-and-image. html), 2016.
[30] A. Khosla, N. Jayadevaprakash, B. Yao, and F.-F. Li, "Novel dataset for fine-grained image categorization: Stanford dogs," in Proc. CVPR Workshop on Fine-Grained Visual Categorization (FGVC), vol. 2, p. 1, 2011.
[31] C.-H. Chen, M.-Y. Lin, and X.-C. Guo, "High-level modeling and synthesis of smart sensor networks for Industrial Internet of Things," Computers & Electrical Engineering, vol. 61, pp. 48-66, 2017.