3DCNN Performance in Hand Gesture Recognition Applied to Robot Arm Interaction
Empreu sempre aquest identificador per citar o enllaçar aquest ítem
http://hdl.handle.net/10045/90590
Títol: | 3DCNN Performance in Hand Gesture Recognition Applied to Robot Arm Interaction |
---|---|
Autors: | Castro-Vargas, John Alejandro | Zapata-Impata, Brayan S. | Gil, Pablo | Garcia-Rodriguez, Jose | Torres, Fernando |
Grups d'investigació o GITE: | Automática, Robótica y Visión Artificial | Informática Industrial y Redes de Computadores |
Centre, Departament o Servei: | Universidad de Alicante. Departamento de Física, Ingeniería de Sistemas y Teoría de la Señal | Universidad de Alicante. Departamento de Tecnología Informática y Computación | Universidad de Alicante. Instituto Universitario de Investigación Informática |
Paraules clau: | Gesture Recognition from Video | 3D Convolutional Neural Network | Interaction human-robot |
Àrees de coneixement: | Ingeniería de Sistemas y Automática | Arquitectura y Tecnología de Computadores |
Data de publicació: | 19-de febrer-2019 |
Editor: | SciTePress |
Citació bibliogràfica: | Castro-Vargas, J.; Zapata-Impata, B.; Gil, P.; Garcia-Rodriguez, J. and Torres, F. (2019). 3DCNN Performance in Hand Gesture Recognition Applied to Robot Arm Interaction.In Proceedings of the 8th International Conference on Pattern Recognition Applications and Methods - Volume 1: ICPRAM, ISBN 978-989-758-351-3, pages 802-806. DOI: 10.5220/0007570208020806 |
Resum: | In the past, methods for hand sign recognition have been successfully tested in Human Robot Interaction (HRI) using traditional methodologies based on static image features and machine learning. However, the recognition of gestures in video sequences is a problem still open, because current detection methods achieve low scores when the background is undefined or in unstructured scenarios. Deep learning techniques are being applied to approach a solution for this problem in recent years. In this paper, we present a study in which we analyse the performance of a 3DCNN architecture for hand gesture recognition in an unstructured scenario. The system yields a score of 73% in both accuracy and F1. The aim of the work is the implementation of a system for commanding robots with gestures recorded by video in real scenarios. |
Patrocinadors: | This work was funded by the Ministry of Economy, Industry and Competitiveness from the Spanish Government through the DPI2015-68087-R and the pre-doctoral grant BES-2016-078290, by the European Commission and FEDER funds through the project COMMANDIA (SOE2/P1/F0638), action supported by Interreg-V Sudoe. |
URI: | http://hdl.handle.net/10045/90590 |
ISBN: | 978-989-758-351-3 |
DOI: | 10.5220/0007570208020806 |
Idioma: | eng |
Tipus: | info:eu-repo/semantics/conferenceObject |
Drets: | © 2019 by SCITEPRESS – Science and Technology Publications, Lda. |
Revisió científica: | si |
Versió de l'editor: | https://doi.org/10.5220/0007570208020806 |
Apareix a la col·lecció: | INV - AUROVA - Comunicaciones a Congresos Internacionales INV - AIA - Comunicaciones a Congresos, Conferencias, etc. INV - I2RC - Comunicaciones a Congresos, Conferencias, etc. |
Arxius per aquest ítem:
Arxiu | Descripció | Tamany | Format | |
---|---|---|---|---|
ICPRAM_2019_118.pdf | Versión final (acceso restringido) | 1,75 MB | Adobe PDF | Obrir Sol·licitar una còpia |
ICPRAM_2019_118_preprint.pdf | Preprint (acceso abierto) | 1,56 MB | Adobe PDF | Obrir Vista prèvia |
Tots els documents dipositats a RUA estan protegits per drets d'autors. Alguns drets reservats.