Gesture recognition method integrating multimodal inter-frame motion and shared attention weights
Guardado en:
| Publicado en: | Discover Artificial Intelligence vol. 5, no. 1 (Dec 2025), p. 405 |
|---|---|
| Autor principal: | |
| Publicado: |
Springer Nature B.V.
|
| Materias: | |
| Acceso en línea: | Citation/Abstract Full Text Full Text - PDF |
| Etiquetas: |
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
| Resumen: | To improve the accuracy and precision of gesture recognition, this study improves YOLOv5 by incorporating a coordinate attention mechanism and a bidirectional feature pyramid network. Based on the improved YOLOv5, a static gesture recognition model is constructed. In addition, this study introduces a multimodal inter-frame motion attention weight module to enhance the model’s ability to recognize dynamic gestures. In the performance evaluation experiments, the proposed model achieves an area under the receiver operating characteristic curve of 0.94, a harmonic mean of 96.4%, and an intersection over union of 0.9. The accuracy of static gesture recognition reaches 100%, while the average accuracy of dynamic gesture recognition achieves 95.7%, which significantly outperforms the comparison models. These results demonstrate that the proposed gesture recognition model offers high accuracy for static gestures and reliable recognition performance for dynamic gestures. This approach provides a potential method and perspective for improving human–computer interaction in virtual reality and intelligent assistance scenarios. |
|---|---|
| ISSN: | 2731-0809 |
| DOI: | 10.1007/s44163-025-00653-7 |
| Fuente: | Research Library |