VOGUE-Based Approach for Segmenting Movement Epenthesis in Continuous Sign Language Recognition

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Ingenierie des Systemes d'Information vol. 30, no. 11 (Nov 2025), p. 2949-2960
1. Verfasser: Thillai, Sivakavi S
Weitere Verfasser: Minu, R I
Veröffentlicht:
International Information and Engineering Technology Association (IIETA)
Schlagworte:
Online-Zugang:Citation/Abstract
Full Text - PDF
Tags: Tag hinzufügen
Keine Tags, Fügen Sie das erste Tag hinzu!

MARC

LEADER 00000nab a2200000uu 4500
001 3293380337
003 UK-CbPIL
022 |a 1633-1311 
022 |a 2116-7125 
022 |a 1290-2926 
024 7 |a 10.18280/isi.301113  |2 doi 
035 |a 3293380337 
045 2 |b d20251101  |b d20251130 
100 1 |a Thillai, Sivakavi S 
245 1 |a VOGUE-Based Approach for Segmenting Movement Epenthesis in Continuous Sign Language Recognition 
260 |b International Information and Engineering Technology Association (IIETA)  |c Nov 2025 
513 |a Journal Article 
520 3 |a When developing a continuous sign language recognition (CSLR) system, a significant challenge lies in processing the vast number of video frames, which demands extensive time and computational resources during both the training and prediction phases. To address this, we propose an efficient and scalable methodology that integrates cluster-based key frame extraction with a VOGUE-based recognition model designed for continuous gestures. The key frame extraction strategy clusters visually similar frames to reduce redundancy while preserving only those with high semantic relevance. To further enhance recognition accuracy, we introduce the Key Curvature Maximum Point (KCMP) technique, which identifies pivotal motion points and captures essential hand trajectory changes inherent to sign language. These refined frames are subsequently used to train a VOGUE-based model that encodes spatial and temporal strokes dynamics, followed by probability distribution modeling for robust prediction. The proposed approach was evaluated using a custom-built Tamil Sign Language dataset. Performance was compared against several established baseline methods, including Dynamic Time Warping (DTW), Hidden Markov Models (HMM), and multiple Conditional Random Field (CRF) variants, as well as the VOM model. The system achieved a recognition accuracy of 86.78% and a sign error rate of 5.3%. A paired t-test confirmed that the improvements over baseline models were statistically significant (p < 0.05). These results demonstrate that the proposed framework provides improved efficiency and competitive accuracy, offering a promising solution for real-time CSLR applications, particularly in low-resource regional sign languages. 
653 |a Problem solving 
653 |a Accuracy 
653 |a Dynamic programming 
653 |a Markov chains 
653 |a Deafness 
653 |a Algorithms 
653 |a Sign language 
653 |a Frames (data processing) 
653 |a Real time 
653 |a Conditional random fields 
653 |a Recognition 
700 1 |a Minu, R I 
773 0 |t Ingenierie des Systemes d'Information  |g vol. 30, no. 11 (Nov 2025), p. 2949-2960 
786 0 |d ProQuest  |t Engineering Database 
856 4 1 |3 Citation/Abstract  |u https://www.proquest.com/docview/3293380337/abstract/embedded/75I98GEZK8WCJMPQ?source=fedsrch 
856 4 0 |3 Full Text - PDF  |u https://www.proquest.com/docview/3293380337/fulltextPDF/embedded/75I98GEZK8WCJMPQ?source=fedsrch