| dc.contributor.author | Amangeldy, Nurzada | |
| dc.contributor.author | Gazizova, Nazerke | |
| dc.contributor.author | Milosz, Marek | |
| dc.contributor.author | Kurmetbek, Bekbolat | |
| dc.contributor.author | Nazyrova, Aizhan | |
| dc.contributor.author | Kassymova, Akmaral | |
| dc.date.accessioned | 2026-02-23T06:28:32Z | |
| dc.date.available | 2026-02-23T06:28:32Z | |
| dc.date.issued | 2025 | |
| dc.identifier.citation | Amangeldy, N.; Gazizova, N.; Milosz, M.; Kurmetbek, B.; Nazyrova, A.; Kassymova, A. Unsupervised Clustering and Ensemble Learning for Classifying Lip Articulation in Fingerspelling. Sensors 2025, 25, 3703. https://doi.org/ 10.3390/s25123703 | ru |
| dc.identifier.issn | 1424-8220 | |
| dc.identifier.other | doi.org/ 10.3390/s25123703 | |
| dc.identifier.uri | http://repository.enu.kz/handle/enu/29303 | |
| dc.description.abstract | This paper presents a new methodology for analyzing lip articulation during fingerspelling aimed at extracting robust visual patterns that can overcome the inherent ambiguity and variability of lip shape. The proposed approach is based on unsupervised clustering of lip movement trajectories to identify consistent articulatory patterns across different time profiles. The methodology is not limited to using a single model. Still, it includes the exploration of varying cluster configurations and an assessment of their robustness, as well as a detailed analysis of the correspondence between individual alphabet letters and specific clusters. In contrast to direct classification based on raw visual features, this approach pre-tests clustered representations using a model-based assessment of their discriminative potential. This structured approach enhances the interpretability and robustness of the extracted features, highlighting the importance of lip dynamics as an auxiliary modality in multimodal sign language recognition. The obtained results demonstrate that trajectory clustering can serve as a practical method for generating features, providing more accurate and context-sensitive gesture interpretation. | ru |
| dc.language.iso | en | ru |
| dc.publisher | Sensors | ru |
| dc.relation.ispartofseries | 25, 3703; | |
| dc.subject | lip articulation | ru |
| dc.subject | trajectory clustering | ru |
| dc.subject | fingerspelling | ru |
| dc.subject | unsupervised learning | ru |
| dc.subject | multimodal recognition | ru |
| dc.subject | visual speech | ru |
| dc.title | Unsupervised Clustering and Ensemble Learning for Classifying Lip Articulation in Fingerspelling | ru |
| dc.type | Article | ru |