Ajou University repository

Attentional bias for hands: Cascade dual-decoder transformer for sign language productionoa mark
Citations

SCOPUS

2

Citation Export

DC Field Value Language
dc.contributor.authorMa, Xiaohan-
dc.contributor.authorJin, Rize-
dc.contributor.authorWang, Jianming-
dc.contributor.authorChung, Tae Sun-
dc.date.issued2024-08-01-
dc.identifier.urihttps://dspace.ajou.ac.kr/dev/handle/2018.oak/34018-
dc.description.abstractSign Language Production (SLP) refers to the task of translating textural forms of spoken language into corresponding sign language expressions. Sign languages convey meaning by means of multiple asynchronous articulators, including manual and non-manual information channels. Recent deep learning-based SLP models directly generate the full-articulatory sign sequence from the text input in an end-to-end manner. However, these models largely down weight the importance of subtle differences in the manual articulation due to the effect of regression to the mean. To explore these neglected aspects, an efficient cascade dual-decoder Transformer (CasDual-Transformer) for SLP is proposed to learn, successively, two mappings SLPhand: Text → Hand pose and SLPsign: Text → Sign pose, utilising an attention-based alignment module that fuses the hand and sign features from previous time steps to predict more expressive sign pose at the current time step. In addition, to provide more efficacious guidance, a novel spatio-temporal loss to penalise shape dissimilarity and temporal distortions of produced sequences is introduced. Experimental studies are performed on two benchmark sign language datasets from distinct cultures to verify the performance of the proposed model. Both quantitative and qualitative results show that the authors’ model demonstrates competitive performance compared to state-of-the-art models, and in some cases, achieves considerable improvements over them.-
dc.description.sponsorshipThis work was supported by the Institute of Information & Communications Technology Planning & Evaluation (IITP) under the Artificial Intelligence Convergence Innovation Human Resources Development (IITP-2023-RS-2023-00255968) grant and the ITRC (Information Technology Research Center) support program (IITP-2021-0-02051) funded by the Korea government (MSIT).-
dc.description.sponsorshipThis work was supported by the Institute of Information & Communications Technology Planning & Evaluation (IITP) under the Artificial Intelligence Convergence Innovation Human Resources Development (IITP\u20102023\u2010RS\u20102023\u201000255968) grant and the ITRC (Information Technology Research Center) support program (IITP\u20102021\u20100\u201002051) funded by the Korea government (MSIT).-
dc.language.isoeng-
dc.publisherJohn Wiley and Sons Inc-
dc.subject.meshInformation channels-
dc.subject.meshLanguage processing-
dc.subject.meshLanguage production-
dc.subject.meshNatural language processing-
dc.subject.meshNatural languages-
dc.subject.meshPose-estimation-
dc.subject.meshSign language-
dc.subject.meshSign language production-
dc.subject.meshSpoken languages-
dc.subject.meshTime step-
dc.titleAttentional bias for hands: Cascade dual-decoder transformer for sign language production-
dc.typeArticle-
dc.citation.endPage708-
dc.citation.startPage696-
dc.citation.titleIET Computer Vision-
dc.citation.volume18-
dc.identifier.bibliographicCitationIET Computer Vision, Vol.18, pp.696-708-
dc.identifier.doi10.1049/cvi2.12273-
dc.identifier.scopusid2-s2.0-85187194070-
dc.identifier.urlhttps://ietresearch.onlinelibrary.wiley.com/journal/17519640-
dc.subject.keywordcomputer vision-
dc.subject.keywordnatural language processing-
dc.subject.keywordpose estimation-
dc.subject.keywordsign language production-
dc.description.isoatrue-
dc.subject.subareaSoftware-
dc.subject.subareaComputer Vision and Pattern Recognition-
Show simple item record

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Chung, Tae-Sun Image
Chung, Tae-Sun정태선
Department of Software and Computer Engineering
Read More

Total Views & Downloads

File Download

  • There are no files associated with this item.