CSF: Contrastive Semantic Features for Direct Multilingual Sign Language Generation
By: Tran Sy Bao
Potential Business Impact:
Translates any language directly to sign language.
Sign language translation systems typically require English as an intermediary language, creating barriers for non-English speakers in the global deaf community. We present Canonical Semantic Form (CSF), a language-agnostic semantic representation framework that enables direct translation from any source language to sign language without English mediation. CSF decomposes utterances into nine universal semantic slots: event, intent, time, condition, agent, object, location, purpose, and modifier. A key contribution is our comprehensive condition taxonomy comprising 35 condition types across eight semantic categories, enabling nuanced representation of conditional expressions common in everyday communication. We train a lightweight transformer-based extractor (0.74 MB) that achieves 99.03% average slot extraction accuracy across four typologically diverse languages: English, Vietnamese, Japanese, and French. The model demonstrates particularly strong performance on condition classification (99.4% accuracy) despite the 35-class complexity. With inference latency of 3.02ms on CPU, our approach enables real-time sign language generation in browser-based applications. We release our code, trained models, and multilingual dataset to support further research in accessible sign language technology.
Similar Papers
Spatio-temporal Sign Language Representation and Translation
Computation and Language
Translates sign language into written words.
Beyond Words: AuralLLM and SignMST-C for Sign Language Production and Bidirectional Accessibility
CV and Pattern Recognition
Helps deaf people talk to hearing people.
Towards Fine-Grained Code-Switch Speech Translation with Semantic Space Alignment
Computation and Language
Translates mixed-language speech accurately and efficiently.