Multilingual JobBERT for Cross-Lingual Job Title Matching
By: Jens-Joris Decorte, Matthias De Lange, Jeroen Van Hautte
Potential Business Impact:
Matches job titles in different languages.
We introduce JobBERT-V3, a contrastive learning-based model for cross-lingual job title matching. Building on the state-of-the-art monolingual JobBERT-V2, our approach extends support to English, German, Spanish, and Chinese by leveraging synthetic translations and a balanced multilingual dataset of over 21 million job titles. The model retains the efficiency-focused architecture of its predecessor while enabling robust alignment across languages without requiring task-specific supervision. Extensive evaluations on the TalentCLEF 2025 benchmark demonstrate that JobBERT-V3 outperforms strong multilingual baselines and achieves consistent performance across both monolingual and cross-lingual settings. While not the primary focus, we also show that the model can be effectively used to rank relevant skills for a given job title, demonstrating its broader applicability in multilingual labor market intelligence. The model is publicly available: https://huggingface.co/TechWolf/JobBERT-v3.
Similar Papers
Lingua Custodi's participation at the WMT 2025 Terminology shared task
Computation and Language
Lets computers understand sentences in many languages.
Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management
Computation and Language
Helps companies find and train workers fairly.
Mitigating Language Bias in Cross-Lingual Job Retrieval: A Recruitment Platform Perspective
Computation and Language
Helps job sites match people to jobs better.