Efficient Text Encoders for Labor Market Analysis
By: Jens-Joris Decorte , Jeroen Van Hautte , Chris Develder and more
Potential Business Impact:
Finds job skills faster and better.
Labor market analysis relies on extracting insights from job advertisements, which provide valuable yet unstructured information on job titles and corresponding skill requirements. While state-of-the-art methods for skill extraction achieve strong performance, they depend on large language models (LLMs), which are computationally expensive and slow. In this paper, we propose \textbf{ConTeXT-match}, a novel contrastive learning approach with token-level attention that is well-suited for the extreme multi-label classification task of skill classification. \textbf{ConTeXT-match} significantly improves skill extraction efficiency and performance, achieving state-of-the-art results with a lightweight bi-encoder model. To support robust evaluation, we introduce \textbf{Skill-XL}, a new benchmark with exhaustive, sentence-level skill annotations that explicitly address the redundancy in the large label space. Finally, we present \textbf{JobBERT V2}, an improved job title normalization model that leverages extracted skills to produce high-quality job title representations. Experiments demonstrate that our models are efficient, accurate, and scalable, making them ideal for large-scale, real-time labor market analysis.
Similar Papers
Resource-Efficient Adaptation of Large Language Models for Text Embeddings via Prompt Engineering and Contrastive Fine-tuning
Computation and Language
Makes computers understand whole sentences better.
Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management
Computation and Language
Helps companies find and train workers fairly.
Enhancing Talent Employment Insights Through Feature Extraction with LLM Finetuning
Computation and Language
Helps computers understand job descriptions better.