TalentCLEF 2025 corpus: Skill and Job Title Intelligence for Human Capital Management
Description
π¨ Current Status: Release of Task B Development set. To check when new data will be uploaded, please consult the calendar of the task
TalentCLEF2025 corpus - Task B Development set release
Introduction:
The first edition of TalentCLEF aims to develop and evaluate models designed to facilitate three essential tasks:
- Finding/ranking candidates for job positions based on their experience and professional skills.
- Implementing upskilling and reskilling strategies that promote the coninuous development of workers
- Detecting emerging skills and skills gaps of importance in organizations.
With that aim, the task is divided into two tasks:
- Task A - Multilingual Job Title Matching. This task involves developing systems to identify and rank the job titles most similar to a given one by generating a ranked list of similar titles from a specified knowledge base for each job title in a provided test set.
- Task B - Job Title-Based Skill Prediction. Task B requires developing systems that can retrieve relevant skills associated with a specified job title.
This data repository contains the data for these two tasks. The data is being released progressively according to the task schedule.
The task evaluation takes place on Codabench (Task A and Task B). Participants must register for the competition through CLEF Lab Registration Page to be part of the evaluation campaign.
File structure:
For a detailed description of the data structure, you can refer to the TalentCLEF2025 data description page, where it is thoroughly explained.
The files is organized into two *.zip
files, TaskA.zip
and TaskB.zip
, each containing training, validation and test folders to support different stages of model development. So far, only the training set for both tasks has been released, but in future releases, as the tasks progress, additional data will be added to the different subfolders for each task.
TaskA includes language-specific subfolders within the training and validation directories, covering English, Spanish, German, and Chinese job title data. The tr*aining folders for TaskA contain language-specific .tsv files for each respective language. Validation folders include three essential filesβqueries, corpus_elements, and q_relsβfor evaluating model relevance to search queries. TaskAβs test folder has queries and corpus_elements files for testing retrieval.
TaskA/
β
βββ training/
β βββ english/
β β βββ taskA_training_en.tsv
β βββ spanish/
β β βββ taskA_training_es.tsv
β βββ german/
β βββ taskA_training_de.tsv
β
βββ validation/
β βββ english/
β β βββ queries
β β βββ corpus_elements
β β βββ qrels
β βββ spanish/
β βββ german/
β βββ chinese/
β
βββ test/
βββ queries
βββ corpus_elements
TaskB follows a similar structure but without language-specific subfolders, providing general .tsv files for training, validation, and testing. This consistent file organization enables efficient data access and structured updates as new data versions are published.
TaskB/
β
βββ training/
β βββ job2skill.tsv
β βββ jobid2terms.json
β βββ skillid2terms.json
β
βββ validation/
β βββ queries
β βββ corpus_elements
β βββ qrels
β
βββ test/
βββ queries
βββ corpus_elements
Tutorials:
Notebook | Link |
Data Download and Load using Python | Link to Colab |
Prepare submission file and run evaluation | Link to Colab |
Task A - Development set Baseline generation | Link to Colab |
Resources: