Welcome to the website of EXIST 2025, the fifth edition of the sEXism Identification in Social neTworks task at CLEF 2025.
EXIST is a series of scientific events and shared tasks on sexism identification in social networks. EXIST aims to capture sexism in a broad sense, from explicit misogyny to other subtle expressions that involve implicit sexist behaviours (EXIST 2021, EXIST 2022, EXIST 2023, EXIST 2024). The fifth edition of the EXIST shared task will be held as a Lab in CLEF 2025, on September 9-12, 2025, in UNED, Madrid, Spain.
Social Networks are the main platforms for social complaint, activism, etc. Movements like #MeTwoo, #8M or #Time’sUp have spread rapidly. Under the umbrella of social networks, many women all around the world have reported abuses, discriminations and other sexist experiences suffered in real life. Social networks are also contributing to the transmission of sexism and other disrespectful and hateful behaviours. In this context, automatic tools not only may help to detect and alert against sexism behaviours and discourses, but also to estimate how often sexist and abusive situations are found in social media platforms, what forms of sexism are more frequent and how sexism is expressed in these media. This Lab will contribute to developing applications to detect sexism.
In 2024 the EXIST campaing included multimedia content in the format of memes, steping forward research on more robust techniques to identify sexism in social networks. Following this line, this year we will focus on TikTok videos in the challenge including in the dataset the three more important sources of sexism spreading: text, images and videos. Sexism on TikTok is also a growing concern as the platform’s algorithm often amplifies content that perpetuates gender stereotypes and internalized misogyny. Consequently, it is essential to develop automated multimodal tools capable of detecting sexism in text, images, and videos, to raise alarms or automatically remove such content from social networks. This lab will contribute to the creation of applications that identify sexist content in social media across all three formats.
Similar to the approach in the 2023 and 2024 edition, this edition will also embrace the Learning With Disagreement (LeWiDi) paradigm for both the development of the dataset and the evaluation of the systems. The LeWiDi paradigm doesn’t rely on a single “correct” label for each example. Instead, the model is trained to handle and learn from conflicting or diverse annotations. This enables the system to consider various annotators’ perspectives, biases, or interpretations, resulting in a fairer learning process.
In previous editions, 223 teams from more than 50 countries submitted their results achieving impressive results, especially in the sexism detection task. However, there is still room for improvement, especially in when the problem is addressed under the LeWeDi paradigm in a multimedia context.
Participants will be asked to identify and characterize sexism in social networks according to different sources: This year the lab comprises nine subtasks in two languages, English and Spanish, which are the same three tasks (sexism identification, source intention detection, and sexism categorization) applied to three different types of data: text (tweets), image (memes) and video (TikToks). This multimedia approach will help identify trends and patterns in sexism across media formats and user interactions, contributing to a deeper understanding of the social dynamics. Also, approches submitted to all tasks will be evaluated to analyze their capacity to detect sexism in a multimodal source.
A condense schema of all tasks included this year in the lab is presented in the following table:
For a more detailed description of each subtask, as well as some examples, check the next sections.
The first subtask is a binary classification. The systems have to decide whether or not a given tweet contains sexist expressions or behaviours (i.e., it is sexist itself, describes a sexist situation or criticizes a sexist behaviour), and classify it according to two categories: YES and NO.
Once a message has been classified as sexist, the second subtask aims to categorize the message according to the intention of the author, which provides insights in the role played by social networks on the emission and dissemination of sexist messages. In this subtask, we propose a ternary classification task:
DIRECT: the intention was to write a message that is sexist by itself or incites to be sexist, as in:
REPORTED: the intention is to report and share a sexist situation suffered by a woman or women in first or third person, as in:
JUDGEMENTAL: the intention was to judge, since the tweet describes sexist situations or behaviours with the aim of condemning them.
Many facets of a woman’s life may be the focus of sexist attitudes including domestic and parenting roles, career opportunities, sexual image, and life expectations, to name a few. Automatically detecting which of these facets of women are being more frequently attacked in social networks will facilitate the development of policies to fight against sexism. According to this, each sexist tweet must be categorized in one or more of the following categories
IDEOLOGICAL AND INEQUALITY: The text discredits the feminist movement, rejects inequality between men and women, or presents men as victims of gender-based oppression.
STEREOTYPING AND DOMINANCE: The text expresses false ideas about women that suggest they are more suitable to fulfill certain roles (mother, wife, family caregiver, faithful, tender, loving, submissive, etc.), or inappropriate for certain tasks (driving, hardwork, etc), or claims that men are somehow superior to women.
OBJECTIFICATION: The text presents women as objects apart from their dignity and personal aspects, or assumes or describes certain physical qualities that women must have in order to fulfill traditional gender roles (compliance with beauty standards, hypersexualization of female attributes, women’s bodies at the disposal of men, etc.).
SEXUAL VIOLENCE: Sexual suggestions, requests for sexual favors or harassment of a sexual nature (rape or sexual assault) are made.
MISOGYNY AND NON-SEXUAL VIOLENCE: The text expressses hatred and violence towards women.
This is a binary classification subtask consisting on determine wheter a meme describes a sexist situation or criticizes a sexist behaviour), and classify it according to two categories: YES and NO. The following figures are some examples of both types of memes, respectively.
As in subtask 1.2, this subtask aims to categorize the meme according to the intention of the author, which provides insights in the role played by social networks on the emission and dissemination of sexist messages. Due to the characteristics of the memes, the REPORTED label is virtually null, so in this task systems should only classify memes with DIRECT or JUDGEMENTAL labels. The following figures are some examples of them, respectively.
This task aims to classify sexist memes according to the categorization provided for subtask 1.3: (i) IDEOLOGICAL AND INEQUALITY, (ii) STEREOTYPING AND DOMINANCE, (iii) OBJECTIFICATION, (iv) SEXUAL VIOLENCE and (v) MISOGYNY AND NON-SEXUAL VIOLENCE. The following figures are some examples of categorized memes.
(a) Stereotyping
(e) Ideological
(c) Objectification
(d) Misogyny
(b) Sexual violence
This subtask is the same subtask as subtasks 1.1 and 2.1. The following figures are some examples of videos classified as YES or NO.
This subtask replicates the subtask 2.2 for memes, but it takes as source videos. The following examples are some videos representing each category.
This subtask aims to classify sexist memes according to the categorization provided for Task 1.3: (i) IDEOLOGICAL AND INEQUALITY, (ii) STEREOTYPING AND DOMINANCE, (iii) OBJECTIFICATION, (iv) SEXUAL VIOLENCE and (v) MISOGYNY AND NON-SEXUAL VIOLENCE. The following figures are some examples of categorized videos.
To be announced!
To be announced!
For any question that concern the shared task, please write to Jorge Carrillo-de-Albornoz.