Skip to the content.

Mu-SHROOM

Welcome to the official shared task website for Mu-SHROOM, a SemEval-2025 shared task!

Mu-SHROOM stands for “Multilingual Shared-task on Hallucinations and Related Observable Overgeneration Mistakes”. Mu-SHROOM will invite participants to detect hallucination spans in the outputs of instruction-tuned LLMs in a multilingual context. This shared task builds upon our previous iteration, SHROOM, with a few key changes:

This website is under construction. More information will be available soon.

What is Mu-SHROOM?

The task consists in detecting spans of text corresponding to hallucinations. Participants are asked to determine which parts of a given text produced by LLMs constitute hallucinations. The task is held in multi-lingual and multi-model context, i.e., we provide data in multiple languages and produced by a variety of public-weights LLMs.´

In practice, we provide an LLM output (as a string of characters, a list of tokens, and a list of logits), and participants have to compute, for every character in the LLM output string, the probability that it is marked as a hallucination. Participants are free to use any approach they deem appropriate, including using external resources.

How will participants be evaluated?

Participants will be ranked along two (character-level) metrics:

  1. intersection-over-union of characters marked as hallucinations in the gold reference vs. predicted as such
  2. how well the probability assigned by the participants’ system that a character is part of a hallucination correlates with the empirical probabilities observed in our annotators.

Rankings and submissions will be done separately per language.

Participants can also download the scoring program on its own here for reference and developing their systems.

Participant info

Register ahead of time on our submission website

Want to be kept in the loop? Join our Google group mailing list or the shared task Slack! We also have a Twitter acount.

Data

Below are links to access the data already released, as well as provisional expected release dates for future splits. Do note that release dates are subject to change.

Dataset split Access
Sample set download (v1)
Validation set download (v2)
Unlabeled train set download (v1)
Unlabeled test set To be published (ETA Jan 10th)
Labeled test set To be published (ETA Feb 1st)

We are releasing a participant kit, which we’ll keep building up. It contains the scoring program as well as two baselines: a random baseline and an XLM-R based system. The latter is a very naive system and is mostly meant for demonstration purposes (in particular how to convert character-level labels into wordpiece-level tags and vice-versa).
You can download the participant kit from here.

Important dates

This information is subject to change.

Organizers of the shared task

Looking for something else?

The website for the previous iteration of the shared task is available here.

The logo is available here (download); we encourage participants to use it where relevant (esp. in your posters)!

Still have questions?

We have compiled a FAQ! You can also reach out via our communicaiton channels.