MLSNT / README.md
RSTZZZZ's picture
Update README.md
ad2bfd4 verified
metadata
license: mit
task_categories:
  - text-classification
  - token-classification
language:
  - zh
  - ja
  - pt
  - fr
  - de
  - ru
tags:
  - toxicity
  - hatespeech
pretty_name: Multi-Lingual Social Network Toxicity
size_categories:
  - 100K<n<1M

MLSNT: Multi-Lingual Social Network Toxicity Dataset

MLSNT is a multi-lingual dataset for toxicity detection created through a large language model-assisted label transfer pipeline. It enables efficient and scalable moderation across languages and platforms, and is built to support span-level and category-specific classification for toxic content.

This dataset is introduced in the following paper:

Unified Game Moderation: Soft-Prompting and LLM-Assisted Label Transfer for Resource-Efficient Toxicity Detection
πŸ† Accepted at KDD 2025, Applied Data Science Track


🧩 Overview

MLSNT harmonizes 15 publicly available toxicity datasets across 7 languages using GPT-4o-mini to create consistent binary and fine-grained labels. It is suitable for both training and evaluating toxicity classifiers in multi-lingual, real-world moderation systems.


🌍 Supported Languages

  • πŸ‡«πŸ‡· French (fr)
  • πŸ‡©πŸ‡ͺ German (de)
  • πŸ‡΅πŸ‡Ή Portuguese (pt)
  • πŸ‡·πŸ‡Ί Russian (ru)
  • πŸ‡¨πŸ‡³ Simplified Chinese (zh-cn)
  • πŸ‡ΉπŸ‡Ό Traditional Chinese (zh-tw)
  • πŸ‡―πŸ‡΅ Japanese (ja)

πŸ—οΈ Construction Method

  1. Source Datasets
    15 human-annotated datasets were gathered from hatespeechdata.com and peer-reviewed publications.

  2. LLM-Assisted Label Transfer
    GPT-4o-mini was prompted to re-annotate each instance into a unified label schema. Only examples where human and LLM annotations agreed were retained.

  3. Toxicity Categories
    Labels are fine-grained categories (e.g., threat, hate, harassment).


πŸ“Š Dataset Statistics

Language Total Samples % Discarded Toxic % (Processed)
German (HASOC, etc.) ~13,800 28–69% 32–56%
French (MLMA) ~3,200 20% 94%
Russian ~14,300 ~40% 33–54%
Portuguese ~21,000 20–44% 26–50%
Japanese ~2,000 10–25% 17–45%
Chinese (Simplified) ~34,000 29–46% 48–61%
Chinese (Traditional) ~65,000 37% ~9%

πŸ’Ύ Format

Each row in the dataset includes:

  • full_text: The original utterance or message.
  • start_string_index: A list of start string indices (start positions of toxic spans).
  • end_string_index: A list of end string indices (end positions of toxic spans).
  • category_id: A list of toxic category IDs (integer values).
  • final_label: A list of toxic category names (string values).
  • min_category_id: The minimum toxic category ID in the row (used as the primary label).
  • match_id: A unique identifier composed of the original dataset name and a row-level ID.

πŸ—‚οΈ Category ID Mapping

ID Friendly Name
0 Non Toxic
1 Threats (Life Threatening)
2 Minor Endangerment
3 Threats (Non-Life Threatening)
4 Hate
5 Sexual Content / Harassment
6 Extremism
7 Insults
8 Controversial / Potentially Toxic Topic

πŸ”¬ Applications

  • Fine-tuning multi-lingual moderation systems
  • Cross-lingual toxicity benchmarking
  • Training span-level and category-specific toxicity detectors
  • Studying LLM label transfer reliability and agreement filtering

πŸ™ Acknowledgments

We thank Ubisoft La Forge, Ubisoft Montreal, and the Ubisoft Data Office for their technical support and valuable feedback throughout this project.

This work was supported by Ubisoft, the CIFAR AI Chair Program, and the Natural Sciences and Engineering Research Council of Canada (NSERC).

πŸ“œ Citation

If you use MLSNT in academic work, please cite:

@inproceedings{yang2025mlsnt,
  title={Unified Game Moderation: Soft-Prompting and LLM-Assisted Label Transfer for Resource-Efficient Toxicity Detection},
  author={Zachary Yang and Domenico Tullo and Reihaneh Rabbany},
  booktitle={Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD)},
  year={2025}
}