--- license: openrail language: - is tags: - bias-detection - icelandic - ner - socially-responsible-ai - prejudice-detection - huggingface - dataset --- # Icelandic Bias-Aware NER Dataset **Trigger warning:** This dataset contains biased, offensive, or harmful language. Examples are included solely for research purposes. ## Dataset Description This dataset contains Icelandic sentences annotated for biased and potentially harmful expressions across 14 categories. It was developed to support research in fairness-oriented NLP, especially for low-resource languages. ### Classes - **B-ADDICTION, I-ADDICTION** - **B-DISABILITY, I-DISABILITY** - **B-ORIGIN, I-ORIGIN** - **B-GENERAL, I-GENERAL** - **B-LGBTQIA, I-LGBTQIA** - **B-LOOKS, I-LOOKS** - **B-PERSONAL, I-PERSONAL** - **B-PROFANITY, I-PROFANITY** - **B-RELIGION, I-RELIGION** - **B-SEXUAL, I-SEXUAL** - **B-SOCIAL_STATUS, I-SOCIAL_STATUS** - **B-STUPIDITY, I-STUPIDITY** - **B-VULGAR, I-VULGAR** - **B-WOMEN, I-WOMEN** Annotations follow the BIO scheme (e.g., `B-WOMEN`, `I-WOMEN`, `O`). ### Contents - **all_balanced.txt**: 184,580 sentence examples created via weak supervision (lemmatized string matching against curated bias lexicon) -**dev.txt**: A development dataset sampled from all_balanced.txt. It includes 15,381 sentence examples. -**test.txt**: A testing dataset sampled from all_balanced.txt. It includes 15,383 sentence examples. -**train.txt**: A training dataset sampled from all_balanced.txt. It includes 153,816 sentence examples -**gold.txt** **Automatically annotated set**: 15,383 sentences created via weak supervision (lemmatized string matching against curated bias lexicon) - **Gold set**: 190 manually reviewed sentence examples from sources not included in the training, testing or development sets. ## Intended Uses & Limitations ### Intended Use - Research on bias detection in Icelandic text - Training and evaluation of bias-aware NLP models - Educational purposes for raising awareness of bias in language ### Limitations - Automatically annotated examples may include false positives and false negatives - Vocabulary-based matching may miss subtle, euphemistic, or emerging forms of bias - Gold set is relatively small ⚠ **Not intended for punitive monitoring or censorship.** Outputs are prompts for reflection, not judgments. ## Ethical Considerations This dataset is released under the **[BigScience OpenRAIL-D License](https://www.licenses.ai/ai-licenses)**, which allows free use with responsible-use restrictions. Prohibited uses include: - Harassment or discrimination - Generating disinformation or hateful content - Surveillance targeting individuals or groups The dataset includes harmful language and should be handled with care. Trigger warnings are recommended for any public deployment. ## Citation Will be added.