IceBiasNER / README.md
steinunnfridriks's picture
Upload README.md
19877a2 verified
---
license: openrail
language:
- is
tags:
- bias-detection
- icelandic
- ner
- socially-responsible-ai
- prejudice-detection
- huggingface
- dataset
---
# Icelandic Bias-Aware NER Dataset
**Trigger warning:** This dataset contains biased, offensive, or harmful language. Examples are included solely for research purposes.
## Dataset Description
This dataset contains Icelandic sentences annotated for biased and potentially harmful expressions across 14 categories. It was developed to support research in fairness-oriented NLP, especially for low-resource languages.
### Classes
- **B-ADDICTION, I-ADDICTION**
- **B-DISABILITY, I-DISABILITY**
- **B-ORIGIN, I-ORIGIN**
- **B-GENERAL, I-GENERAL**
- **B-LGBTQIA, I-LGBTQIA**
- **B-LOOKS, I-LOOKS**
- **B-PERSONAL, I-PERSONAL**
- **B-PROFANITY, I-PROFANITY**
- **B-RELIGION, I-RELIGION**
- **B-SEXUAL, I-SEXUAL**
- **B-SOCIAL_STATUS, I-SOCIAL_STATUS**
- **B-STUPIDITY, I-STUPIDITY**
- **B-VULGAR, I-VULGAR**
- **B-WOMEN, I-WOMEN**
Annotations follow the BIO scheme (e.g., `B-WOMEN`, `I-WOMEN`, `O`).
### Contents
- **all_balanced.txt**: 184,580 sentence examples created via weak supervision (lemmatized string matching against curated bias lexicon)
-**dev.txt**: A development dataset sampled from all_balanced.txt. It includes 15,381 sentence examples.
-**test.txt**: A testing dataset sampled from all_balanced.txt. It includes 15,383 sentence examples.
-**train.txt**: A training dataset sampled from all_balanced.txt. It includes 153,816 sentence examples
-**gold.txt**
**Automatically annotated set**: 15,383 sentences created via weak supervision (lemmatized string matching against curated bias lexicon)
- **Gold set**: 190 manually reviewed sentence examples from sources not included in the training, testing or development sets.
## Intended Uses & Limitations
### Intended Use
- Research on bias detection in Icelandic text
- Training and evaluation of bias-aware NLP models
- Educational purposes for raising awareness of bias in language
### Limitations
- Automatically annotated examples may include false positives and false negatives
- Vocabulary-based matching may miss subtle, euphemistic, or emerging forms of bias
- Gold set is relatively small
**Not intended for punitive monitoring or censorship.** Outputs are prompts for reflection, not judgments.
## Ethical Considerations
This dataset is released under the **[BigScience OpenRAIL-D License](https://www.licenses.ai/ai-licenses)**, which allows free use with responsible-use restrictions.
Prohibited uses include:
- Harassment or discrimination
- Generating disinformation or hateful content
- Surveillance targeting individuals or groups
The dataset includes harmful language and should be handled with care. Trigger warnings are recommended for any public deployment.
## Citation
Will be added.