File size: 4,937 Bytes
790cca2 87927ad 15dae38 87927ad 8794657 87927ad 790cca2 15dae38 790cca2 ff1a585 5cf4c03 76b2d26 5cf4c03 15dae38 8794657 76b2d26 790cca2 115964b 928972b 87927ad 928972b 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 b79f015 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 115964b 790cca2 115964b 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 15dae38 790cca2 928972b 15dae38 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 |
---
base_model:
- FacebookAI/roberta-base
datasets:
- mediabiasgroup/BABE
language:
- en
library_name: transformers
license: apache-2.0
metrics:
- f1
pipeline_tag: text-classification
tags:
- Bias Detection
- Text Classification
Author:
- Himel Ghosh
---
## Citation
Paper Link: https://arxiv.org/abs/2505.13010
If you are using this model, please cite this paper:
```bibtex
@misc{ghosh2025biasbiasdetectingbias,
title={To Bias or Not to Bias: Detecting bias in News with bias-detector},
author={Himel Ghosh and Ahmed Mosharafa and Georg Groh},
year={2025},
eprint={2505.13010},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2505.13010},
}
```
This is a RoBERTa-based binary classification model fine-tuned on the BABE (URL: https://huggingface.co/datasets/mediabiasgroup/BABE)
dataset for bias detection in English news statements.
The model predicts whether a given sentence contains biased language (LABEL_1) or is unbiased (LABEL_0).
It is intended for applications in media bias analysis, content moderation, and social computing research.
- Example usage with Hugging Face’s pipeline:
```python
from transformers import pipeline
classifier = pipeline("text-classification", model="himel7/bias-detector", tokenizer="roberta-base")
result = classifier("Immigrants are criminals.")
```
## Evaluation
The model was evaluated on the entire BABE dataset with a K-fold Cross Validation and yielded the following metrics at K=5:
- **Accuracy: 0.9202**
- **Precision: 0.9615**
- **Recall: 0.8927**
- **F1 Score: 0.9257**
## Model Details
### Model Description
This model is a fine-tuned version of roberta-base trained to detect linguistic bias in English-language news statements.
The task is framed as binary classification: the model outputs LABEL_1 for biased statements and LABEL_0 for non-biased statements.
Fine-tuning was performed on the BABE dataset, which contains annotated news snippets across various topics and political leanings.
The annotations focus on whether the language used expresses subjective bias rather than factual reporting.
The goal of this model is to assist in detecting subtle forms of bias in media content, such as emotionally loaded language, stereotypical
phrasing, or exaggerated claims, and can be useful in journalistic analysis, media monitoring, or NLP research into framing and stance.
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** Himel Ghosh
- **Language(s) (NLP):** Python
- **Finetuned from model:** roberta-base
- **Code:** https://github.com/Himel1996/NewsBiasDetector/
## Uses
This model is intended to support the detection and analysis of biased language in English news content. It can be used as a tool by:
- **Media researchers** and **social scientists** studying framing, bias, or political discourse.
- **Journalists and editors** aiming to assess the neutrality of their writing or compare outlets.
- **Developers** integrating bias detection into NLP pipelines for content moderation, misinformation detection, or AI-assisted writing tools.
### Foreseeable Uses:
- Annotating datasets for bias.
- Measuring bias across different news outlets or topics.
- Serving as an assistive tool in editorial decision-making or media monitoring.
### Direct Use
This model can be used directly for binary classification of English-language news statements to determine whether they exhibit biased language.
It returns one of two labels:
- **LABEL_0** :Non-biased
- **LABEL_1** : Biased
## Bias, Risks, and Limitations
While this model is designed to detect linguistic bias, it carries several limitations and risks, both technical and sociotechnical:
- The model was fine-tuned on the BABE dataset, which includes annotations based on human judgments that may reflect specific cultural or political perspectives.
- It may not generalize well to non-news text or out-of-domain content (e.g., social media, informal writing).
- Subtle forms of bias, sarcasm, irony, or coded language may not be reliably detected.
- Bias is inherently subjective: What one annotator considers biased may be seen as neutral by another. The model reflects those subjective judgments.
- The model does not detect factual correctness or misinformation — only linguistic bias cues.
- Labeling a text as “biased” may have reputational or ethical implications, especially if used in moderation, censorship, or journalistic evaluations.
## Training Details
### Training Data
Training was done on the BABE Dataset: https://huggingface.co/datasets/mediabiasgroup/BABE
#### Summary
The model achieved 92.02% Accuracy, with very high Precision of 96.15% and 89.27% Recall.
This means the model predicts very few false positives and detects the biases that are actually biases. |