Datasets:
Tasks:
Text Classification
Modalities:
Text
Formats:
csv
Sub-tasks:
multi-class-classification
Languages:
English
Size:
< 1K
License:
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,63 @@
|
|
| 1 |
---
|
| 2 |
license: cc-by-nc-sa-4.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: cc-by-nc-sa-4.0
|
| 3 |
+
tags:
|
| 4 |
+
- text-classification
|
| 5 |
+
- bias-detection
|
| 6 |
+
- media
|
| 7 |
+
- news
|
| 8 |
+
task_categories:
|
| 9 |
+
- text-classification
|
| 10 |
+
task_ids:
|
| 11 |
+
- multi-class-classification
|
| 12 |
+
language:
|
| 13 |
+
- en
|
| 14 |
+
pretty_name: NEWSUNFOLD
|
| 15 |
+
size_categories:
|
| 16 |
+
- n<1K
|
| 17 |
---
|
| 18 |
+
|
| 19 |
+
# Dataset Card for NEWSUNFOLD
|
| 20 |
+
|
| 21 |
+
## Dataset Summary
|
| 22 |
+
**NEWSUNFOLD** contains sentence-level labels of linguistic media bias collected for the paper
|
| 23 |
+
*“NewsUnfold: Creating a News-Reading Application That Indicates Linguistic Media Bias and Collects Feedback”* (https://doi.org/10.1609/icwsm.v19i1.35847).
|
| 24 |
+
|
| 25 |
+
The dataset has **310** English news sentences with binary bias labels derived from user feedback on the NewsUnfold web app. Each record includes the sentence text, a bias label, an identifier, the source outlet, the original article URL, and a topic tag.
|
| 26 |
+
|
| 27 |
+
- **Collection window:** March 4-11, 2023
|
| 28 |
+
- **Label meaning:** `1` = biased, `0` = not biased
|
| 29 |
+
- **Selection rule:** Only sentences with a clear decision (≥5 votes) are included.
|
| 30 |
+
|
| 31 |
+
GitHub: https://github.com/Media-Bias-Group/NewsUnfold/
|
| 32 |
+
|
| 33 |
+
## Use Cases
|
| 34 |
+
- Training/evaluating bias-detection models
|
| 35 |
+
- Auditing or probing linguistic bias signals in news sentences
|
| 36 |
+
- Complementing existing datasets (e.g., BABE) with user-feedback–grounded labels
|
| 37 |
+
|
| 38 |
+
## Dataset Structure
|
| 39 |
+
|
| 40 |
+
### Data Fields
|
| 41 |
+
- `id` *(string)*: Unique sentence identifier.
|
| 42 |
+
- `text` *(string)*: The news sentence.
|
| 43 |
+
- `label` *(int; 0 or 1)*: Bias label (`1` = biased, `0` = not biased).
|
| 44 |
+
- `outlet` *(string)*: News outlet/source.
|
| 45 |
+
- `news_link` *(string/url)*: Link to the original article.
|
| 46 |
+
- `topic` *(string)*: Topic/category of the article/sentence.
|
| 47 |
+
|
| 48 |
+
## Citation
|
| 49 |
+
```
|
| 50 |
+
@article{Hinterreiter_Wessel_Schliski_Echizen_Latoschik_Spinde_2025,
|
| 51 |
+
title = {NewsUnfold: Creating a News-Reading Application That Indicates Linguistic Media Bias and Collects Feedback},
|
| 52 |
+
volume = {19},
|
| 53 |
+
url = {https://doi.org/10.1609/icwsm.v19i1.35847},
|
| 54 |
+
number = {1},
|
| 55 |
+
journal = {Proceedings of the International AAAI Conference on Web and Social Media},
|
| 56 |
+
author = {Hinterreiter, Maximilian and Wessel, Daniel and Schliski, Johanna and Echizen, Isao and Latoschik, Marc Erich and Spinde, Timo},
|
| 57 |
+
year = {2025},
|
| 58 |
+
doi = {10.1609/icwsm.v19i1.35847},
|
| 59 |
+
pages = {1237-1248},
|
| 60 |
+
abstract = {We present NewsUnfold, a news-reading web application that (1) indicates sentence-level linguistic media bias predictions to readers, (2) collects reader feedback on those predictions, and (3) releases a new dataset of sentences with community labels indicating perceived bias. In an in-the-wild study, participants used the app to read curated articles while seeing sentences highlighted as biased or not biased (per a baseline classifier). They could then agree or disagree with these highlights, producing crowd-sourced labels. Using only sentences with a clear decision (≥5 votes), we provide a 310-sentence dataset and report inter-annotator agreement (Krippendorff’s α = .504). We further train a classifier combining the BABE dataset with our new sentences, improving F1 to .824 (+2.49% vs. BABE alone), demonstrating that community feedback can refine bias-detection models and support transparent, user-in-the-loop media analysis.},
|
| 61 |
+
keywords = {media bias, bias detection, user feedback, dataset, news, sentence-level labeling}
|
| 62 |
+
}
|
| 63 |
+
```
|