himel7 commited on
Commit
928972b
·
verified ·
1 Parent(s): 115964b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -19
README.md CHANGED
@@ -37,6 +37,22 @@ dataset for bias detection in English news statements.
37
  The model predicts whether a given sentence contains biased language (LABEL_1) or is unbiased (LABEL_0).
38
  It is intended for applications in media bias analysis, content moderation, and social computing research.
39
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ## Model Details
41
 
42
  ### Model Description
@@ -83,16 +99,6 @@ It returns one of two labels:
83
 
84
  - **LABEL_1** : Biased
85
 
86
- - Example usage with Hugging Face’s pipeline:
87
-
88
- -
89
- ```
90
- from transformers import pipeline
91
-
92
- classifier = pipeline("text-classification", model="himel7/bias-detector", tokenizer="roberta-base")
93
- result = classifier("Immigrants are criminals.")
94
- ```
95
-
96
 
97
  ## Bias, Risks, and Limitations
98
 
@@ -111,14 +117,6 @@ While this model is designed to detect linguistic bias, it carries several limit
111
  Training was done on the BABE Dataset: https://huggingface.co/datasets/mediabiasgroup/BABE
112
 
113
 
114
- ## Evaluation
115
-
116
- The model was evaluated on the entire BABE dataset with a K-fold Cross Validation and yielded the following metrics at K=5:
117
- - **Accuracy: 0.9202**
118
- - **Precision: 0.9615**
119
- - **Recall: 0.8927**
120
- - **F1 Score: 0.9257**
121
-
122
  #### Summary
123
- The model achieved 85.2% Accuracy on the BABE test split, with very high Precision of 92.37% and 80.14% Recall.
124
  This means the model predicts very few false positives and detects the biases that are actually biases.
 
37
  The model predicts whether a given sentence contains biased language (LABEL_1) or is unbiased (LABEL_0).
38
  It is intended for applications in media bias analysis, content moderation, and social computing research.
39
 
40
+ - Example usage with Hugging Face’s pipeline:
41
+ ```
42
+ from transformers import pipeline
43
+
44
+ classifier = pipeline("text-classification", model="himel7/bias-detector", tokenizer="roberta-base")
45
+ result = classifier("Immigrants are criminals.")
46
+ ```
47
+
48
+ ## Evaluation
49
+
50
+ The model was evaluated on the entire BABE dataset with a K-fold Cross Validation and yielded the following metrics at K=5:
51
+ - **Accuracy: 0.9202**
52
+ - **Precision: 0.9615**
53
+ - **Recall: 0.8927**
54
+ - **F1 Score: 0.9257**
55
+
56
  ## Model Details
57
 
58
  ### Model Description
 
99
 
100
  - **LABEL_1** : Biased
101
 
 
 
 
 
 
 
 
 
 
 
102
 
103
  ## Bias, Risks, and Limitations
104
 
 
117
  Training was done on the BABE Dataset: https://huggingface.co/datasets/mediabiasgroup/BABE
118
 
119
 
 
 
 
 
 
 
 
 
120
  #### Summary
121
+ The model achieved 92.02% Accuracy, with very high Precision of 96.15% and 89.27% Recall.
122
  This means the model predicts very few false positives and detects the biases that are actually biases.