Fix model configuration for proper Hugging Face compatibility
Browse files- README.md +16 -3
- config.json +32 -0
README.md
CHANGED
|
@@ -6,6 +6,7 @@ tags:
|
|
| 6 |
- transformers
|
| 7 |
- pytorch
|
| 8 |
- multilingual
|
|
|
|
| 9 |
license: mit
|
| 10 |
---
|
| 11 |
|
|
@@ -15,14 +16,18 @@ Multilingual text classification model trained on XLM-RoBERTa base for sentiment
|
|
| 15 |
|
| 16 |
## Model Description
|
| 17 |
|
| 18 |
-
This is a multilingual text classification model based on XLM-RoBERTa. It has been
|
| 19 |
|
| 20 |
## Model Details
|
| 21 |
|
| 22 |
- **Base Model**: XLM-RoBERTa Base
|
| 23 |
-
- **
|
|
|
|
| 24 |
- **Languages**: Multilingual (English, Russian, Tajik, and others)
|
| 25 |
- **Max Sequence Length**: 512 tokens
|
|
|
|
|
|
|
|
|
|
| 26 |
|
| 27 |
## Performance
|
| 28 |
|
|
@@ -75,6 +80,14 @@ This model was trained using:
|
|
| 75 |
- **Training Epochs**: 2
|
| 76 |
- **Languages**: English, Russian, Tajik
|
| 77 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 78 |
## Limitations
|
| 79 |
|
| 80 |
- The model's performance may vary across different languages
|
|
@@ -89,7 +102,7 @@ If you use this model in your research, please cite:
|
|
| 89 |
```bibtex
|
| 90 |
@misc{multilingual-text-classifier,
|
| 91 |
title={Multilingual Text Classification Model},
|
| 92 |
-
author={
|
| 93 |
year={2024},
|
| 94 |
publisher={Hugging Face},
|
| 95 |
journal={Hugging Face Hub},
|
|
|
|
| 6 |
- transformers
|
| 7 |
- pytorch
|
| 8 |
- multilingual
|
| 9 |
+
- xlm-roberta
|
| 10 |
license: mit
|
| 11 |
---
|
| 12 |
|
|
|
|
| 16 |
|
| 17 |
## Model Description
|
| 18 |
|
| 19 |
+
This is a multilingual text classification model based on XLM-RoBERTa. It has been fine-tuned for sentiment analysis across multiple languages and can classify text into positive, negative, and neutral categories.
|
| 20 |
|
| 21 |
## Model Details
|
| 22 |
|
| 23 |
- **Base Model**: XLM-RoBERTa Base
|
| 24 |
+
- **Model Type**: XLMRobertaForSequenceClassification
|
| 25 |
+
- **Number of Labels**: 3 (Negative, Neutral, Positive)
|
| 26 |
- **Languages**: Multilingual (English, Russian, Tajik, and others)
|
| 27 |
- **Max Sequence Length**: 512 tokens
|
| 28 |
+
- **Hidden Size**: 768
|
| 29 |
+
- **Attention Heads**: 12
|
| 30 |
+
- **Layers**: 12
|
| 31 |
|
| 32 |
## Performance
|
| 33 |
|
|
|
|
| 80 |
- **Training Epochs**: 2
|
| 81 |
- **Languages**: English, Russian, Tajik
|
| 82 |
|
| 83 |
+
## Model Architecture
|
| 84 |
+
|
| 85 |
+
The model uses the XLM-RoBERTa architecture with:
|
| 86 |
+
- 12 transformer layers
|
| 87 |
+
- 768 hidden dimensions
|
| 88 |
+
- 12 attention heads
|
| 89 |
+
- 3 classification heads for sentiment analysis
|
| 90 |
+
|
| 91 |
## Limitations
|
| 92 |
|
| 93 |
- The model's performance may vary across different languages
|
|
|
|
| 102 |
```bibtex
|
| 103 |
@misc{multilingual-text-classifier,
|
| 104 |
title={Multilingual Text Classification Model},
|
| 105 |
+
author={Advexon},
|
| 106 |
year={2024},
|
| 107 |
publisher={Hugging Face},
|
| 108 |
journal={Hugging Face Hub},
|
config.json
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_type": "xlm-roberta",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"XLMRobertaForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"bos_token_id": 0,
|
| 8 |
+
"eos_token_id": 2,
|
| 9 |
+
"hidden_act": "gelu",
|
| 10 |
+
"hidden_dropout_prob": 0.1,
|
| 11 |
+
"hidden_size": 768,
|
| 12 |
+
"initializer_range": 0.02,
|
| 13 |
+
"intermediate_size": 3072,
|
| 14 |
+
"layer_norm_eps": 1e-12,
|
| 15 |
+
"max_position_embeddings": 514,
|
| 16 |
+
"num_attention_heads": 12,
|
| 17 |
+
"num_hidden_layers": 12,
|
| 18 |
+
"pad_token_id": 1,
|
| 19 |
+
"type_vocab_size": 1,
|
| 20 |
+
"vocab_size": 250002,
|
| 21 |
+
"num_labels": 3,
|
| 22 |
+
"id2label": {
|
| 23 |
+
"0": "negative",
|
| 24 |
+
"1": "neutral",
|
| 25 |
+
"2": "positive"
|
| 26 |
+
},
|
| 27 |
+
"label2id": {
|
| 28 |
+
"negative": 0,
|
| 29 |
+
"neutral": 1,
|
| 30 |
+
"positive": 2
|
| 31 |
+
}
|
| 32 |
+
}
|