advexon commited on
Commit
b7acf27
·
verified ·
1 Parent(s): a1a4186

Fix model configuration for proper Hugging Face compatibility

Browse files
Files changed (2) hide show
  1. README.md +16 -3
  2. config.json +32 -0
README.md CHANGED
@@ -6,6 +6,7 @@ tags:
6
  - transformers
7
  - pytorch
8
  - multilingual
 
9
  license: mit
10
  ---
11
 
@@ -15,14 +16,18 @@ Multilingual text classification model trained on XLM-RoBERTa base for sentiment
15
 
16
  ## Model Description
17
 
18
- This is a multilingual text classification model based on XLM-RoBERTa. It has been trained for sentiment analysis across multiple languages and can classify text into positive, negative, and neutral categories.
19
 
20
  ## Model Details
21
 
22
  - **Base Model**: XLM-RoBERTa Base
23
- - **Number of Labels**: 3 (Positive, Negative, Neutral)
 
24
  - **Languages**: Multilingual (English, Russian, Tajik, and others)
25
  - **Max Sequence Length**: 512 tokens
 
 
 
26
 
27
  ## Performance
28
 
@@ -75,6 +80,14 @@ This model was trained using:
75
  - **Training Epochs**: 2
76
  - **Languages**: English, Russian, Tajik
77
 
 
 
 
 
 
 
 
 
78
  ## Limitations
79
 
80
  - The model's performance may vary across different languages
@@ -89,7 +102,7 @@ If you use this model in your research, please cite:
89
  ```bibtex
90
  @misc{multilingual-text-classifier,
91
  title={Multilingual Text Classification Model},
92
- author={Your Name},
93
  year={2024},
94
  publisher={Hugging Face},
95
  journal={Hugging Face Hub},
 
6
  - transformers
7
  - pytorch
8
  - multilingual
9
+ - xlm-roberta
10
  license: mit
11
  ---
12
 
 
16
 
17
  ## Model Description
18
 
19
+ This is a multilingual text classification model based on XLM-RoBERTa. It has been fine-tuned for sentiment analysis across multiple languages and can classify text into positive, negative, and neutral categories.
20
 
21
  ## Model Details
22
 
23
  - **Base Model**: XLM-RoBERTa Base
24
+ - **Model Type**: XLMRobertaForSequenceClassification
25
+ - **Number of Labels**: 3 (Negative, Neutral, Positive)
26
  - **Languages**: Multilingual (English, Russian, Tajik, and others)
27
  - **Max Sequence Length**: 512 tokens
28
+ - **Hidden Size**: 768
29
+ - **Attention Heads**: 12
30
+ - **Layers**: 12
31
 
32
  ## Performance
33
 
 
80
  - **Training Epochs**: 2
81
  - **Languages**: English, Russian, Tajik
82
 
83
+ ## Model Architecture
84
+
85
+ The model uses the XLM-RoBERTa architecture with:
86
+ - 12 transformer layers
87
+ - 768 hidden dimensions
88
+ - 12 attention heads
89
+ - 3 classification heads for sentiment analysis
90
+
91
  ## Limitations
92
 
93
  - The model's performance may vary across different languages
 
102
  ```bibtex
103
  @misc{multilingual-text-classifier,
104
  title={Multilingual Text Classification Model},
105
+ author={Advexon},
106
  year={2024},
107
  publisher={Hugging Face},
108
  journal={Hugging Face Hub},
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "xlm-roberta",
3
+ "architectures": [
4
+ "XLMRobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 514,
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "pad_token_id": 1,
19
+ "type_vocab_size": 1,
20
+ "vocab_size": 250002,
21
+ "num_labels": 3,
22
+ "id2label": {
23
+ "0": "negative",
24
+ "1": "neutral",
25
+ "2": "positive"
26
+ },
27
+ "label2id": {
28
+ "negative": 0,
29
+ "neutral": 1,
30
+ "positive": 2
31
+ }
32
+ }