DeepDream2045 commited on
Commit
be6bc2b
·
verified ·
1 Parent(s): d931199

Training in progress, step 25, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: NousResearch/CodeLlama-7b-hf
3
  library_name: peft
4
  ---
5
 
@@ -199,4 +199,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
- - PEFT 0.13.2
 
1
  ---
2
+ base_model: unsloth/Qwen2.5-Coder-7B
3
  library_name: peft
4
  ---
5
 
 
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
+ - PEFT 0.14.0
last-checkpoint/adapter_config.json CHANGED
@@ -1,8 +1,10 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "NousResearch/CodeLlama-7b-hf",
5
  "bias": "none",
 
 
6
  "fan_in_fan_out": null,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
@@ -11,6 +13,7 @@
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 64,
 
14
  "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
@@ -20,13 +23,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "k_proj",
24
- "o_proj",
25
  "up_proj",
26
- "q_proj",
27
- "v_proj",
28
  "down_proj",
29
- "gate_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/Qwen2.5-Coder-7B",
5
  "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
  "fan_in_fan_out": null,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
 
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
  "lora_alpha": 64,
16
+ "lora_bias": false,
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "v_proj",
27
  "k_proj",
28
+ "gate_proj",
29
  "up_proj",
 
 
30
  "down_proj",
31
+ "o_proj",
32
+ "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a16782ad6003cc249729eaf787b2f12c4580a66df7a4ad911042842ad8b37f3
3
- size 319876032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c8d32944e4454be4bdddf2e55f719383daf3c04aa97315c721bcb3c26a5548b
3
+ size 323014168
last-checkpoint/added_tokens.json CHANGED
@@ -1,13 +1,25 @@
1
  {
2
- "<|assistant|>": 32001,
3
- "<|endoftext|>": 32000,
4
- "<|end|>": 32007,
5
- "<|placeholder1|>": 32002,
6
- "<|placeholder2|>": 32003,
7
- "<|placeholder3|>": 32004,
8
- "<|placeholder4|>": 32005,
9
- "<|placeholder5|>": 32008,
10
- "<|placeholder6|>": 32009,
11
- "<|system|>": 32006,
12
- "<|user|>": 32010
 
 
 
 
 
 
 
 
 
 
 
 
13
  }
 
1
  {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|PAD_TOKEN|>": 151665,
5
+ "<|box_end|>": 151649,
6
+ "<|box_start|>": 151648,
7
+ "<|endoftext|>": 151643,
8
+ "<|file_sep|>": 151664,
9
+ "<|fim_middle|>": 151660,
10
+ "<|fim_pad|>": 151662,
11
+ "<|fim_prefix|>": 151659,
12
+ "<|fim_suffix|>": 151661,
13
+ "<|im_end|>": 151645,
14
+ "<|im_start|>": 151644,
15
+ "<|image_pad|>": 151655,
16
+ "<|object_ref_end|>": 151647,
17
+ "<|object_ref_start|>": 151646,
18
+ "<|quad_end|>": 151651,
19
+ "<|quad_start|>": 151650,
20
+ "<|repo_name|>": 151663,
21
+ "<|video_pad|>": 151656,
22
+ "<|vision_end|>": 151653,
23
+ "<|vision_pad|>": 151654,
24
+ "<|vision_start|>": 151652
25
  }
last-checkpoint/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:07b7bb7ee1bd78fc86f3b3d334c0f723efc54a0e89cee8a3e6addb29cdc51e70
3
- size 640009682
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50fac20b8af7089797be365f66d5653744e9733a3a115f1d45df97adb16cbd02
3
+ size 646253418
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa6f173b1b026c367b70c1fce82f123265c6bc34d13a332bc2c892e213fccbb7
3
- size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e3cfaa35ecb32c58a62906e3cd360d45beac47233854c9b1e0fc070e573e580
3
+ size 14512
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc3583ab6d26a0a402902114950375996567c99dccbbc5e283def019711614f9
3
- size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5d333ab5eb402c3ad15d011ab1cda0069a64cfa2aa10d382897e584855eb9e0
3
+ size 14512
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1df0528620c07325b8faa7567e59b0c1e86a1f1ee6af1245a69c6c0463fe4e2
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f37b2aa490ccb1598b01e14cda36e9081f7ce646deab4d3c2d03de0d2169a755
3
  size 1064
last-checkpoint/special_tokens_map.json CHANGED
@@ -1,27 +1,28 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
 
 
 
 
 
 
 
 
9
  "eos_token": {
10
- "content": "</s>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "unk_token": {
24
- "content": "<unk>",
25
  "lstrip": false,
26
  "normalized": false,
27
  "rstrip": false,
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
  "eos_token": {
18
+ "content": "<|endoftext|>",
19
  "lstrip": false,
20
  "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
  "pad_token": {
25
+ "content": "<|PAD_TOKEN|>",
 
 
 
 
 
 
 
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
last-checkpoint/tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/tokenizer_config.json CHANGED
@@ -1,26 +1,185 @@
1
  {
2
- "add_bos_token": true,
3
- "add_eos_token": false,
4
- "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
- "0": {
7
- "content": "<unk>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
11
  "single_word": false,
12
  "special": true
13
  },
14
- "1": {
15
- "content": "<s>",
16
  "lstrip": false,
17
  "normalized": false,
18
  "rstrip": false,
19
  "single_word": false,
20
  "special": true
21
  },
22
- "2": {
23
- "content": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  "lstrip": false,
25
  "normalized": false,
26
  "rstrip": false,
@@ -28,15 +187,31 @@
28
  "special": true
29
  }
30
  },
31
- "bos_token": "<s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
33
  "clean_up_tokenization_spaces": false,
34
- "eos_token": "</s>",
35
- "legacy": true,
36
- "model_max_length": 1000000000000000019884624838656,
37
- "pad_token": "</s>",
38
- "sp_model_kwargs": {},
39
- "tokenizer_class": "LlamaTokenizer",
40
- "unk_token": "<unk>",
41
- "use_default_system_prompt": false
 
42
  }
 
1
  {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
 
4
  "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
  "lstrip": false,
8
  "normalized": false,
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
12
  },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
  "lstrip": false,
16
  "normalized": false,
17
  "rstrip": false,
18
  "single_word": false,
19
  "special": true
20
  },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<|PAD_TOKEN|>",
183
  "lstrip": false,
184
  "normalized": false,
185
  "rstrip": false,
 
187
  "special": true
188
  }
189
  },
190
+ "additional_special_tokens": [
191
+ "<|im_start|>",
192
+ "<|im_end|>",
193
+ "<|object_ref_start|>",
194
+ "<|object_ref_end|>",
195
+ "<|box_start|>",
196
+ "<|box_end|>",
197
+ "<|quad_start|>",
198
+ "<|quad_end|>",
199
+ "<|vision_start|>",
200
+ "<|vision_end|>",
201
+ "<|vision_pad|>",
202
+ "<|image_pad|>",
203
+ "<|video_pad|>"
204
+ ],
205
+ "bos_token": null,
206
  "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
207
  "clean_up_tokenization_spaces": false,
208
+ "eos_token": "<|endoftext|>",
209
+ "errors": "replace",
210
+ "extra_special_tokens": {},
211
+ "model_max_length": 131072,
212
+ "pad_token": "<|PAD_TOKEN|>",
213
+ "padding_side": "left",
214
+ "split_special_tokens": false,
215
+ "tokenizer_class": "Qwen2Tokenizer",
216
+ "unk_token": null
217
  }
last-checkpoint/trainer_state.json CHANGED
@@ -1,392 +1,209 @@
1
  {
2
- "best_metric": 4.127087231609039e-05,
3
- "best_model_checkpoint": "miner_id_24/checkpoint-50",
4
- "epoch": 1.0496732026143791,
5
  "eval_steps": 25,
6
- "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.02091503267973856,
13
- "grad_norm": 102.72461700439453,
14
  "learning_rate": 5e-05,
15
- "loss": 148.2262,
16
  "step": 1
17
  },
18
  {
19
- "epoch": 0.02091503267973856,
20
- "eval_loss": 9.406306266784668,
21
- "eval_runtime": 26.0996,
22
- "eval_samples_per_second": 3.103,
23
- "eval_steps_per_second": 1.571,
24
  "step": 1
25
  },
26
  {
27
- "epoch": 0.04183006535947712,
28
- "grad_norm": 90.18787384033203,
29
  "learning_rate": 0.0001,
30
- "loss": 145.0047,
31
  "step": 2
32
  },
33
  {
34
- "epoch": 0.06274509803921569,
35
- "grad_norm": 102.96981048583984,
36
  "learning_rate": 9.989294616193017e-05,
37
- "loss": 137.3407,
38
  "step": 3
39
  },
40
  {
41
- "epoch": 0.08366013071895424,
42
- "grad_norm": 145.56475830078125,
43
  "learning_rate": 9.957224306869053e-05,
44
- "loss": 134.6917,
45
  "step": 4
46
  },
47
  {
48
- "epoch": 0.10457516339869281,
49
- "grad_norm": 140.44082641601562,
50
  "learning_rate": 9.903926402016153e-05,
51
- "loss": 118.558,
52
  "step": 5
53
  },
54
  {
55
- "epoch": 0.12549019607843137,
56
- "grad_norm": 135.83241271972656,
57
  "learning_rate": 9.829629131445342e-05,
58
- "loss": 101.6476,
59
  "step": 6
60
  },
61
  {
62
- "epoch": 0.14640522875816994,
63
- "grad_norm": 116.04879760742188,
64
  "learning_rate": 9.73465064747553e-05,
65
- "loss": 77.2923,
66
  "step": 7
67
  },
68
  {
69
- "epoch": 0.16732026143790849,
70
- "grad_norm": 116.68311309814453,
71
  "learning_rate": 9.619397662556435e-05,
72
- "loss": 60.8297,
73
  "step": 8
74
  },
75
  {
76
- "epoch": 0.18823529411764706,
77
- "grad_norm": 101.5295639038086,
78
  "learning_rate": 9.484363707663442e-05,
79
- "loss": 45.67,
80
  "step": 9
81
  },
82
  {
83
- "epoch": 0.20915032679738563,
84
- "grad_norm": 85.99949645996094,
85
  "learning_rate": 9.330127018922194e-05,
86
- "loss": 38.6756,
87
  "step": 10
88
  },
89
  {
90
- "epoch": 0.23006535947712417,
91
- "grad_norm": 64.54862213134766,
92
  "learning_rate": 9.157348061512727e-05,
93
- "loss": 26.7462,
94
  "step": 11
95
  },
96
  {
97
- "epoch": 0.25098039215686274,
98
- "grad_norm": 37.021934509277344,
99
  "learning_rate": 8.966766701456177e-05,
100
- "loss": 11.3077,
101
  "step": 12
102
  },
103
  {
104
- "epoch": 0.2718954248366013,
105
- "grad_norm": 33.68046569824219,
106
  "learning_rate": 8.759199037394887e-05,
107
- "loss": 8.8502,
108
  "step": 13
109
  },
110
  {
111
- "epoch": 0.2928104575163399,
112
- "grad_norm": 69.80364990234375,
113
  "learning_rate": 8.535533905932738e-05,
114
- "loss": 11.204,
115
  "step": 14
116
  },
117
  {
118
- "epoch": 0.3137254901960784,
119
- "grad_norm": 64.46556091308594,
120
  "learning_rate": 8.296729075500344e-05,
121
- "loss": 8.0732,
122
  "step": 15
123
  },
124
  {
125
- "epoch": 0.33464052287581697,
126
- "grad_norm": 22.950719833374023,
127
  "learning_rate": 8.043807145043604e-05,
128
- "loss": 2.3239,
129
  "step": 16
130
  },
131
  {
132
- "epoch": 0.35555555555555557,
133
- "grad_norm": 27.570941925048828,
134
  "learning_rate": 7.777851165098012e-05,
135
- "loss": 3.8739,
136
  "step": 17
137
  },
138
  {
139
- "epoch": 0.3764705882352941,
140
- "grad_norm": 52.724632263183594,
141
  "learning_rate": 7.500000000000001e-05,
142
- "loss": 2.7403,
143
  "step": 18
144
  },
145
  {
146
- "epoch": 0.39738562091503266,
147
- "grad_norm": 33.629058837890625,
148
  "learning_rate": 7.211443451095007e-05,
149
- "loss": 1.5774,
150
  "step": 19
151
  },
152
  {
153
- "epoch": 0.41830065359477125,
154
- "grad_norm": 10.682621002197266,
155
  "learning_rate": 6.91341716182545e-05,
156
- "loss": 0.6792,
157
  "step": 20
158
  },
159
  {
160
- "epoch": 0.4392156862745098,
161
- "grad_norm": 6.780605792999268,
162
  "learning_rate": 6.607197326515808e-05,
163
- "loss": 0.2117,
164
  "step": 21
165
  },
166
  {
167
- "epoch": 0.46013071895424834,
168
- "grad_norm": 22.594688415527344,
169
  "learning_rate": 6.294095225512603e-05,
170
- "loss": 0.998,
171
  "step": 22
172
  },
173
  {
174
- "epoch": 0.48104575163398694,
175
- "grad_norm": 6.626839637756348,
176
  "learning_rate": 5.9754516100806423e-05,
177
- "loss": 0.3769,
178
  "step": 23
179
  },
180
  {
181
- "epoch": 0.5019607843137255,
182
- "grad_norm": 1.138240933418274,
183
  "learning_rate": 5.6526309611002594e-05,
184
- "loss": 0.0633,
185
  "step": 24
186
  },
187
  {
188
- "epoch": 0.5228758169934641,
189
- "grad_norm": 1.8117948770523071,
190
  "learning_rate": 5.327015646150716e-05,
191
- "loss": 0.0665,
192
  "step": 25
193
  },
194
  {
195
- "epoch": 0.5228758169934641,
196
- "eval_loss": 0.00532330758869648,
197
- "eval_runtime": 26.0795,
198
- "eval_samples_per_second": 3.106,
199
- "eval_steps_per_second": 1.572,
200
  "step": 25
201
- },
202
- {
203
- "epoch": 0.5437908496732026,
204
- "grad_norm": 8.163567543029785,
205
- "learning_rate": 5e-05,
206
- "loss": 0.172,
207
- "step": 26
208
- },
209
- {
210
- "epoch": 0.5647058823529412,
211
- "grad_norm": 2.176654577255249,
212
- "learning_rate": 4.6729843538492847e-05,
213
- "loss": 0.0438,
214
- "step": 27
215
- },
216
- {
217
- "epoch": 0.5856209150326798,
218
- "grad_norm": 0.09583692997694016,
219
- "learning_rate": 4.347369038899744e-05,
220
- "loss": 0.0033,
221
- "step": 28
222
- },
223
- {
224
- "epoch": 0.6065359477124183,
225
- "grad_norm": 2.4841692447662354,
226
- "learning_rate": 4.0245483899193595e-05,
227
- "loss": 0.0293,
228
- "step": 29
229
- },
230
- {
231
- "epoch": 0.6274509803921569,
232
- "grad_norm": 0.9432563185691833,
233
- "learning_rate": 3.705904774487396e-05,
234
- "loss": 0.0176,
235
- "step": 30
236
- },
237
- {
238
- "epoch": 0.6483660130718955,
239
- "grad_norm": 0.29981163144111633,
240
- "learning_rate": 3.392802673484193e-05,
241
- "loss": 0.0095,
242
- "step": 31
243
- },
244
- {
245
- "epoch": 0.6692810457516339,
246
- "grad_norm": 0.13397172093391418,
247
- "learning_rate": 3.086582838174551e-05,
248
- "loss": 0.0034,
249
- "step": 32
250
- },
251
- {
252
- "epoch": 0.6901960784313725,
253
- "grad_norm": 0.26384004950523376,
254
- "learning_rate": 2.7885565489049946e-05,
255
- "loss": 0.0036,
256
- "step": 33
257
- },
258
- {
259
- "epoch": 0.7111111111111111,
260
- "grad_norm": 0.017268378287553787,
261
- "learning_rate": 2.500000000000001e-05,
262
- "loss": 0.0007,
263
- "step": 34
264
- },
265
- {
266
- "epoch": 0.7320261437908496,
267
- "grad_norm": 0.0771813616156578,
268
- "learning_rate": 2.2221488349019903e-05,
269
- "loss": 0.0022,
270
- "step": 35
271
- },
272
- {
273
- "epoch": 0.7529411764705882,
274
- "grad_norm": 0.06307241320610046,
275
- "learning_rate": 1.9561928549563968e-05,
276
- "loss": 0.0011,
277
- "step": 36
278
- },
279
- {
280
- "epoch": 0.7738562091503268,
281
- "grad_norm": 0.009056968614459038,
282
- "learning_rate": 1.703270924499656e-05,
283
- "loss": 0.0003,
284
- "step": 37
285
- },
286
- {
287
- "epoch": 0.7947712418300653,
288
- "grad_norm": 0.02682294324040413,
289
- "learning_rate": 1.4644660940672627e-05,
290
- "loss": 0.0008,
291
- "step": 38
292
- },
293
- {
294
- "epoch": 0.8156862745098039,
295
- "grad_norm": 0.03674069046974182,
296
- "learning_rate": 1.2408009626051137e-05,
297
- "loss": 0.0013,
298
- "step": 39
299
- },
300
- {
301
- "epoch": 0.8366013071895425,
302
- "grad_norm": 0.011068358086049557,
303
- "learning_rate": 1.0332332985438248e-05,
304
- "loss": 0.0005,
305
- "step": 40
306
- },
307
- {
308
- "epoch": 0.857516339869281,
309
- "grad_norm": 0.013999490067362785,
310
- "learning_rate": 8.426519384872733e-06,
311
- "loss": 0.0007,
312
- "step": 41
313
- },
314
- {
315
- "epoch": 0.8784313725490196,
316
- "grad_norm": 0.011875639669597149,
317
- "learning_rate": 6.698729810778065e-06,
318
- "loss": 0.0006,
319
- "step": 42
320
- },
321
- {
322
- "epoch": 0.8993464052287582,
323
- "grad_norm": 0.12404092401266098,
324
- "learning_rate": 5.156362923365588e-06,
325
- "loss": 0.0015,
326
- "step": 43
327
- },
328
- {
329
- "epoch": 0.9202614379084967,
330
- "grad_norm": 0.03556958585977554,
331
- "learning_rate": 3.8060233744356633e-06,
332
- "loss": 0.001,
333
- "step": 44
334
- },
335
- {
336
- "epoch": 0.9411764705882353,
337
- "grad_norm": 6.352353096008301,
338
- "learning_rate": 2.653493525244721e-06,
339
- "loss": 0.0426,
340
- "step": 45
341
- },
342
- {
343
- "epoch": 0.9620915032679739,
344
- "grad_norm": 0.0167868509888649,
345
- "learning_rate": 1.70370868554659e-06,
346
- "loss": 0.0007,
347
- "step": 46
348
- },
349
- {
350
- "epoch": 0.9830065359477124,
351
- "grad_norm": 0.024576369673013687,
352
- "learning_rate": 9.607359798384785e-07,
353
- "loss": 0.0005,
354
- "step": 47
355
- },
356
- {
357
- "epoch": 1.007843137254902,
358
- "grad_norm": 0.009638285264372826,
359
- "learning_rate": 4.277569313094809e-07,
360
- "loss": 0.0005,
361
- "step": 48
362
- },
363
- {
364
- "epoch": 1.0287581699346404,
365
- "grad_norm": 0.007980446331202984,
366
- "learning_rate": 1.0705383806982606e-07,
367
- "loss": 0.0004,
368
- "step": 49
369
- },
370
- {
371
- "epoch": 1.0496732026143791,
372
- "grad_norm": 0.006425910629332066,
373
- "learning_rate": 0.0,
374
- "loss": 0.0003,
375
- "step": 50
376
- },
377
- {
378
- "epoch": 1.0496732026143791,
379
- "eval_loss": 4.127087231609039e-05,
380
- "eval_runtime": 26.0704,
381
- "eval_samples_per_second": 3.107,
382
- "eval_steps_per_second": 1.573,
383
- "step": 50
384
  }
385
  ],
386
  "logging_steps": 1,
387
  "max_steps": 50,
388
  "num_input_tokens_seen": 0,
389
- "num_train_epochs": 2,
390
  "save_steps": 25,
391
  "stateful_callbacks": {
392
  "EarlyStoppingCallback": {
@@ -404,12 +221,12 @@
404
  "should_evaluate": false,
405
  "should_log": false,
406
  "should_save": true,
407
- "should_training_stop": true
408
  },
409
  "attributes": {}
410
  }
411
  },
412
- "total_flos": 1.314789078859776e+17,
413
  "train_batch_size": 2,
414
  "trial_name": null,
415
  "trial_params": null
 
1
  {
2
+ "best_metric": 1.957382082939148,
3
+ "best_model_checkpoint": "miner_id_24/checkpoint-25",
4
+ "epoch": 0.016421709499958945,
5
  "eval_steps": 25,
6
+ "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.0006568683799983579,
13
+ "grad_norm": 21.116296768188477,
14
  "learning_rate": 5e-05,
15
+ "loss": 47.7589,
16
  "step": 1
17
  },
18
  {
19
+ "epoch": 0.0006568683799983579,
20
+ "eval_loss": 4.154504299163818,
21
+ "eval_runtime": 573.0772,
22
+ "eval_samples_per_second": 8.948,
23
+ "eval_steps_per_second": 2.237,
24
  "step": 1
25
  },
26
  {
27
+ "epoch": 0.0013137367599967158,
28
+ "grad_norm": 18.675199508666992,
29
  "learning_rate": 0.0001,
30
+ "loss": 53.2593,
31
  "step": 2
32
  },
33
  {
34
+ "epoch": 0.0019706051399950735,
35
+ "grad_norm": 20.739986419677734,
36
  "learning_rate": 9.989294616193017e-05,
37
+ "loss": 51.8736,
38
  "step": 3
39
  },
40
  {
41
+ "epoch": 0.0026274735199934315,
42
+ "grad_norm": 21.701404571533203,
43
  "learning_rate": 9.957224306869053e-05,
44
+ "loss": 56.3313,
45
  "step": 4
46
  },
47
  {
48
+ "epoch": 0.003284341899991789,
49
+ "grad_norm": 21.279020309448242,
50
  "learning_rate": 9.903926402016153e-05,
51
+ "loss": 50.3041,
52
  "step": 5
53
  },
54
  {
55
+ "epoch": 0.003941210279990147,
56
+ "grad_norm": 19.682722091674805,
57
  "learning_rate": 9.829629131445342e-05,
58
+ "loss": 46.2417,
59
  "step": 6
60
  },
61
  {
62
+ "epoch": 0.004598078659988505,
63
+ "grad_norm": 21.489429473876953,
64
  "learning_rate": 9.73465064747553e-05,
65
+ "loss": 44.0281,
66
  "step": 7
67
  },
68
  {
69
+ "epoch": 0.005254947039986863,
70
+ "grad_norm": 23.246824264526367,
71
  "learning_rate": 9.619397662556435e-05,
72
+ "loss": 44.2746,
73
  "step": 8
74
  },
75
  {
76
+ "epoch": 0.00591181541998522,
77
+ "grad_norm": 25.966197967529297,
78
  "learning_rate": 9.484363707663442e-05,
79
+ "loss": 42.628,
80
  "step": 9
81
  },
82
  {
83
+ "epoch": 0.006568683799983578,
84
+ "grad_norm": 29.68147850036621,
85
  "learning_rate": 9.330127018922194e-05,
86
+ "loss": 39.1083,
87
  "step": 10
88
  },
89
  {
90
+ "epoch": 0.007225552179981936,
91
+ "grad_norm": 29.226259231567383,
92
  "learning_rate": 9.157348061512727e-05,
93
+ "loss": 37.2143,
94
  "step": 11
95
  },
96
  {
97
+ "epoch": 0.007882420559980294,
98
+ "grad_norm": 29.88581657409668,
99
  "learning_rate": 8.966766701456177e-05,
100
+ "loss": 37.5903,
101
  "step": 12
102
  },
103
  {
104
+ "epoch": 0.008539288939978651,
105
+ "grad_norm": 24.698503494262695,
106
  "learning_rate": 8.759199037394887e-05,
107
+ "loss": 38.8273,
108
  "step": 13
109
  },
110
  {
111
+ "epoch": 0.00919615731997701,
112
+ "grad_norm": 22.38409423828125,
113
  "learning_rate": 8.535533905932738e-05,
114
+ "loss": 39.2722,
115
  "step": 14
116
  },
117
  {
118
+ "epoch": 0.009853025699975367,
119
+ "grad_norm": 17.25520133972168,
120
  "learning_rate": 8.296729075500344e-05,
121
+ "loss": 37.6769,
122
  "step": 15
123
  },
124
  {
125
+ "epoch": 0.010509894079973726,
126
+ "grad_norm": 15.956023216247559,
127
  "learning_rate": 8.043807145043604e-05,
128
+ "loss": 34.8531,
129
  "step": 16
130
  },
131
  {
132
+ "epoch": 0.011166762459972083,
133
+ "grad_norm": 18.063295364379883,
134
  "learning_rate": 7.777851165098012e-05,
135
+ "loss": 36.7135,
136
  "step": 17
137
  },
138
  {
139
+ "epoch": 0.01182363083997044,
140
+ "grad_norm": 18.19580841064453,
141
  "learning_rate": 7.500000000000001e-05,
142
+ "loss": 36.0943,
143
  "step": 18
144
  },
145
  {
146
+ "epoch": 0.0124804992199688,
147
+ "grad_norm": 19.19350242614746,
148
  "learning_rate": 7.211443451095007e-05,
149
+ "loss": 34.9283,
150
  "step": 19
151
  },
152
  {
153
+ "epoch": 0.013137367599967156,
154
+ "grad_norm": 15.737515449523926,
155
  "learning_rate": 6.91341716182545e-05,
156
+ "loss": 35.9358,
157
  "step": 20
158
  },
159
  {
160
+ "epoch": 0.013794235979965515,
161
+ "grad_norm": 17.68427276611328,
162
  "learning_rate": 6.607197326515808e-05,
163
+ "loss": 35.4988,
164
  "step": 21
165
  },
166
  {
167
+ "epoch": 0.014451104359963872,
168
+ "grad_norm": 15.05672550201416,
169
  "learning_rate": 6.294095225512603e-05,
170
+ "loss": 34.0991,
171
  "step": 22
172
  },
173
  {
174
+ "epoch": 0.01510797273996223,
175
+ "grad_norm": 15.203287124633789,
176
  "learning_rate": 5.9754516100806423e-05,
177
+ "loss": 35.6672,
178
  "step": 23
179
  },
180
  {
181
+ "epoch": 0.015764841119960588,
182
+ "grad_norm": 15.746703147888184,
183
  "learning_rate": 5.6526309611002594e-05,
184
+ "loss": 35.0715,
185
  "step": 24
186
  },
187
  {
188
+ "epoch": 0.016421709499958945,
189
+ "grad_norm": 16.753942489624023,
190
  "learning_rate": 5.327015646150716e-05,
191
+ "loss": 35.2284,
192
  "step": 25
193
  },
194
  {
195
+ "epoch": 0.016421709499958945,
196
+ "eval_loss": 1.957382082939148,
197
+ "eval_runtime": 575.1539,
198
+ "eval_samples_per_second": 8.916,
199
+ "eval_steps_per_second": 2.229,
200
  "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
  }
202
  ],
203
  "logging_steps": 1,
204
  "max_steps": 50,
205
  "num_input_tokens_seen": 0,
206
+ "num_train_epochs": 1,
207
  "save_steps": 25,
208
  "stateful_callbacks": {
209
  "EarlyStoppingCallback": {
 
221
  "should_evaluate": false,
222
  "should_log": false,
223
  "should_save": true,
224
+ "should_training_stop": false
225
  },
226
  "attributes": {}
227
  }
228
  },
229
+ "total_flos": 1.406014452662272e+17,
230
  "train_batch_size": 2,
231
  "trial_name": null,
232
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10fa75858312944c66542aee6d6a3e31ea121a457c0727d1a86e3c74f020b648
3
- size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bac41a9514476834778779d021c3d120cffc738ddcbe07213824d214167aadd3
3
+ size 6904
last-checkpoint/vocab.json ADDED
The diff for this file is too large to render. See raw diff