mamung commited on
Commit
8662a16
·
verified ·
1 Parent(s): e6e3956

Training in progress, step 8, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:606e3fb61641782e892a9608a299d16746f85867c3542c27511b23f93bedfd1f
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5bbe16e6dfcc4d1beaccc8733c052e51cbd7b501affc325bf0a82113f81bcff
3
  size 167832240
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8cbd6288f0543eb4076c501009a036aca4a86980e8c6a2cbbe6086be1748d60
3
  size 85723284
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71108f33983ec8897e395094938fa7b51e50487fc5a9efc5f33b34fbf18b58ed
3
  size 85723284
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6607a2634317352a1c5d3805f188f5d499b7a4803740f0c3b3b493e984216ea3
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a82b9f2cb69cffb5c91e61597e033eef893451c68faed84d6275b923928391fd
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92b4cf85d7ba7a497f88ff799bf4dec5af7dd95be6e00f78bf46ba5deb56bbf8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a3d18e607a27cfab9cec7cfaa7384cdb877a2330c4bcd4e1efcae25be9908cb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.20168067226890757,
5
  "eval_steps": 1,
6
- "global_step": 6,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -97,6 +97,36 @@
97
  "eval_samples_per_second": 11.624,
98
  "eval_steps_per_second": 6.259,
99
  "step": 6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
100
  }
101
  ],
102
  "logging_steps": 1,
@@ -116,7 +146,7 @@
116
  "attributes": {}
117
  }
118
  },
119
- "total_flos": 2225661850681344.0,
120
  "train_batch_size": 2,
121
  "trial_name": null,
122
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.2689075630252101,
5
  "eval_steps": 1,
6
+ "global_step": 8,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
97
  "eval_samples_per_second": 11.624,
98
  "eval_steps_per_second": 6.259,
99
  "step": 6
100
+ },
101
+ {
102
+ "epoch": 0.23529411764705882,
103
+ "grad_norm": 5.040204048156738,
104
+ "learning_rate": 7e-05,
105
+ "loss": 0.4876,
106
+ "step": 7
107
+ },
108
+ {
109
+ "epoch": 0.23529411764705882,
110
+ "eval_loss": 0.17421294748783112,
111
+ "eval_runtime": 1.1209,
112
+ "eval_samples_per_second": 11.598,
113
+ "eval_steps_per_second": 6.245,
114
+ "step": 7
115
+ },
116
+ {
117
+ "epoch": 0.2689075630252101,
118
+ "grad_norm": 2.6843221187591553,
119
+ "learning_rate": 8e-05,
120
+ "loss": 0.183,
121
+ "step": 8
122
+ },
123
+ {
124
+ "epoch": 0.2689075630252101,
125
+ "eval_loss": 0.059619225561618805,
126
+ "eval_runtime": 1.1225,
127
+ "eval_samples_per_second": 11.581,
128
+ "eval_steps_per_second": 6.236,
129
+ "step": 8
130
  }
131
  ],
132
  "logging_steps": 1,
 
146
  "attributes": {}
147
  }
148
  },
149
+ "total_flos": 2967549134241792.0,
150
  "train_batch_size": 2,
151
  "trial_name": null,
152
  "trial_params": null