real-jiakai commited on
Commit
d831723
verified
1 Parent(s): 55343cc

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ eval_nbest_predictions.json filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ checkpoint-*/
2
+ .ipynb_checkpoints
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: google-bert/bert-base-uncased
5
+ tags:
6
+ - generated_from_trainer
7
+ datasets:
8
+ - squad_v2
9
+ model-index:
10
+ - name: debug_squad2
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # debug_squad2
18
+
19
+ This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on the squad_v2 dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 3e-05
39
+ - train_batch_size: 12
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
+ - lr_scheduler_type: linear
44
+ - num_epochs: 5.0
45
+
46
+ ### Training results
47
+
48
+
49
+
50
+ ### Framework versions
51
+
52
+ - Transformers 4.47.0.dev0
53
+ - Pytorch 2.5.1+cu124
54
+ - Datasets 3.1.0
55
+ - Tokenizers 0.20.3
all_results.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 5.0,
3
+ "eval_HasAns_exact": 71.25506072874494,
4
+ "eval_HasAns_f1": 78.777963677423,
5
+ "eval_HasAns_total": 5928,
6
+ "eval_NoAns_exact": 73.42304457527334,
7
+ "eval_NoAns_f1": 73.42304457527334,
8
+ "eval_NoAns_total": 5945,
9
+ "eval_best_exact": 72.34060473342879,
10
+ "eval_best_exact_thresh": 0.0,
11
+ "eval_best_f1": 76.0966704859567,
12
+ "eval_best_f1_thresh": 0.0,
13
+ "eval_exact": 72.34060473342879,
14
+ "eval_f1": 76.09667048595679,
15
+ "eval_runtime": 42.8962,
16
+ "eval_samples": 12134,
17
+ "eval_samples_per_second": 282.869,
18
+ "eval_steps_per_second": 35.364,
19
+ "eval_total": 11873,
20
+ "total_flos": 1.2910084832623104e+17,
21
+ "train_loss": 0.07113273273183131,
22
+ "train_runtime": 1918.8519,
23
+ "train_samples": 131754,
24
+ "train_samples_per_second": 343.315,
25
+ "train_steps_per_second": 28.611
26
+ }
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google-bert/bert-base-uncased",
3
+ "architectures": [
4
+ "BertForQuestionAnswering"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.47.0.dev0",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 30522
26
+ }
eval_nbest_predictions.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a567cccb720de34a692a1a3fdda9920e4801ee8ec75b462895255574714ed7f
3
+ size 56496143
eval_null_odds.json ADDED
The diff for this file is too large to render. See raw diff
 
eval_predictions.json ADDED
The diff for this file is too large to render. See raw diff
 
eval_results.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 5.0,
3
+ "eval_HasAns_exact": 71.25506072874494,
4
+ "eval_HasAns_f1": 78.777963677423,
5
+ "eval_HasAns_total": 5928,
6
+ "eval_NoAns_exact": 73.42304457527334,
7
+ "eval_NoAns_f1": 73.42304457527334,
8
+ "eval_NoAns_total": 5945,
9
+ "eval_best_exact": 72.34060473342879,
10
+ "eval_best_exact_thresh": 0.0,
11
+ "eval_best_f1": 76.0966704859567,
12
+ "eval_best_f1_thresh": 0.0,
13
+ "eval_exact": 72.34060473342879,
14
+ "eval_f1": 76.09667048595679,
15
+ "eval_runtime": 42.8962,
16
+ "eval_samples": 12134,
17
+ "eval_samples_per_second": 282.869,
18
+ "eval_steps_per_second": 35.364,
19
+ "eval_total": 11873
20
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ba97d2a9ba1362b605cd7b79756c0debbcb4434e738847ca1cacb52730b7e43
3
+ size 435596088
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 5.0,
3
+ "total_flos": 1.2910084832623104e+17,
4
+ "train_loss": 0.07113273273183131,
5
+ "train_runtime": 1918.8519,
6
+ "train_samples": 131754,
7
+ "train_samples_per_second": 343.315,
8
+ "train_steps_per_second": 28.611
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,805 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 5.0,
5
+ "eval_steps": 500,
6
+ "global_step": 54900,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.04553734061930783,
13
+ "grad_norm": 18.591150283813477,
14
+ "learning_rate": 2.9726775956284154e-05,
15
+ "loss": 2.5583,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.09107468123861566,
20
+ "grad_norm": 15.338476181030273,
21
+ "learning_rate": 2.9453551912568308e-05,
22
+ "loss": 1.6841,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.1366120218579235,
27
+ "grad_norm": 14.975069999694824,
28
+ "learning_rate": 2.9180327868852458e-05,
29
+ "loss": 1.5423,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 0.18214936247723132,
34
+ "grad_norm": 17.511743545532227,
35
+ "learning_rate": 2.890710382513661e-05,
36
+ "loss": 1.4113,
37
+ "step": 2000
38
+ },
39
+ {
40
+ "epoch": 0.22768670309653916,
41
+ "grad_norm": 18.18260955810547,
42
+ "learning_rate": 2.8633879781420765e-05,
43
+ "loss": 1.3533,
44
+ "step": 2500
45
+ },
46
+ {
47
+ "epoch": 0.273224043715847,
48
+ "grad_norm": 16.508344650268555,
49
+ "learning_rate": 2.836065573770492e-05,
50
+ "loss": 1.299,
51
+ "step": 3000
52
+ },
53
+ {
54
+ "epoch": 0.31876138433515483,
55
+ "grad_norm": 25.282175064086914,
56
+ "learning_rate": 2.8087431693989072e-05,
57
+ "loss": 1.282,
58
+ "step": 3500
59
+ },
60
+ {
61
+ "epoch": 0.36429872495446264,
62
+ "grad_norm": 24.085981369018555,
63
+ "learning_rate": 2.7814207650273226e-05,
64
+ "loss": 1.2146,
65
+ "step": 4000
66
+ },
67
+ {
68
+ "epoch": 0.4098360655737705,
69
+ "grad_norm": 12.455412864685059,
70
+ "learning_rate": 2.754098360655738e-05,
71
+ "loss": 1.2332,
72
+ "step": 4500
73
+ },
74
+ {
75
+ "epoch": 0.4553734061930783,
76
+ "grad_norm": 10.222314834594727,
77
+ "learning_rate": 2.7267759562841533e-05,
78
+ "loss": 1.1607,
79
+ "step": 5000
80
+ },
81
+ {
82
+ "epoch": 0.5009107468123861,
83
+ "grad_norm": 10.9562406539917,
84
+ "learning_rate": 2.6994535519125683e-05,
85
+ "loss": 1.1293,
86
+ "step": 5500
87
+ },
88
+ {
89
+ "epoch": 0.546448087431694,
90
+ "grad_norm": 16.84743309020996,
91
+ "learning_rate": 2.6721311475409837e-05,
92
+ "loss": 1.129,
93
+ "step": 6000
94
+ },
95
+ {
96
+ "epoch": 0.5919854280510018,
97
+ "grad_norm": 21.29459571838379,
98
+ "learning_rate": 2.644808743169399e-05,
99
+ "loss": 1.1002,
100
+ "step": 6500
101
+ },
102
+ {
103
+ "epoch": 0.6375227686703097,
104
+ "grad_norm": 25.021621704101562,
105
+ "learning_rate": 2.6174863387978144e-05,
106
+ "loss": 1.0397,
107
+ "step": 7000
108
+ },
109
+ {
110
+ "epoch": 0.6830601092896175,
111
+ "grad_norm": 11.044733047485352,
112
+ "learning_rate": 2.5901639344262294e-05,
113
+ "loss": 1.0806,
114
+ "step": 7500
115
+ },
116
+ {
117
+ "epoch": 0.7285974499089253,
118
+ "grad_norm": 10.266705513000488,
119
+ "learning_rate": 2.5628415300546447e-05,
120
+ "loss": 1.047,
121
+ "step": 8000
122
+ },
123
+ {
124
+ "epoch": 0.7741347905282332,
125
+ "grad_norm": 13.432859420776367,
126
+ "learning_rate": 2.53551912568306e-05,
127
+ "loss": 1.0566,
128
+ "step": 8500
129
+ },
130
+ {
131
+ "epoch": 0.819672131147541,
132
+ "grad_norm": 25.581703186035156,
133
+ "learning_rate": 2.5081967213114754e-05,
134
+ "loss": 1.0277,
135
+ "step": 9000
136
+ },
137
+ {
138
+ "epoch": 0.8652094717668488,
139
+ "grad_norm": 14.602388381958008,
140
+ "learning_rate": 2.4808743169398908e-05,
141
+ "loss": 1.0154,
142
+ "step": 9500
143
+ },
144
+ {
145
+ "epoch": 0.9107468123861566,
146
+ "grad_norm": 17.17100715637207,
147
+ "learning_rate": 2.453551912568306e-05,
148
+ "loss": 1.0149,
149
+ "step": 10000
150
+ },
151
+ {
152
+ "epoch": 0.9562841530054644,
153
+ "grad_norm": 10.567970275878906,
154
+ "learning_rate": 2.4262295081967215e-05,
155
+ "loss": 1.007,
156
+ "step": 10500
157
+ },
158
+ {
159
+ "epoch": 1.0018214936247722,
160
+ "grad_norm": 22.679931640625,
161
+ "learning_rate": 2.398907103825137e-05,
162
+ "loss": 0.995,
163
+ "step": 11000
164
+ },
165
+ {
166
+ "epoch": 1.04735883424408,
167
+ "grad_norm": 12.259163856506348,
168
+ "learning_rate": 2.371584699453552e-05,
169
+ "loss": 0.7177,
170
+ "step": 11500
171
+ },
172
+ {
173
+ "epoch": 1.092896174863388,
174
+ "grad_norm": 18.350513458251953,
175
+ "learning_rate": 2.3442622950819672e-05,
176
+ "loss": 0.7203,
177
+ "step": 12000
178
+ },
179
+ {
180
+ "epoch": 1.138433515482696,
181
+ "grad_norm": 21.86613655090332,
182
+ "learning_rate": 2.3169398907103826e-05,
183
+ "loss": 0.7585,
184
+ "step": 12500
185
+ },
186
+ {
187
+ "epoch": 1.1839708561020037,
188
+ "grad_norm": 18.372119903564453,
189
+ "learning_rate": 2.289617486338798e-05,
190
+ "loss": 0.7324,
191
+ "step": 13000
192
+ },
193
+ {
194
+ "epoch": 1.2295081967213115,
195
+ "grad_norm": 25.320674896240234,
196
+ "learning_rate": 2.262295081967213e-05,
197
+ "loss": 0.7241,
198
+ "step": 13500
199
+ },
200
+ {
201
+ "epoch": 1.2750455373406193,
202
+ "grad_norm": 10.435434341430664,
203
+ "learning_rate": 2.2349726775956283e-05,
204
+ "loss": 0.7322,
205
+ "step": 14000
206
+ },
207
+ {
208
+ "epoch": 1.3205828779599271,
209
+ "grad_norm": 29.3476505279541,
210
+ "learning_rate": 2.2076502732240437e-05,
211
+ "loss": 0.745,
212
+ "step": 14500
213
+ },
214
+ {
215
+ "epoch": 1.366120218579235,
216
+ "grad_norm": 8.349024772644043,
217
+ "learning_rate": 2.180327868852459e-05,
218
+ "loss": 0.7543,
219
+ "step": 15000
220
+ },
221
+ {
222
+ "epoch": 1.4116575591985427,
223
+ "grad_norm": 10.698665618896484,
224
+ "learning_rate": 2.153005464480874e-05,
225
+ "loss": 0.7475,
226
+ "step": 15500
227
+ },
228
+ {
229
+ "epoch": 1.4571948998178508,
230
+ "grad_norm": 10.810667037963867,
231
+ "learning_rate": 2.1256830601092897e-05,
232
+ "loss": 0.7336,
233
+ "step": 16000
234
+ },
235
+ {
236
+ "epoch": 1.5027322404371586,
237
+ "grad_norm": 14.816246032714844,
238
+ "learning_rate": 2.098360655737705e-05,
239
+ "loss": 0.7393,
240
+ "step": 16500
241
+ },
242
+ {
243
+ "epoch": 1.5482695810564664,
244
+ "grad_norm": 11.557518005371094,
245
+ "learning_rate": 2.0710382513661204e-05,
246
+ "loss": 0.7333,
247
+ "step": 17000
248
+ },
249
+ {
250
+ "epoch": 1.5938069216757742,
251
+ "grad_norm": 16.669546127319336,
252
+ "learning_rate": 2.0437158469945358e-05,
253
+ "loss": 0.7255,
254
+ "step": 17500
255
+ },
256
+ {
257
+ "epoch": 1.639344262295082,
258
+ "grad_norm": 14.28069019317627,
259
+ "learning_rate": 2.0163934426229508e-05,
260
+ "loss": 0.7369,
261
+ "step": 18000
262
+ },
263
+ {
264
+ "epoch": 1.6848816029143898,
265
+ "grad_norm": 22.36986541748047,
266
+ "learning_rate": 1.9890710382513662e-05,
267
+ "loss": 0.7287,
268
+ "step": 18500
269
+ },
270
+ {
271
+ "epoch": 1.7304189435336976,
272
+ "grad_norm": 11.109138488769531,
273
+ "learning_rate": 1.9617486338797815e-05,
274
+ "loss": 0.7406,
275
+ "step": 19000
276
+ },
277
+ {
278
+ "epoch": 1.7759562841530054,
279
+ "grad_norm": 16.205177307128906,
280
+ "learning_rate": 1.934426229508197e-05,
281
+ "loss": 0.734,
282
+ "step": 19500
283
+ },
284
+ {
285
+ "epoch": 1.8214936247723132,
286
+ "grad_norm": 16.029556274414062,
287
+ "learning_rate": 1.907103825136612e-05,
288
+ "loss": 0.7261,
289
+ "step": 20000
290
+ },
291
+ {
292
+ "epoch": 1.867030965391621,
293
+ "grad_norm": 17.520740509033203,
294
+ "learning_rate": 1.8797814207650273e-05,
295
+ "loss": 0.7289,
296
+ "step": 20500
297
+ },
298
+ {
299
+ "epoch": 1.9125683060109289,
300
+ "grad_norm": 31.18975257873535,
301
+ "learning_rate": 1.8524590163934426e-05,
302
+ "loss": 0.7012,
303
+ "step": 21000
304
+ },
305
+ {
306
+ "epoch": 1.9581056466302367,
307
+ "grad_norm": 11.743003845214844,
308
+ "learning_rate": 1.825136612021858e-05,
309
+ "loss": 0.7185,
310
+ "step": 21500
311
+ },
312
+ {
313
+ "epoch": 2.0036429872495445,
314
+ "grad_norm": 13.794853210449219,
315
+ "learning_rate": 1.7978142076502733e-05,
316
+ "loss": 0.6816,
317
+ "step": 22000
318
+ },
319
+ {
320
+ "epoch": 2.0491803278688523,
321
+ "grad_norm": 28.844125747680664,
322
+ "learning_rate": 1.7704918032786887e-05,
323
+ "loss": 0.4469,
324
+ "step": 22500
325
+ },
326
+ {
327
+ "epoch": 2.09471766848816,
328
+ "grad_norm": 5.327408790588379,
329
+ "learning_rate": 1.743169398907104e-05,
330
+ "loss": 0.4577,
331
+ "step": 23000
332
+ },
333
+ {
334
+ "epoch": 2.1402550091074684,
335
+ "grad_norm": 29.065576553344727,
336
+ "learning_rate": 1.7158469945355194e-05,
337
+ "loss": 0.4617,
338
+ "step": 23500
339
+ },
340
+ {
341
+ "epoch": 2.185792349726776,
342
+ "grad_norm": 7.028495788574219,
343
+ "learning_rate": 1.6885245901639344e-05,
344
+ "loss": 0.4542,
345
+ "step": 24000
346
+ },
347
+ {
348
+ "epoch": 2.231329690346084,
349
+ "grad_norm": 17.614421844482422,
350
+ "learning_rate": 1.6612021857923497e-05,
351
+ "loss": 0.4809,
352
+ "step": 24500
353
+ },
354
+ {
355
+ "epoch": 2.276867030965392,
356
+ "grad_norm": 16.79141616821289,
357
+ "learning_rate": 1.633879781420765e-05,
358
+ "loss": 0.4717,
359
+ "step": 25000
360
+ },
361
+ {
362
+ "epoch": 2.3224043715846996,
363
+ "grad_norm": 29.323766708374023,
364
+ "learning_rate": 1.6065573770491805e-05,
365
+ "loss": 0.4758,
366
+ "step": 25500
367
+ },
368
+ {
369
+ "epoch": 2.3679417122040074,
370
+ "grad_norm": 19.425329208374023,
371
+ "learning_rate": 1.5792349726775955e-05,
372
+ "loss": 0.4679,
373
+ "step": 26000
374
+ },
375
+ {
376
+ "epoch": 2.413479052823315,
377
+ "grad_norm": 5.9673380851745605,
378
+ "learning_rate": 1.5519125683060108e-05,
379
+ "loss": 0.4885,
380
+ "step": 26500
381
+ },
382
+ {
383
+ "epoch": 2.459016393442623,
384
+ "grad_norm": 12.733803749084473,
385
+ "learning_rate": 1.5245901639344264e-05,
386
+ "loss": 0.4811,
387
+ "step": 27000
388
+ },
389
+ {
390
+ "epoch": 2.504553734061931,
391
+ "grad_norm": 22.212753295898438,
392
+ "learning_rate": 1.4972677595628415e-05,
393
+ "loss": 0.4613,
394
+ "step": 27500
395
+ },
396
+ {
397
+ "epoch": 2.5500910746812386,
398
+ "grad_norm": 18.736804962158203,
399
+ "learning_rate": 1.4699453551912569e-05,
400
+ "loss": 0.4786,
401
+ "step": 28000
402
+ },
403
+ {
404
+ "epoch": 2.5956284153005464,
405
+ "grad_norm": 20.595645904541016,
406
+ "learning_rate": 1.442622950819672e-05,
407
+ "loss": 0.4997,
408
+ "step": 28500
409
+ },
410
+ {
411
+ "epoch": 2.6411657559198543,
412
+ "grad_norm": 18.30236053466797,
413
+ "learning_rate": 1.4153005464480874e-05,
414
+ "loss": 0.4792,
415
+ "step": 29000
416
+ },
417
+ {
418
+ "epoch": 2.686703096539162,
419
+ "grad_norm": 19.671579360961914,
420
+ "learning_rate": 1.3879781420765026e-05,
421
+ "loss": 0.4556,
422
+ "step": 29500
423
+ },
424
+ {
425
+ "epoch": 2.73224043715847,
426
+ "grad_norm": 14.210960388183594,
427
+ "learning_rate": 1.3606557377049181e-05,
428
+ "loss": 0.4737,
429
+ "step": 30000
430
+ },
431
+ {
432
+ "epoch": 2.7777777777777777,
433
+ "grad_norm": 15.706062316894531,
434
+ "learning_rate": 1.3333333333333333e-05,
435
+ "loss": 0.4773,
436
+ "step": 30500
437
+ },
438
+ {
439
+ "epoch": 2.8233151183970855,
440
+ "grad_norm": 2.883192300796509,
441
+ "learning_rate": 1.3060109289617487e-05,
442
+ "loss": 0.4734,
443
+ "step": 31000
444
+ },
445
+ {
446
+ "epoch": 2.8688524590163933,
447
+ "grad_norm": 12.449463844299316,
448
+ "learning_rate": 1.2786885245901639e-05,
449
+ "loss": 0.4579,
450
+ "step": 31500
451
+ },
452
+ {
453
+ "epoch": 2.9143897996357016,
454
+ "grad_norm": 19.76999855041504,
455
+ "learning_rate": 1.2513661202185792e-05,
456
+ "loss": 0.4772,
457
+ "step": 32000
458
+ },
459
+ {
460
+ "epoch": 2.959927140255009,
461
+ "grad_norm": 38.29131317138672,
462
+ "learning_rate": 1.2240437158469946e-05,
463
+ "loss": 0.4689,
464
+ "step": 32500
465
+ },
466
+ {
467
+ "epoch": 3.0054644808743167,
468
+ "grad_norm": 4.822471618652344,
469
+ "learning_rate": 1.19672131147541e-05,
470
+ "loss": 0.4374,
471
+ "step": 33000
472
+ },
473
+ {
474
+ "epoch": 3.051001821493625,
475
+ "grad_norm": 21.936952590942383,
476
+ "learning_rate": 1.1693989071038253e-05,
477
+ "loss": 0.2893,
478
+ "step": 33500
479
+ },
480
+ {
481
+ "epoch": 3.096539162112933,
482
+ "grad_norm": 25.08475685119629,
483
+ "learning_rate": 1.1420765027322405e-05,
484
+ "loss": 0.2888,
485
+ "step": 34000
486
+ },
487
+ {
488
+ "epoch": 3.1420765027322406,
489
+ "grad_norm": 13.904535293579102,
490
+ "learning_rate": 1.1147540983606558e-05,
491
+ "loss": 0.2915,
492
+ "step": 34500
493
+ },
494
+ {
495
+ "epoch": 3.1876138433515484,
496
+ "grad_norm": 3.9972009658813477,
497
+ "learning_rate": 1.087431693989071e-05,
498
+ "loss": 0.3081,
499
+ "step": 35000
500
+ },
501
+ {
502
+ "epoch": 3.233151183970856,
503
+ "grad_norm": 30.3763370513916,
504
+ "learning_rate": 1.0601092896174864e-05,
505
+ "loss": 0.3158,
506
+ "step": 35500
507
+ },
508
+ {
509
+ "epoch": 3.278688524590164,
510
+ "grad_norm": 4.742646217346191,
511
+ "learning_rate": 1.0327868852459017e-05,
512
+ "loss": 0.2869,
513
+ "step": 36000
514
+ },
515
+ {
516
+ "epoch": 3.324225865209472,
517
+ "grad_norm": 42.35346984863281,
518
+ "learning_rate": 1.005464480874317e-05,
519
+ "loss": 0.2932,
520
+ "step": 36500
521
+ },
522
+ {
523
+ "epoch": 3.3697632058287796,
524
+ "grad_norm": 13.786494255065918,
525
+ "learning_rate": 9.781420765027323e-06,
526
+ "loss": 0.2976,
527
+ "step": 37000
528
+ },
529
+ {
530
+ "epoch": 3.4153005464480874,
531
+ "grad_norm": 11.841999053955078,
532
+ "learning_rate": 9.508196721311476e-06,
533
+ "loss": 0.2938,
534
+ "step": 37500
535
+ },
536
+ {
537
+ "epoch": 3.4608378870673953,
538
+ "grad_norm": 29.609792709350586,
539
+ "learning_rate": 9.234972677595628e-06,
540
+ "loss": 0.3006,
541
+ "step": 38000
542
+ },
543
+ {
544
+ "epoch": 3.506375227686703,
545
+ "grad_norm": 18.663509368896484,
546
+ "learning_rate": 8.961748633879782e-06,
547
+ "loss": 0.3039,
548
+ "step": 38500
549
+ },
550
+ {
551
+ "epoch": 3.551912568306011,
552
+ "grad_norm": 1.889535903930664,
553
+ "learning_rate": 8.688524590163935e-06,
554
+ "loss": 0.3082,
555
+ "step": 39000
556
+ },
557
+ {
558
+ "epoch": 3.5974499089253187,
559
+ "grad_norm": 19.884906768798828,
560
+ "learning_rate": 8.415300546448089e-06,
561
+ "loss": 0.3085,
562
+ "step": 39500
563
+ },
564
+ {
565
+ "epoch": 3.6429872495446265,
566
+ "grad_norm": 12.94946002960205,
567
+ "learning_rate": 8.14207650273224e-06,
568
+ "loss": 0.3086,
569
+ "step": 40000
570
+ },
571
+ {
572
+ "epoch": 3.6885245901639343,
573
+ "grad_norm": 27.61246681213379,
574
+ "learning_rate": 7.868852459016394e-06,
575
+ "loss": 0.3109,
576
+ "step": 40500
577
+ },
578
+ {
579
+ "epoch": 3.734061930783242,
580
+ "grad_norm": 14.3275146484375,
581
+ "learning_rate": 7.595628415300546e-06,
582
+ "loss": 0.3025,
583
+ "step": 41000
584
+ },
585
+ {
586
+ "epoch": 3.7795992714025504,
587
+ "grad_norm": 6.992326259613037,
588
+ "learning_rate": 7.3224043715847e-06,
589
+ "loss": 0.3108,
590
+ "step": 41500
591
+ },
592
+ {
593
+ "epoch": 3.8251366120218577,
594
+ "grad_norm": 5.084273815155029,
595
+ "learning_rate": 7.049180327868853e-06,
596
+ "loss": 0.3012,
597
+ "step": 42000
598
+ },
599
+ {
600
+ "epoch": 3.870673952641166,
601
+ "grad_norm": 27.192960739135742,
602
+ "learning_rate": 6.775956284153006e-06,
603
+ "loss": 0.2996,
604
+ "step": 42500
605
+ },
606
+ {
607
+ "epoch": 3.9162112932604733,
608
+ "grad_norm": 42.71589279174805,
609
+ "learning_rate": 6.502732240437159e-06,
610
+ "loss": 0.31,
611
+ "step": 43000
612
+ },
613
+ {
614
+ "epoch": 3.9617486338797816,
615
+ "grad_norm": 11.759929656982422,
616
+ "learning_rate": 6.229508196721312e-06,
617
+ "loss": 0.2953,
618
+ "step": 43500
619
+ },
620
+ {
621
+ "epoch": 4.007285974499089,
622
+ "grad_norm": 30.232032775878906,
623
+ "learning_rate": 5.956284153005465e-06,
624
+ "loss": 0.2612,
625
+ "step": 44000
626
+ },
627
+ {
628
+ "epoch": 4.052823315118397,
629
+ "grad_norm": 62.14134979248047,
630
+ "learning_rate": 5.683060109289618e-06,
631
+ "loss": 0.1843,
632
+ "step": 44500
633
+ },
634
+ {
635
+ "epoch": 4.098360655737705,
636
+ "grad_norm": 7.216005802154541,
637
+ "learning_rate": 5.409836065573771e-06,
638
+ "loss": 0.1859,
639
+ "step": 45000
640
+ },
641
+ {
642
+ "epoch": 4.143897996357013,
643
+ "grad_norm": 17.130714416503906,
644
+ "learning_rate": 5.136612021857924e-06,
645
+ "loss": 0.2106,
646
+ "step": 45500
647
+ },
648
+ {
649
+ "epoch": 4.18943533697632,
650
+ "grad_norm": 4.9142985343933105,
651
+ "learning_rate": 4.863387978142076e-06,
652
+ "loss": 0.1932,
653
+ "step": 46000
654
+ },
655
+ {
656
+ "epoch": 4.2349726775956285,
657
+ "grad_norm": 13.510531425476074,
658
+ "learning_rate": 4.59016393442623e-06,
659
+ "loss": 0.2081,
660
+ "step": 46500
661
+ },
662
+ {
663
+ "epoch": 4.280510018214937,
664
+ "grad_norm": 5.9509172439575195,
665
+ "learning_rate": 4.316939890710383e-06,
666
+ "loss": 0.1861,
667
+ "step": 47000
668
+ },
669
+ {
670
+ "epoch": 4.326047358834244,
671
+ "grad_norm": 7.964664936065674,
672
+ "learning_rate": 4.043715846994535e-06,
673
+ "loss": 0.1863,
674
+ "step": 47500
675
+ },
676
+ {
677
+ "epoch": 4.371584699453552,
678
+ "grad_norm": 25.88454246520996,
679
+ "learning_rate": 3.770491803278689e-06,
680
+ "loss": 0.1938,
681
+ "step": 48000
682
+ },
683
+ {
684
+ "epoch": 4.41712204007286,
685
+ "grad_norm": 2.440138339996338,
686
+ "learning_rate": 3.4972677595628415e-06,
687
+ "loss": 0.1963,
688
+ "step": 48500
689
+ },
690
+ {
691
+ "epoch": 4.462659380692168,
692
+ "grad_norm": 6.778371334075928,
693
+ "learning_rate": 3.2240437158469947e-06,
694
+ "loss": 0.1829,
695
+ "step": 49000
696
+ },
697
+ {
698
+ "epoch": 4.508196721311475,
699
+ "grad_norm": 31.79254722595215,
700
+ "learning_rate": 2.9508196721311474e-06,
701
+ "loss": 0.1897,
702
+ "step": 49500
703
+ },
704
+ {
705
+ "epoch": 4.553734061930784,
706
+ "grad_norm": 42.17744064331055,
707
+ "learning_rate": 2.6775956284153005e-06,
708
+ "loss": 0.1921,
709
+ "step": 50000
710
+ },
711
+ {
712
+ "epoch": 4.599271402550091,
713
+ "grad_norm": 0.5351423621177673,
714
+ "learning_rate": 2.4043715846994536e-06,
715
+ "loss": 0.2084,
716
+ "step": 50500
717
+ },
718
+ {
719
+ "epoch": 4.644808743169399,
720
+ "grad_norm": 19.15989112854004,
721
+ "learning_rate": 2.1311475409836063e-06,
722
+ "loss": 0.1898,
723
+ "step": 51000
724
+ },
725
+ {
726
+ "epoch": 4.6903460837887065,
727
+ "grad_norm": 3.171184539794922,
728
+ "learning_rate": 1.8579234972677596e-06,
729
+ "loss": 0.1928,
730
+ "step": 51500
731
+ },
732
+ {
733
+ "epoch": 4.735883424408015,
734
+ "grad_norm": 61.36581802368164,
735
+ "learning_rate": 1.5846994535519126e-06,
736
+ "loss": 0.198,
737
+ "step": 52000
738
+ },
739
+ {
740
+ "epoch": 4.781420765027322,
741
+ "grad_norm": 6.246685028076172,
742
+ "learning_rate": 1.3114754098360657e-06,
743
+ "loss": 0.1794,
744
+ "step": 52500
745
+ },
746
+ {
747
+ "epoch": 4.82695810564663,
748
+ "grad_norm": 10.183932304382324,
749
+ "learning_rate": 1.0382513661202186e-06,
750
+ "loss": 0.1845,
751
+ "step": 53000
752
+ },
753
+ {
754
+ "epoch": 4.872495446265938,
755
+ "grad_norm": 29.7720947265625,
756
+ "learning_rate": 7.650273224043716e-07,
757
+ "loss": 0.1889,
758
+ "step": 53500
759
+ },
760
+ {
761
+ "epoch": 4.918032786885246,
762
+ "grad_norm": 9.350275039672852,
763
+ "learning_rate": 4.918032786885246e-07,
764
+ "loss": 0.1941,
765
+ "step": 54000
766
+ },
767
+ {
768
+ "epoch": 4.963570127504553,
769
+ "grad_norm": 16.810739517211914,
770
+ "learning_rate": 2.185792349726776e-07,
771
+ "loss": 0.2008,
772
+ "step": 54500
773
+ },
774
+ {
775
+ "epoch": 5.0,
776
+ "step": 54900,
777
+ "total_flos": 1.2910084832623104e+17,
778
+ "train_loss": 0.07113273273183131,
779
+ "train_runtime": 1918.8519,
780
+ "train_samples_per_second": 343.315,
781
+ "train_steps_per_second": 28.611
782
+ }
783
+ ],
784
+ "logging_steps": 500,
785
+ "max_steps": 54900,
786
+ "num_input_tokens_seen": 0,
787
+ "num_train_epochs": 5,
788
+ "save_steps": 500,
789
+ "stateful_callbacks": {
790
+ "TrainerControl": {
791
+ "args": {
792
+ "should_epoch_stop": false,
793
+ "should_evaluate": false,
794
+ "should_log": false,
795
+ "should_save": true,
796
+ "should_training_stop": true
797
+ },
798
+ "attributes": {}
799
+ }
800
+ },
801
+ "total_flos": 1.2910084832623104e+17,
802
+ "train_batch_size": 12,
803
+ "trial_name": null,
804
+ "trial_params": null
805
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a564d101b061d721d938cc1ef56b2bc15757c8cf5e7104dab8df30d58cdedfa2
3
+ size 5368
vocab.txt ADDED
The diff for this file is too large to render. See raw diff