Basma2423 commited on
Commit
d30e3a9
·
verified ·
1 Parent(s): b73d67e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -1,58 +1,32 @@
1
  {
2
- "activation_dropout": 0.1,
3
- "activation_function": "gelu",
4
- "add_bias_logits": false,
5
- "add_final_layer_norm": true,
6
  "architectures": [
7
- "MBartForConditionalGeneration"
8
  ],
9
- "attention_dropout": 0.1,
10
- "bos_token_id": 0,
11
- "classif_dropout": 0.1,
12
  "classifier_dropout": 0.0,
 
 
13
  "d_model": 768,
14
- "decoder_attention_heads": 12,
15
- "decoder_ffn_dim": 3072,
16
- "decoder_layerdrop": 0.0,
17
- "decoder_layers": 6,
18
- "decoder_start_token_id": 2,
19
- "do_blenderbot_90_layernorm": false,
20
- "dropout": 0.1,
21
- "early_stopping": null,
22
- "encoder_attention_heads": 12,
23
- "encoder_ffn_dim": 3072,
24
- "encoder_layerdrop": 0.0,
25
- "encoder_layers": 6,
26
- "eos_token_id": 2,
27
- "extra_pos_embeddings": 2,
28
- "force_bos_token_to_be_generated": false,
29
- "forced_eos_token_id": 2,
30
- "gradient_checkpointing": false,
31
- "id2label": {
32
- "0": "LABEL_0",
33
- "1": "LABEL_1",
34
- "2": "LABEL_2"
35
- },
36
- "init_std": 0.02,
37
  "is_encoder_decoder": true,
38
- "label2id": {
39
- "LABEL_0": 0,
40
- "LABEL_1": 1,
41
- "LABEL_2": 2
42
- },
43
- "max_position_embeddings": 1024,
44
- "model_type": "mbart",
45
- "no_repeat_ngram_size": null,
46
- "normalize_before": true,
47
- "normalize_embedding": true,
48
- "num_beams": null,
49
- "num_hidden_layers": 6,
50
- "pad_token_id": 1,
51
- "scale_embedding": false,
52
- "static_position_embeddings": false,
53
- "tokenizer_class": "BarthezTokenizer",
54
  "torch_dtype": "float32",
55
- "transformers_version": "4.51.3",
56
- "use_cache": false,
57
- "vocab_size": 50594
58
  }
 
1
  {
 
 
 
 
2
  "architectures": [
3
+ "T5ForConditionalGeneration"
4
  ],
 
 
 
5
  "classifier_dropout": 0.0,
6
+ "d_ff": 2048,
7
+ "d_kv": 64,
8
  "d_model": 768,
9
+ "decoder_start_token_id": 0,
10
+ "dense_act_fn": "gelu_new",
11
+ "dropout_rate": 0.1,
12
+ "eos_token_id": 1,
13
+ "feed_forward_proj": "gated-gelu",
14
+ "initializer_factor": 1.0,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  "is_encoder_decoder": true,
16
+ "is_gated_act": true,
17
+ "layer_norm_epsilon": 1e-06,
18
+ "model_type": "t5",
19
+ "num_decoder_layers": 12,
20
+ "num_heads": 12,
21
+ "num_layers": 12,
22
+ "output_past": true,
23
+ "pad_token_id": 0,
24
+ "relative_attention_max_distance": 128,
25
+ "relative_attention_num_buckets": 32,
26
+ "tie_word_embeddings": false,
27
+ "tokenizer_class": "T5Tokenizer",
 
 
 
 
28
  "torch_dtype": "float32",
29
+ "transformers_version": "4.53.0",
30
+ "use_cache": true,
31
+ "vocab_size": 110208
32
  }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "pad_token_id": 0,
5
+ "transformers_version": "4.53.0"
6
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2139aa0df60063c5f2a89b26029f1e46a73d98ca4786df7e9f5e8f8aa2ce3fd8
3
- size 558937304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:030be25143ccd1d78f909c41318fb2ecc301e9c81c4835b77323d13e3a68f361
3
+ size 1470068592
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38b1742741bbc7df44be28b94d10d9ab05ad1cce2d8e70ac745824b916791397
3
+ size 2996811
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd54aa42eef0dd482320f74af4126bc2997a856ba65d8619476c5b9c42dc0d7e
3
+ size 14645
scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9380964b4b2206fff8b4f1535e87c70866b61a1d77c927a05f1aaa0e184890d1
3
+ size 1383
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6218a7bb0fe5f214de8ffed80e034797b70aa538fcf9b5e6aedd7d40c0317f31
3
+ size 1465
special_tokens_map.json CHANGED
@@ -1,18 +1,106 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "cls_token": {
10
- "content": "<s>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  "eos_token": {
17
  "content": "</s>",
18
  "lstrip": false,
@@ -20,13 +108,6 @@
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
- "mask_token": {
24
- "content": "<mask>",
25
- "lstrip": true,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
  "pad_token": {
31
  "content": "<pad>",
32
  "lstrip": false,
@@ -34,13 +115,6 @@
34
  "rstrip": false,
35
  "single_word": false
36
  },
37
- "sep_token": {
38
- "content": "</s>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false
43
- },
44
  "unk_token": {
45
  "content": "<unk>",
46
  "lstrip": false,
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>",
99
+ "<extra_id_96>",
100
+ "<extra_id_97>",
101
+ "<extra_id_98>",
102
+ "<extra_id_99>"
103
+ ],
104
  "eos_token": {
105
  "content": "</s>",
106
  "lstrip": false,
 
108
  "rstrip": false,
109
  "single_word": false
110
  },
 
 
 
 
 
 
 
111
  "pad_token": {
112
  "content": "<pad>",
113
  "lstrip": false,
 
115
  "rstrip": false,
116
  "single_word": false
117
  },
 
 
 
 
 
 
 
118
  "unk_token": {
119
  "content": "<unk>",
120
  "lstrip": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff
 
trainer_state.json CHANGED
@@ -1,142 +1,364 @@
1
  {
2
- "best_global_step": 5570,
3
- "best_metric": 0.7717538371980056,
4
- "best_model_checkpoint": "./checkpoints/checkpoint-5570",
5
- "epoch": 0.9998653923812088,
6
- "eval_steps": 1114,
7
- "global_step": 5571,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "epoch": 0.0897384125274824,
14
- "grad_norm": 0.10910310596227646,
15
- "learning_rate": 4.901672070547205e-05,
16
- "loss": 2.0607,
17
  "step": 500
18
  },
19
  {
20
- "epoch": 0.1794768250549648,
21
- "grad_norm": 0.11283092200756073,
22
- "learning_rate": 4.6136704554635066e-05,
23
- "loss": 0.1219,
24
  "step": 1000
25
  },
26
  {
27
- "epoch": 0.19993718311123077,
28
- "eval_cer": 0.7744026050691318,
29
- "eval_loss": 0.06905069202184677,
30
- "eval_runtime": 2200.6581,
31
- "eval_samples_per_second": 80.463,
32
- "eval_steps_per_second": 0.21,
33
- "step": 1114
34
- },
35
- {
36
- "epoch": 0.26921523758244714,
37
- "grad_norm": 0.07702942937612534,
38
- "learning_rate": 4.9413107833705604e-05,
39
- "loss": 0.0709,
40
  "step": 1500
41
  },
42
  {
43
- "epoch": 0.3589536501099296,
44
- "grad_norm": 0.048379454761743546,
45
- "learning_rate": 4.695072154439477e-05,
46
- "loss": 0.0526,
47
- "step": 2000
 
 
 
 
 
48
  },
49
  {
50
- "epoch": 0.39987436622246153,
51
- "eval_cer": 0.7726060406484112,
52
- "eval_loss": 0.03332927078008652,
53
- "eval_runtime": 436.8886,
54
- "eval_samples_per_second": 405.3,
55
- "eval_steps_per_second": 1.057,
56
- "step": 2228
57
  },
58
  {
59
- "epoch": 0.44869206263741196,
60
- "grad_norm": 0.04356146231293678,
61
- "learning_rate": 4.970863550514242e-05,
62
- "loss": 0.0398,
63
  "step": 2500
64
  },
65
  {
66
- "epoch": 0.5384304751648943,
67
- "grad_norm": 0.03718848526477814,
68
- "learning_rate": 4.767405209364289e-05,
69
- "loss": 0.0348,
70
  "step": 3000
71
  },
72
  {
73
- "epoch": 0.5998115493336923,
74
- "eval_cer": 0.7720533296293449,
75
- "eval_loss": 0.022723019123077393,
76
- "eval_runtime": 439.0491,
77
- "eval_samples_per_second": 403.306,
78
- "eval_steps_per_second": 1.052,
79
- "step": 3342
 
 
 
80
  },
81
  {
82
- "epoch": 0.6281688876923768,
83
- "grad_norm": 0.03279432654380798,
84
- "learning_rate": 4.384876624004233e-05,
85
- "loss": 0.03,
86
  "step": 3500
87
  },
88
  {
89
- "epoch": 0.7179073002198592,
90
- "grad_norm": 0.032072387635707855,
91
- "learning_rate": 3.8534883130104934e-05,
92
- "loss": 0.027,
93
  "step": 4000
94
  },
95
  {
96
- "epoch": 0.7997487324449231,
97
- "eval_cer": 0.7718523934618883,
98
- "eval_loss": 0.01841309294104576,
99
- "eval_runtime": 436.0956,
100
- "eval_samples_per_second": 406.037,
101
- "eval_steps_per_second": 1.059,
102
- "step": 4456
103
  },
104
  {
105
- "epoch": 0.8076457127473415,
106
- "grad_norm": 0.0274814460426569,
107
- "learning_rate": 3.2152071176988856e-05,
108
- "loss": 0.0248,
109
- "step": 4500
 
 
 
 
 
110
  },
111
  {
112
- "epoch": 0.8973841252748239,
113
- "grad_norm": 0.025461740791797638,
114
- "learning_rate": 2.520441835335607e-05,
115
- "loss": 0.0232,
116
  "step": 5000
117
  },
118
  {
119
- "epoch": 0.9871225378023063,
120
- "grad_norm": 0.026277432218194008,
121
- "learning_rate": 1.8240621417367316e-05,
122
- "loss": 0.0222,
123
  "step": 5500
124
  },
125
  {
126
- "epoch": 0.9996859155561538,
127
- "eval_cer": 0.7717538371980056,
128
- "eval_loss": 0.016575267538428307,
129
- "eval_runtime": 437.5161,
130
- "eval_samples_per_second": 404.719,
131
- "eval_steps_per_second": 1.056,
132
- "step": 5570
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
133
  }
134
  ],
135
  "logging_steps": 500,
136
- "max_steps": 5571,
137
  "num_input_tokens_seen": 0,
138
- "num_train_epochs": 1,
139
- "save_steps": 1114,
140
  "stateful_callbacks": {
141
  "TrainerControl": {
142
  "args": {
@@ -144,13 +366,13 @@
144
  "should_evaluate": false,
145
  "should_log": false,
146
  "should_save": true,
147
- "should_training_stop": true
148
  },
149
  "attributes": {}
150
  }
151
  },
152
- "total_flos": 6.522138159281603e+17,
153
- "train_batch_size": 384,
154
  "trial_name": null,
155
  "trial_params": null
156
  }
 
1
  {
2
+ "best_global_step": 16280,
3
+ "best_metric": 0.05104936846864332,
4
+ "best_model_checkpoint": "checkpoints/checkpoint-16280",
5
+ "epoch": 1.9992631708215645,
6
+ "eval_steps": 1628,
7
+ "global_step": 16280,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "epoch": 0.06140243153628884,
14
+ "grad_norm": 0.31788545846939087,
15
+ "learning_rate": 4.99487482593161e-05,
16
+ "loss": 0.639,
17
  "step": 500
18
  },
19
  {
20
+ "epoch": 0.12280486307257768,
21
+ "grad_norm": 0.21716584265232086,
22
+ "learning_rate": 4.9794381003990576e-05,
23
+ "loss": 0.3961,
24
  "step": 1000
25
  },
26
  {
27
+ "epoch": 0.1842072946088665,
28
+ "grad_norm": 0.32310155034065247,
29
+ "learning_rate": 4.9537535426562465e-05,
30
+ "loss": 0.3303,
 
 
 
 
 
 
 
 
 
31
  "step": 1500
32
  },
33
  {
34
+ "epoch": 0.19992631708215644,
35
+ "eval_avg": 0.07433640047160056,
36
+ "eval_cer": 0.048444434926464684,
37
+ "eval_der": 0.054747223027151605,
38
+ "eval_loss": 0.2203781008720398,
39
+ "eval_runtime": 1149.449,
40
+ "eval_samples_per_second": 45.035,
41
+ "eval_steps_per_second": 0.352,
42
+ "eval_wer": 0.1198175434611854,
43
+ "step": 1628
44
  },
45
  {
46
+ "epoch": 0.24560972614515536,
47
+ "grad_norm": 0.1829567700624466,
48
+ "learning_rate": 4.997155120906588e-05,
49
+ "loss": 0.287,
50
+ "step": 2000
 
 
51
  },
52
  {
53
+ "epoch": 0.3070121576814442,
54
+ "grad_norm": 0.1720687448978424,
55
+ "learning_rate": 4.984333190353011e-05,
56
+ "loss": 0.2682,
57
  "step": 2500
58
  },
59
  {
60
+ "epoch": 0.368414589217733,
61
+ "grad_norm": 0.19322410225868225,
62
+ "learning_rate": 4.961243195561407e-05,
63
+ "loss": 0.2506,
64
  "step": 3000
65
  },
66
  {
67
+ "epoch": 0.3998526341643129,
68
+ "eval_avg": 0.061321180853411116,
69
+ "eval_cer": 0.04388846969480041,
70
+ "eval_der": 0.03895160822949064,
71
+ "eval_loss": 0.18064290285110474,
72
+ "eval_runtime": 1155.5329,
73
+ "eval_samples_per_second": 44.798,
74
+ "eval_steps_per_second": 0.35,
75
+ "eval_wer": 0.10112346463594228,
76
+ "step": 3256
77
  },
78
  {
79
+ "epoch": 0.42981702075402184,
80
+ "grad_norm": 0.16168585419654846,
81
+ "learning_rate": 4.998779393767303e-05,
82
+ "loss": 0.2339,
83
  "step": 3500
84
  },
85
  {
86
+ "epoch": 0.4912194522903107,
87
+ "grad_norm": 0.12988826632499695,
88
+ "learning_rate": 4.988596305882391e-05,
89
+ "loss": 0.2271,
90
  "step": 4000
91
  },
92
  {
93
+ "epoch": 0.5526218838265995,
94
+ "grad_norm": 0.11692152917385101,
95
+ "learning_rate": 4.9681275337621505e-05,
96
+ "loss": 0.2181,
97
+ "step": 4500
 
 
98
  },
99
  {
100
+ "epoch": 0.5997789512464694,
101
+ "eval_avg": 0.06223662921729023,
102
+ "eval_cer": 0.03700322837976026,
103
+ "eval_der": 0.055956727518601486,
104
+ "eval_loss": 0.19135934114456177,
105
+ "eval_runtime": 1153.4306,
106
+ "eval_samples_per_second": 44.879,
107
+ "eval_steps_per_second": 0.351,
108
+ "eval_wer": 0.09374993175350896,
109
+ "step": 4884
110
  },
111
  {
112
+ "epoch": 0.6140243153628884,
113
+ "grad_norm": 0.12457627058029175,
114
+ "learning_rate": 4.9374576774384186e-05,
115
+ "loss": 0.2112,
116
  "step": 5000
117
  },
118
  {
119
+ "epoch": 0.6754267468991773,
120
+ "grad_norm": 0.13500584661960602,
121
+ "learning_rate": 4.8967134993181585e-05,
122
+ "loss": 0.2042,
123
  "step": 5500
124
  },
125
  {
126
+ "epoch": 0.736829178435466,
127
+ "grad_norm": 0.14300259947776794,
128
+ "learning_rate": 4.846063400258344e-05,
129
+ "loss": 0.1994,
130
+ "step": 6000
131
+ },
132
+ {
133
+ "epoch": 0.7982316099717549,
134
+ "grad_norm": 0.15229490399360657,
135
+ "learning_rate": 4.785716723543856e-05,
136
+ "loss": 0.1941,
137
+ "step": 6500
138
+ },
139
+ {
140
+ "epoch": 0.7997052683286258,
141
+ "eval_avg": 0.05848093903880644,
142
+ "eval_cer": 0.034985685323039174,
143
+ "eval_der": 0.05257239447503411,
144
+ "eval_loss": 0.17743617296218872,
145
+ "eval_runtime": 1154.8064,
146
+ "eval_samples_per_second": 44.826,
147
+ "eval_steps_per_second": 0.351,
148
+ "eval_wer": 0.08788473731834603,
149
+ "step": 6512
150
+ },
151
+ {
152
+ "epoch": 0.8596340415080437,
153
+ "grad_norm": 0.10620284080505371,
154
+ "learning_rate": 4.9950986655011213e-05,
155
+ "loss": 0.1892,
156
+ "step": 7000
157
+ },
158
+ {
159
+ "epoch": 0.9210364730443326,
160
+ "grad_norm": 0.10538509488105774,
161
+ "learning_rate": 4.9798882653846754e-05,
162
+ "loss": 0.1843,
163
+ "step": 7500
164
+ },
165
+ {
166
+ "epoch": 0.9824389045806214,
167
+ "grad_norm": 0.11998545378446579,
168
+ "learning_rate": 4.954428172468978e-05,
169
+ "loss": 0.1814,
170
+ "step": 8000
171
+ },
172
+ {
173
+ "epoch": 0.9996315854107822,
174
+ "eval_avg": 0.05574338139826937,
175
+ "eval_cer": 0.0337392713322537,
176
+ "eval_der": 0.05011513570945588,
177
+ "eval_loss": 0.16507230699062347,
178
+ "eval_runtime": 1158.6086,
179
+ "eval_samples_per_second": 44.679,
180
+ "eval_steps_per_second": 0.35,
181
+ "eval_wer": 0.08337573715309854,
182
+ "step": 8140
183
+ },
184
+ {
185
+ "epoch": 1.0438413361169103,
186
+ "grad_norm": 0.12560783326625824,
187
+ "learning_rate": 4.9189048589485774e-05,
188
+ "loss": 0.1764,
189
+ "step": 8500
190
+ },
191
+ {
192
+ "epoch": 1.105243767653199,
193
+ "grad_norm": 0.10523095726966858,
194
+ "learning_rate": 4.8733228183318665e-05,
195
+ "loss": 0.1735,
196
+ "step": 9000
197
+ },
198
+ {
199
+ "epoch": 1.1666461991894879,
200
+ "grad_norm": 0.11041487008333206,
201
+ "learning_rate": 4.817931533427457e-05,
202
+ "loss": 0.1715,
203
+ "step": 9500
204
+ },
205
+ {
206
+ "epoch": 1.1995579024929388,
207
+ "eval_avg": 0.0544335846636687,
208
+ "eval_cer": 0.03342891994782228,
209
+ "eval_der": 0.048513281174538105,
210
+ "eval_loss": 0.1576041877269745,
211
+ "eval_runtime": 1159.9625,
212
+ "eval_samples_per_second": 44.626,
213
+ "eval_steps_per_second": 0.349,
214
+ "eval_wer": 0.08135855286864571,
215
+ "step": 9768
216
+ },
217
+ {
218
+ "epoch": 1.2280486307257767,
219
+ "grad_norm": 0.1073341891169548,
220
+ "learning_rate": 4.998896962187747e-05,
221
+ "loss": 0.1674,
222
+ "step": 10000
223
+ },
224
+ {
225
+ "epoch": 1.2894510622620656,
226
+ "grad_norm": 0.11093363165855408,
227
+ "learning_rate": 4.988961418344719e-05,
228
+ "loss": 0.167,
229
+ "step": 10500
230
+ },
231
+ {
232
+ "epoch": 1.3508534937983545,
233
+ "grad_norm": 0.09265608340501785,
234
+ "learning_rate": 4.968738681210237e-05,
235
+ "loss": 0.165,
236
+ "step": 11000
237
+ },
238
+ {
239
+ "epoch": 1.3994842195750952,
240
+ "eval_avg": 0.05450751736133735,
241
+ "eval_cer": 0.03384634860456726,
242
+ "eval_der": 0.04814585144402502,
243
+ "eval_loss": 0.15283821523189545,
244
+ "eval_runtime": 1150.9092,
245
+ "eval_samples_per_second": 44.977,
246
+ "eval_steps_per_second": 0.352,
247
+ "eval_wer": 0.08153035203541975,
248
+ "step": 11396
249
+ },
250
+ {
251
+ "epoch": 1.4122559253346432,
252
+ "grad_norm": 0.10535308718681335,
253
+ "learning_rate": 4.999780685937908e-05,
254
+ "loss": 0.1617,
255
+ "step": 11500
256
+ },
257
+ {
258
+ "epoch": 1.473658356870932,
259
+ "grad_norm": 0.09392343461513519,
260
+ "learning_rate": 4.992486965915939e-05,
261
+ "loss": 0.161,
262
+ "step": 12000
263
+ },
264
+ {
265
+ "epoch": 1.535060788407221,
266
+ "grad_norm": 0.0898473933339119,
267
+ "learning_rate": 4.974891481067152e-05,
268
+ "loss": 0.1593,
269
+ "step": 12500
270
+ },
271
+ {
272
+ "epoch": 1.5964632199435098,
273
+ "grad_norm": 0.10352090746164322,
274
+ "learning_rate": 4.94706695576259e-05,
275
+ "loss": 0.1581,
276
+ "step": 13000
277
+ },
278
+ {
279
+ "epoch": 1.5994105366572517,
280
+ "eval_avg": 0.052648695144827624,
281
+ "eval_cer": 0.03269155476372104,
282
+ "eval_der": 0.04647078141601352,
283
+ "eval_loss": 0.1520785242319107,
284
+ "eval_runtime": 1154.3982,
285
+ "eval_samples_per_second": 44.842,
286
+ "eval_steps_per_second": 0.351,
287
+ "eval_wer": 0.07878374925474832,
288
+ "step": 13024
289
+ },
290
+ {
291
+ "epoch": 1.6578656514797987,
292
+ "grad_norm": 0.08162818104028702,
293
+ "learning_rate": 4.995337159613658e-05,
294
+ "loss": 0.1563,
295
+ "step": 13500
296
+ },
297
+ {
298
+ "epoch": 1.7192680830160874,
299
+ "grad_norm": 0.10093415528535843,
300
+ "learning_rate": 4.9803736946497604e-05,
301
+ "loss": 0.1549,
302
+ "step": 14000
303
+ },
304
+ {
305
+ "epoch": 1.7806705145523762,
306
+ "grad_norm": 0.0779104083776474,
307
+ "learning_rate": 4.9551585305458784e-05,
308
+ "loss": 0.1531,
309
+ "step": 14500
310
+ },
311
+ {
312
+ "epoch": 1.799336853739408,
313
+ "eval_avg": 0.05215597602151043,
314
+ "eval_cer": 0.032804504015484064,
315
+ "eval_der": 0.046157249106538095,
316
+ "eval_loss": 0.1437014639377594,
317
+ "eval_runtime": 1158.5459,
318
+ "eval_samples_per_second": 44.681,
319
+ "eval_steps_per_second": 0.35,
320
+ "eval_wer": 0.07750617494250915,
321
+ "step": 14652
322
+ },
323
+ {
324
+ "epoch": 1.8420729460886651,
325
+ "grad_norm": 0.09926008433103561,
326
+ "learning_rate": 4.919795884774044e-05,
327
+ "loss": 0.1519,
328
+ "step": 15000
329
+ },
330
+ {
331
+ "epoch": 1.903475377624954,
332
+ "grad_norm": 0.0932171419262886,
333
+ "learning_rate": 4.874431915635338e-05,
334
+ "loss": 0.1508,
335
+ "step": 15500
336
+ },
337
+ {
338
+ "epoch": 1.9648778091612429,
339
+ "grad_norm": 0.08440640568733215,
340
+ "learning_rate": 4.819254118169078e-05,
341
+ "loss": 0.1493,
342
+ "step": 16000
343
+ },
344
+ {
345
+ "epoch": 1.9992631708215645,
346
+ "eval_avg": 0.05104936846864332,
347
+ "eval_cer": 0.03196083873281993,
348
+ "eval_der": 0.04519598184101095,
349
+ "eval_loss": 0.1432473361492157,
350
+ "eval_runtime": 1163.7456,
351
+ "eval_samples_per_second": 44.481,
352
+ "eval_steps_per_second": 0.348,
353
+ "eval_wer": 0.07599128483209908,
354
+ "step": 16280
355
  }
356
  ],
357
  "logging_steps": 500,
358
+ "max_steps": 24429,
359
  "num_input_tokens_seen": 0,
360
+ "num_train_epochs": 3,
361
+ "save_steps": 1628,
362
  "stateful_callbacks": {
363
  "TrainerControl": {
364
  "args": {
 
366
  "should_evaluate": false,
367
  "should_log": false,
368
  "should_save": true,
369
+ "should_training_stop": false
370
  },
371
  "attributes": {}
372
  }
373
  },
374
+ "total_flos": 1.8107754052964844e+18,
375
+ "train_batch_size": 128,
376
  "trial_name": null,
377
  "trial_params": null
378
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4bce4c2a570e5406234205443c86622d09135dc71fc8963eb2e85ed7f5af2b7c
3
- size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:994dcd51f10df37a598334a01dc75b4d6c85da6c809dc43933bd3b4a8f34c7ee
3
+ size 5905