antalvdb commited on
Commit
96f93c0
1 Parent(s): 050e67d

Upload 12 files

Browse files
all_results.json CHANGED
@@ -1,14 +1,8 @@
1
  {
2
  "epoch": 2.0,
3
- "eval_cer": 0.01165665167457964,
4
- "eval_loss": 0.005274078343063593,
5
- "eval_runtime": 1990.4275,
6
- "eval_samples": 2000,
7
- "eval_samples_per_second": 1.005,
8
- "eval_steps_per_second": 0.251,
9
- "train_loss": 0.016045775709704408,
10
- "train_runtime": 206173.8627,
11
- "train_samples": 2960086,
12
- "train_samples_per_second": 28.714,
13
- "train_steps_per_second": 0.897
14
  }
 
1
  {
2
  "epoch": 2.0,
3
+ "train_loss": 0.013614018808192265,
4
+ "train_runtime": 415192.129,
5
+ "train_samples": 6342779,
6
+ "train_samples_per_second": 30.553,
7
+ "train_steps_per_second": 0.955
 
 
 
 
 
 
8
  }
config.json CHANGED
@@ -69,7 +69,7 @@
69
  }
70
  },
71
  "torch_dtype": "float32",
72
- "transformers_version": "4.27.3",
73
  "use_cache": true,
74
  "vocab_size": 50265
75
  }
 
69
  }
70
  },
71
  "torch_dtype": "float32",
72
+ "transformers_version": "4.30.2",
73
  "use_cache": true,
74
  "vocab_size": 50265
75
  }
generation_config.json CHANGED
@@ -8,5 +8,5 @@
8
  "no_repeat_ngram_size": 3,
9
  "num_beams": 4,
10
  "pad_token_id": 1,
11
- "transformers_version": "4.27.3"
12
  }
 
8
  "no_repeat_ngram_size": 3,
9
  "num_beams": 4,
10
  "pad_token_id": 1,
11
+ "transformers_version": "4.30.2"
12
  }
merges.txt CHANGED
@@ -1,4 +1,4 @@
1
- #version: 0.2 - Trained by `huggingface/tokenizers`
2
  Ġ t
3
  Ġ a
4
  h e
 
1
+ #version: 0.2
2
  Ġ t
3
  Ġ a
4
  h e
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96696783fd7389a128057a8994d2b0151034ffed0645a6a6c5b271c1514f2b0b
3
  size 557971229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ed26ed7830bddb2e53106a9f1e2d9eac7db42270845f34d6f3315c387c147b5
3
  size 557971229
tokenizer.json CHANGED
@@ -87,6 +87,7 @@
87
  "continuing_subword_prefix": "",
88
  "end_of_word_suffix": "",
89
  "fuse_unk": false,
 
90
  "vocab": {
91
  "<s>": 0,
92
  "<pad>": 1,
 
87
  "continuing_subword_prefix": "",
88
  "end_of_word_suffix": "",
89
  "fuse_unk": false,
90
+ "byte_fallback": false,
91
  "vocab": {
92
  "<s>": 0,
93
  "<pad>": 1,
tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<s>",
 
4
  "cls_token": "<s>",
5
  "eos_token": "</s>",
6
  "errors": "replace",
@@ -8,7 +9,6 @@
8
  "model_max_length": 1024,
9
  "pad_token": "<pad>",
10
  "sep_token": "</s>",
11
- "special_tokens_map_file": null,
12
  "tokenizer_class": "BartTokenizer",
13
  "trim_offsets": true,
14
  "unk_token": "<unk>"
 
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<s>",
4
+ "clean_up_tokenization_spaces": true,
5
  "cls_token": "<s>",
6
  "eos_token": "</s>",
7
  "errors": "replace",
 
9
  "model_max_length": 1024,
10
  "pad_token": "<pad>",
11
  "sep_token": "</s>",
 
12
  "tokenizer_class": "BartTokenizer",
13
  "trim_offsets": true,
14
  "unk_token": "<unk>"
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 2.0,
3
- "train_loss": 0.016045775709704408,
4
- "train_runtime": 206173.8627,
5
- "train_samples": 2960086,
6
- "train_samples_per_second": 28.714,
7
- "train_steps_per_second": 0.897
8
  }
 
1
  {
2
  "epoch": 2.0,
3
+ "train_loss": 0.013614018808192265,
4
+ "train_runtime": 415192.129,
5
+ "train_samples": 6342779,
6
+ "train_samples_per_second": 30.553,
7
+ "train_steps_per_second": 0.955
8
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e466518c4579bd6bee53fdf30377e705184fd9f7e036056a6b09423ab880b7fd
3
- size 3707
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cd15191290a2c0e137c6f35a6f650076ece6690fadfa42b7621140db2393fbb
3
+ size 4155