ZZ99 commited on
Commit
c82b424
1 Parent(s): 82ee79c

commit from zz

Browse files
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - generated_from_trainer
4
+ metrics:
5
+ - accuracy
6
+ model-index:
7
+ - name: test-mlm
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # test-mlm
15
+
16
+ This model is a fine-tuned version of [/root/autodl-tmp/nbme/deberta-v3-large](https://huggingface.co//root/autodl-tmp/nbme/deberta-v3-large) on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 1.5185
19
+ - Accuracy: 0.6987
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 1e-05
39
+ - train_batch_size: 4
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
+ - lr_scheduler_type: linear
44
+ - num_epochs: 3.0
45
+
46
+ ### Training results
47
+
48
+
49
+
50
+ ### Framework versions
51
+
52
+ - Transformers 4.19.0.dev0
53
+ - Pytorch 1.11.0
54
+ - Datasets 2.1.0
55
+ - Tokenizers 0.12.1
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[MASK]": 128000}
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.6986770866102621,
4
+ "eval_loss": 1.5184571743011475,
5
+ "eval_runtime": 50.0544,
6
+ "eval_samples": 794,
7
+ "eval_samples_per_second": 15.863,
8
+ "eval_steps_per_second": 1.998,
9
+ "perplexity": 4.565176487471998,
10
+ "train_loss": 1.5083582298156009,
11
+ "train_runtime": 7793.681,
12
+ "train_samples": 14828,
13
+ "train_samples_per_second": 5.708,
14
+ "train_steps_per_second": 1.427
15
+ }
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/autodl-tmp/nbme/deberta-v3-large",
3
+ "architectures": [
4
+ "NewDebertaV2ForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_eps": 1e-07,
13
+ "max_position_embeddings": 512,
14
+ "max_relative_positions": -1,
15
+ "model_type": "deberta-v2",
16
+ "norm_rel_ebd": "layer_norm",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
+ "pad_token_id": 0,
20
+ "pooler_dropout": 0,
21
+ "pooler_hidden_act": "gelu",
22
+ "pooler_hidden_size": 1024,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "position_buckets": 256,
29
+ "relative_attention": true,
30
+ "share_att_key": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.19.0.dev0",
33
+ "type_vocab_size": 0,
34
+ "vocab_size": 128001
35
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.6986770866102621,
4
+ "eval_loss": 1.5184571743011475,
5
+ "eval_runtime": 50.0544,
6
+ "eval_samples": 794,
7
+ "eval_samples_per_second": 15.863,
8
+ "eval_steps_per_second": 1.998,
9
+ "perplexity": 4.565176487471998
10
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f2f66fc7f71282ad9572e83c58abc6aa20f49951290f54bbc8c48b5b57654c7
3
+ size 1740500457
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": false, "vocab_type": "spm", "special_tokens_map_file": null, "name_or_path": "/root/autodl-tmp/nbme/deberta-v3-large", "sp_model_kwargs": {}, "tokenizer_class": "DebertaV2Tokenizer"}
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "train_loss": 1.5083582298156009,
4
+ "train_runtime": 7793.681,
5
+ "train_samples": 14828,
6
+ "train_samples_per_second": 5.708,
7
+ "train_steps_per_second": 1.427
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "global_step": 11121,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.27,
12
+ "learning_rate": 9.100800287743908e-06,
13
+ "loss": 2.8188,
14
+ "step": 1000
15
+ },
16
+ {
17
+ "epoch": 0.54,
18
+ "learning_rate": 8.201600575487818e-06,
19
+ "loss": 1.8053,
20
+ "step": 2000
21
+ },
22
+ {
23
+ "epoch": 0.81,
24
+ "learning_rate": 7.3024008632317245e-06,
25
+ "loss": 1.5897,
26
+ "step": 3000
27
+ },
28
+ {
29
+ "epoch": 1.08,
30
+ "learning_rate": 6.403201150975632e-06,
31
+ "loss": 1.4814,
32
+ "step": 4000
33
+ },
34
+ {
35
+ "epoch": 1.35,
36
+ "learning_rate": 5.50400143871954e-06,
37
+ "loss": 1.3795,
38
+ "step": 5000
39
+ },
40
+ {
41
+ "epoch": 1.62,
42
+ "learning_rate": 4.604801726463448e-06,
43
+ "loss": 1.3396,
44
+ "step": 6000
45
+ },
46
+ {
47
+ "epoch": 1.89,
48
+ "learning_rate": 3.705602014207356e-06,
49
+ "loss": 1.2938,
50
+ "step": 7000
51
+ },
52
+ {
53
+ "epoch": 2.16,
54
+ "learning_rate": 2.806402301951264e-06,
55
+ "loss": 1.2645,
56
+ "step": 8000
57
+ },
58
+ {
59
+ "epoch": 2.43,
60
+ "learning_rate": 1.9072025896951715e-06,
61
+ "loss": 1.2359,
62
+ "step": 9000
63
+ },
64
+ {
65
+ "epoch": 2.7,
66
+ "learning_rate": 1.0080028774390793e-06,
67
+ "loss": 1.2172,
68
+ "step": 10000
69
+ },
70
+ {
71
+ "epoch": 2.97,
72
+ "learning_rate": 1.0880316518298715e-07,
73
+ "loss": 1.205,
74
+ "step": 11000
75
+ },
76
+ {
77
+ "epoch": 3.0,
78
+ "step": 11121,
79
+ "total_flos": 4.147380631930061e+16,
80
+ "train_loss": 1.5083582298156009,
81
+ "train_runtime": 7793.681,
82
+ "train_samples_per_second": 5.708,
83
+ "train_steps_per_second": 1.427
84
+ }
85
+ ],
86
+ "max_steps": 11121,
87
+ "num_train_epochs": 3,
88
+ "total_flos": 4.147380631930061e+16,
89
+ "trial_name": null,
90
+ "trial_params": null
91
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a21949356dc8e894e20a5e23cc4b9ad03bab37ed07c14be91035a352f0b7a0a
3
+ size 3119