nadahlberg commited on
Commit
4d0220e
1 Parent(s): b235196

Model save

Browse files
README.md CHANGED
@@ -16,8 +16,8 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [docketanalyzer/docket-lm-xs](https://huggingface.co/docketanalyzer/docket-lm-xs) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.0635
20
- - F1: 0.9828
21
 
22
  ## Model description
23
 
@@ -37,27 +37,27 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 5e-05
40
- - train_batch_size: 4
41
  - eval_batch_size: 16
42
  - seed: 42
43
- - gradient_accumulation_steps: 4
44
- - total_train_batch_size: 16
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_ratio: 0.02
48
- - num_epochs: 10
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | F1 |
53
  |:-------------:|:-----:|:----:|:---------------:|:------:|
54
- | 0.0787 | 1.42 | 60 | 0.0636 | 0.9739 |
55
- | 0.0053 | 2.84 | 120 | 0.0489 | 0.9828 |
56
- | 0.0029 | 4.26 | 180 | 0.0556 | 0.9828 |
57
- | 0.0019 | 5.68 | 240 | 0.0636 | 0.9828 |
58
- | 0.0014 | 7.1 | 300 | 0.0638 | 0.9828 |
59
- | 0.0012 | 8.52 | 360 | 0.0635 | 0.9828 |
60
- | 0.0012 | 9.94 | 420 | 0.0635 | 0.9828 |
 
 
61
 
62
 
63
  ### Framework versions
 
16
 
17
  This model is a fine-tuned version of [docketanalyzer/docket-lm-xs](https://huggingface.co/docketanalyzer/docket-lm-xs) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.0363
20
+ - F1: 0.9892
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 5e-05
40
+ - train_batch_size: 16
41
  - eval_batch_size: 16
42
  - seed: 42
 
 
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - lr_scheduler_warmup_ratio: 0.02
46
+ - num_epochs: 6
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | F1 |
51
  |:-------------:|:-----:|:----:|:---------------:|:------:|
52
+ | 0.0334 | 0.61 | 60 | 0.0383 | 0.9928 |
53
+ | 0.0077 | 1.22 | 120 | 0.0542 | 0.9786 |
54
+ | 0.004 | 1.84 | 180 | 0.0340 | 0.9892 |
55
+ | 0.0031 | 2.45 | 240 | 0.1027 | 0.9716 |
56
+ | 0.0019 | 3.06 | 300 | 0.0067 | 0.9964 |
57
+ | 0.0036 | 3.67 | 360 | 0.0076 | 0.9964 |
58
+ | 0.0019 | 4.29 | 420 | 0.0472 | 0.9856 |
59
+ | 0.1193 | 4.9 | 480 | 0.0503 | 0.9856 |
60
+ | 0.0014 | 5.51 | 540 | 0.0350 | 0.9892 |
61
 
62
 
63
  ### Framework versions
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "eval_loss": 0.063520148396492,
3
- "eval_f1": 0.9827586206896551,
4
- "eval_runtime": 0.5302,
5
- "eval_samples_per_second": 316.883,
6
- "eval_steps_per_second": 20.748,
7
- "epoch": 9.94
8
  }
 
1
  {
2
+ "eval_loss": 0.03632596880197525,
3
+ "eval_f1": 0.9891696750902527,
4
+ "eval_runtime": 1.1875,
5
+ "eval_samples_per_second": 327.572,
6
+ "eval_steps_per_second": 21.052,
7
+ "epoch": 6.0
8
  }
logs/events.out.tfevents.1707484442.nathan.87297.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bc341721618039a513aedfa54e62d15982df221b2567c0bfea6aa490b85a79f
3
+ size 53839
logs/events.out.tfevents.1707484559.nathan.87297.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3f63791ccaf27577d3176b5bd7b3e603645b4027d72189812e6316d693bdba2
3
+ size 405
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:775b6e2366b5a83c4357af857e43e4c1b0ee2fcbd6ae05e572b0eea078d1bc9f
3
  size 283347432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d95c53f7217430e4e71a07498ef8eb0b8bf6873166aceb1d54205cf5e1475a5a
3
  size 283347432
run_config.json CHANGED
@@ -4,10 +4,10 @@
4
  "max_length": 256
5
  },
6
  "training_args": {
7
- "num_train_epochs": 10,
8
- "per_device_train_batch_size": 4,
9
  "per_device_eval_batch_size": 16,
10
- "gradient_accumulation_steps": 4,
11
  "learning_rate": 5e-05,
12
  "weight_decay": 0.1,
13
  "warmup_ratio": 0.02,
 
4
  "max_length": 256
5
  },
6
  "training_args": {
7
+ "num_train_epochs": 6,
8
+ "per_device_train_batch_size": 16,
9
  "per_device_eval_batch_size": 16,
10
+ "gradient_accumulation_steps": 1,
11
  "learning_rate": 5e-05,
12
  "weight_decay": 0.1,
13
  "warmup_ratio": 0.02,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f32359f205619e9ab451ce60f0473430ad675ed490d4afde3895b1223c970e56
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:257fe51d7cf53fa40026935e0b9ddc3ee3878866616085e360f7c39fa63f0239
3
  size 4664