nayohan commited on
Commit
2bfaaf0
1 Parent(s): 4ced979

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -14
README.md CHANGED
@@ -7,32 +7,36 @@ license: cc-by-4.0
7
  KoQuality-Polyglot-5.8b is a fine-tuned version of [EleutherAI/polyglot-ko-5.8b](https://huggingface.co/EleutherAI/polyglot-ko-5.8b) on [KoQuality dataset](https://huggingface.co/datasets/DILAB-HYU/KoQuality), which is curated by proposed method (len_group=5, k=100, n=0.01, method=ppl_sampling).
8
 
9
 
10
- ### Overall Average accuracy score of the KoBEST datasets
11
 
12
  We use [KoBEST benchmark](https://huggingface.co/datasets/skt/kobest_v1) datasets(BoolQ, COPA, HellaSwag, SentiNeg, WiC) to compare the performance of our best model and other models accuracy. Our model outperforms other models in the average accuracy score of the KoBEST datasets.
13
  <img src=https://cdn-uploads.huggingface.co/production/uploads/650fecfd247f564485f8fbcf/q4cCUCzRJa3m2f7oxI_FY.png style="max-width: 500px; width: 300%"/>
14
 
15
  | Model | 0-shot | 1-shot | 2-shot | 5-shot | 10-shot
16
  | --- | --- | --- | --- | --- | --- |
17
- | koquality-polyglot-5.8b | 0.5472 | 0.5979 | 0.6260 | 0.6486 | 0.6535
18
  | polyglot-ko-5.8b | 0.5587 | 0.5977 | 0.6138 | 0.6431 | 0.6457
19
  | koalpcaca-polyglot-5.8b | 0.5085 | 0.5561 | 0.5768 | 0.6097 | 0.6059
20
  | kullm-polyglot-5.8b | 0.5409 | 0.6072 | 0.5945 | 0.6345 | 0.6530
 
21
 
 
 
22
 
23
- ### Evaluation results
24
- COPA (F1)
25
  <img src=https://cdn-uploads.huggingface.co/production/uploads/650fecfd247f564485f8fbcf/7EKl1OAgKgPBFcSlGzBiW.png style="max-width: 800px; width: 400%"/>
 
 
 
 
 
 
26
 
27
- HellaSwag (F1)
28
-
29
- BoolQ (F1)
30
-
31
- SentiNeg (F1)
32
 
33
- WiC (F1)
34
 
 
35
 
 
36
 
37
 
38
 
@@ -48,10 +52,10 @@ WiC (F1)
48
  - num_epochs: 2.0
49
 
50
  ## Framework versions
51
- Transformers 4.30.2
52
- Pytorch 2.0.1+cu117
53
- Datasets 2.11.0
54
- deepspeed 0.9.5
55
 
56
  ## Citation
57
 
 
7
  KoQuality-Polyglot-5.8b is a fine-tuned version of [EleutherAI/polyglot-ko-5.8b](https://huggingface.co/EleutherAI/polyglot-ko-5.8b) on [KoQuality dataset](https://huggingface.co/datasets/DILAB-HYU/KoQuality), which is curated by proposed method (len_group=5, k=100, n=0.01, method=ppl_sampling).
8
 
9
 
10
+ ## Overall Average accuracy score of the KoBEST datasets
11
 
12
  We use [KoBEST benchmark](https://huggingface.co/datasets/skt/kobest_v1) datasets(BoolQ, COPA, HellaSwag, SentiNeg, WiC) to compare the performance of our best model and other models accuracy. Our model outperforms other models in the average accuracy score of the KoBEST datasets.
13
  <img src=https://cdn-uploads.huggingface.co/production/uploads/650fecfd247f564485f8fbcf/q4cCUCzRJa3m2f7oxI_FY.png style="max-width: 500px; width: 300%"/>
14
 
15
  | Model | 0-shot | 1-shot | 2-shot | 5-shot | 10-shot
16
  | --- | --- | --- | --- | --- | --- |
 
17
  | polyglot-ko-5.8b | 0.5587 | 0.5977 | 0.6138 | 0.6431 | 0.6457
18
  | koalpcaca-polyglot-5.8b | 0.5085 | 0.5561 | 0.5768 | 0.6097 | 0.6059
19
  | kullm-polyglot-5.8b | 0.5409 | 0.6072 | 0.5945 | 0.6345 | 0.6530
20
+ | koquality-polyglot-5.8b | 0.5472 | 0.5979 | 0.6260 | 0.6486 | 0.6535
21
 
22
+ ## Evaluation results
23
+ ### COPA (F1)
24
 
 
 
25
  <img src=https://cdn-uploads.huggingface.co/production/uploads/650fecfd247f564485f8fbcf/7EKl1OAgKgPBFcSlGzBiW.png style="max-width: 800px; width: 400%"/>
26
+ | Model | 0-shot | 1-shot | 2-shot | 5-shot | 10-shot
27
+ | --- | --- | --- | --- | --- | --- |
28
+ | polyglot-ko-5.8b | 0.5587 | 0.5977 | 0.6138 | 0.6431 | 0.6457
29
+ | koalpcaca-polyglot-5.8b | 0.5085 | 0.5561 | 0.5768 | 0.6097 | 0.6059
30
+ | kullm-polyglot-5.8b | 0.5409 | 0.6072 | 0.5945 | 0.6345 | 0.6530
31
+ | koquality-polyglot-5.8b | 0.5472 | 0.5979 | 0.6260 | 0.6486 | 0.6535
32
 
33
+ ### HellaSwag (F1)
 
 
 
 
34
 
35
+ ### BoolQ (F1)
36
 
37
+ ### SentiNeg (F1)
38
 
39
+ ### WiC (F1)
40
 
41
 
42
 
 
52
  - num_epochs: 2.0
53
 
54
  ## Framework versions
55
+ - Transformers 4.30.2
56
+ - Pytorch 2.0.1+cu117
57
+ - Datasets 2.11.0
58
+ - deepspeed 0.9.5
59
 
60
  ## Citation
61