bongsoo commited on
Commit
fa97e79
1 Parent(s): cbac07d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -91,8 +91,9 @@ For an automated evaluation of this model, see the *Sentence Embeddings Benchmar
91
 
92
 
93
  ## Training
94
- -[albert-small-kor-v1](https://huggingface.co/bongsoo/albert-small-kor-v1) 모델을 **sts(10)-distil(10) 훈련만 시킴** (**nli-sts추가 훈련시키면 정합도가 떨어짐**)
95
-
 
96
  The model was trained with the parameters:
97
 
98
  **공통**
@@ -104,7 +105,7 @@ The model was trained with the parameters:
104
  - 훈련코드 [여기](https://github.com/kobongsoo/BERT/blob/master/sbert/sentece-bert-sts.ipynb) 참조
105
 
106
  **2.distilation**
107
- - 교사 모델 : kpf-sbert-v1(max_token_len:128)
108
  - 말뭉치 : news_talk_ko_en_train.tsv (한국어-영어 대화-뉴스 병렬 말뭉치 : 1.38M)
109
  - Param : **lr: 5e-5, epochs: 10, train_batch: 32, eval/test_batch: 64, max_token_len: 128(교사모델이 128이므로 맟춰줌)**
110
  - 훈련코드 [여기](https://github.com/kobongsoo/BERT/blob/master/sbert/sbert-distillaton2.ipynb) 참조
 
91
 
92
 
93
  ## Training
94
+ - [albert-small-kor-v1](https://huggingface.co/bongsoo/albert-small-kor-v1) 모델을 **sts(10)-distil(10) 훈련만 시킴** (**nli-sts추가 훈련시키면 정합도가 떨어짐**)
95
+ - 교사모델은 [**kpf-sbert-v1.1**](https://huggingface.co/bongsoo/kpf-sbert-v1.1) 이용함.
96
+
97
  The model was trained with the parameters:
98
 
99
  **공통**
 
105
  - 훈련코드 [여기](https://github.com/kobongsoo/BERT/blob/master/sbert/sentece-bert-sts.ipynb) 참조
106
 
107
  **2.distilation**
108
+ - 교사 모델 : [**kpf-sbert-v1.1**](https://huggingface.co/bongsoo/kpf-sbert-v1.1)(max_token_len:128)
109
  - 말뭉치 : news_talk_ko_en_train.tsv (한국어-영어 대화-뉴스 병렬 말뭉치 : 1.38M)
110
  - Param : **lr: 5e-5, epochs: 10, train_batch: 32, eval/test_batch: 64, max_token_len: 128(교사모델이 128이므로 맟춰줌)**
111
  - 훈련코드 [여기](https://github.com/kobongsoo/BERT/blob/master/sbert/sbert-distillaton2.ipynb) 참조