MDDDDR commited on
Commit
43d0420
1 Parent(s): 292e627

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -7
README.md CHANGED
@@ -1,5 +1,13 @@
 
 
 
 
 
 
 
 
1
  ### Model Card for Model ID
2
- base_model : [Ko-Llama3-Luxia-8B](https://huggingface.co/saltlux/Ko-Llama3-Luxia-8B)
3
 
4
  ### Basic usage
5
  ```python
@@ -22,21 +30,24 @@ print(tokenizer.decode(outputs[0]))
22
  ```
23
 
24
  ### Training dataset
25
- dataset : [kyujinpy/KOpen-platypus](https://huggingface.co/datasets/kyujinpy/KOpen-platypus)
26
 
27
  ### lora_config and bnb_config in Training
28
  ```python
29
  bnd_config = BitsAndBytesConfig(
30
- load_in_4bit = True
 
 
 
31
  )
32
 
33
  lora_config = LoraConfig(
34
- r = 16,
35
- lora_alpha = 16,
36
  lora_dropout = 0.05,
37
- target_modules = ['gate_proj', 'up_proj', 'down_proj']
38
  )
39
  ```
40
 
41
  ### Hardware
42
- RTX 3090 Ti 24GB x 1
 
1
+ ---
2
+ datasets:
3
+ - sean0042/KorMedMCQA
4
+ language:
5
+ - ko
6
+ - en
7
+ pipeline_tag: text-generation
8
+ ---
9
  ### Model Card for Model ID
10
+ base_model : [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it)
11
 
12
  ### Basic usage
13
  ```python
 
30
  ```
31
 
32
  ### Training dataset
33
+ dataset : [sean0042/KorMedMCQA](https://huggingface.co/datasets/sean0042/KorMedMCQA)
34
 
35
  ### lora_config and bnb_config in Training
36
  ```python
37
  bnd_config = BitsAndBytesConfig(
38
+ load_in_4bit = True,
39
+ bnb_4bit_use_double_quant = True,
40
+ bnb_4bit_quant_type = 'nf4',
41
+ bnb_4bit_compute_dtype = torch.bfloat16
42
  )
43
 
44
  lora_config = LoraConfig(
45
+ r = 32,
46
+ lora_alpha = 32,
47
  lora_dropout = 0.05,
48
+ target_modules = ['q_proj', 'k_proj', 'v_proj', 'o_proj', 'gate_proj', 'up_proj', 'down_proj']
49
  )
50
  ```
51
 
52
  ### Hardware
53
+ A100 40GB x 1