bhenrym14 commited on
Commit
f50ba57
1 Parent(s): cd1d9ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -23,3 +23,5 @@ This is a finetune of Llama-2-13b, intended to extend the useful context window
23
  All training was performed with 1x RTX 6000 Ada.
24
 
25
  For full model card, including how to use PNTK, see any of the two merged models linked above.
 
 
 
23
  All training was performed with 1x RTX 6000 Ada.
24
 
25
  For full model card, including how to use PNTK, see any of the two merged models linked above.
26
+
27
+ **IMPORTANT** There are TWO sets of adapter weights. `adapter_model_base` is to be applied to llama-2-13b. The result can then be merged with `adapter_model`. These adapters correspond to the first and second training phase respectively. Applying the second phase adapter to base llama-2-13b will produce different results to what the merged full airophin model.