akoksal commited on
Commit
33bd403
1 Parent(s): 690de87

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -5
README.md CHANGED
@@ -46,11 +46,7 @@ The LongForm dataset is created by leveraging English corpus examples with augme
46
 
47
  Github Repo: https://github.com/akoksal/LongForm
48
 
49
- ### For LongForm-OPT models: Use [EOI] to indicate the end of instruction.
50
-
51
- LongForm-**T5-XL**: https://huggingface.co/akoksal/LongForm-T5-XL
52
-
53
- LongForm-**OPT-2.7B**: https://huggingface.co/akoksal/LongForm-OPT-2.7B
54
 
55
  ## How to Load
56
  ```python
@@ -85,11 +81,19 @@ We provide in-depth evaluation of LongForm models and baselines in the paper. We
85
  | [**LongForm-OPT-6.7B**](https://huggingface.co/akoksal/LongForm-OPT-6.7B) | 17.7 | 16.9 | 17.2 | 19.0 |
86
  | [**LongForm-LLaMA-7B**](https://huggingface.co/akoksal/LongForm-LLaMA-7B-diff)‡ | **19.7** | **21.7** | **18.6** | 18.9 |
87
 
 
 
 
 
 
 
88
  ‡: We can just release the difference between LongForm-LLaMA-7B and pretrained LLaMA-7B publicly due to restrictions of LLaMA models.
89
 
90
  ## Limitations
91
  The LongForm dataset and models mainly focus on long text generation and have limitations regarding structured prediction tasks in NLP. Additionally, we observe that LongForm models may present hallucination problems similar to those found in LLMs.
92
 
 
 
93
 
94
  ## Citation
95
  ```
 
46
 
47
  Github Repo: https://github.com/akoksal/LongForm
48
 
49
+ ### For LongForm OPT and LLaMA models: Use [EOI] to indicate the end of instruction.
 
 
 
 
50
 
51
  ## How to Load
52
  ```python
 
81
  | [**LongForm-OPT-6.7B**](https://huggingface.co/akoksal/LongForm-OPT-6.7B) | 17.7 | 16.9 | 17.2 | 19.0 |
82
  | [**LongForm-LLaMA-7B**](https://huggingface.co/akoksal/LongForm-LLaMA-7B-diff)‡ | **19.7** | **21.7** | **18.6** | 18.9 |
83
 
84
+ Smaller versions of LongForm-OPT models are also available:
85
+ - [**LongForm-OPT-1.3B**](https://huggingface.co/akoksal/LongForm-OPT-1.3B)
86
+ - [**LongForm-OPT-350M**](https://huggingface.co/akoksal/LongForm-OPT-350M)
87
+ - [**LongForm-OPT-125M**](https://huggingface.co/akoksal/LongForm-OPT-125M)
88
+
89
+
90
  ‡: We can just release the difference between LongForm-LLaMA-7B and pretrained LLaMA-7B publicly due to restrictions of LLaMA models.
91
 
92
  ## Limitations
93
  The LongForm dataset and models mainly focus on long text generation and have limitations regarding structured prediction tasks in NLP. Additionally, we observe that LongForm models may present hallucination problems similar to those found in LLMs.
94
 
95
+ ## License
96
+ The LongForm project is subject to a MIT License with custom limitations for restrictions imposed by OpenAI (for the instruction generation part), as well as the license of language models (OPT, LLaMA, and T5).
97
 
98
  ## Citation
99
  ```