afaji commited on
Commit
aacdbd9
1 Parent(s): 73b3371

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -12
README.md CHANGED
@@ -23,7 +23,7 @@ should probably proofread and complete it, then remove this comment. -->
23
  [![Model License](https://img.shields.io/badge/Model%20License-CC%20By%20NC%204.0-red.svg)]()
24
 
25
  This model is one of our LaMini model series in paper "[LaMini: A Diverse Herd of Distilled Models from Large-Scale Instructions](https://github.com/mbzuai-nlp/lamini)". This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on [LaMini dataset](https://huggingface.co/datasets/MBZUAI/LaMini-instruction) that contains 2.58M samples for instruction fine-tuning. For more information about our dataset, please refer to our [project repository](https://github.com/mbzuai-nlp/lamini/).
26
- You can view other LaMini model series as follow. Note that not all models are performing as well. More details can be seen in our paper.
27
 
28
  <table>
29
  <thead>
@@ -42,9 +42,9 @@ You can view other LaMini model series as follow. Note that not all models are p
42
  </tr>
43
  <tr>
44
  <td>Flan-T5</td>
45
- <td><a href="https://huggingface.co/MBZUAI/lamini-flan-t5-77m" target="_blank" rel="noopener noreferrer">LaMini-Flan-T5-77M</a></td>
46
- <td><a href="https://huggingface.co/MBZUAI/lamini-flan-t5-248m" target="_blank" rel="noopener noreferrer">LaMini-Flan-T5-248M</a></td>
47
- <td><a href="https://huggingface.co/MBZUAI/lamini-flan-t5-783m" target="_blank" rel="noopener noreferrer">LaMini-Flan-T5-783M</a></td>
48
  <td></td>
49
  </tr>
50
  <tr>
@@ -56,9 +56,9 @@ You can view other LaMini model series as follow. Note that not all models are p
56
  </tr>
57
  <tr>
58
  <td>GPT-2</td>
59
- <td><a href="https://huggingface.co/MBZUAI/lamini-gpt-124m" target="_blank" rel="noopener noreferrer">LaMini-GPT-124M</a></td>
60
- <td><a href="https://huggingface.co/MBZUAI/lamini-gpt-774m" target="_blank" rel="noopener noreferrer">LaMini-GPT-774M</a></td>
61
- <td><a href="https://huggingface.co/MBZUAI/lamini-gpt-1.5b" target="_blank" rel="noopener noreferrer">LaMini-GPT-1.5B</a></td>
62
  <td></td>
63
  </tr>
64
  <tr>
@@ -130,12 +130,11 @@ More information needed
130
  # Citation
131
 
132
  ```bibtex
133
- @misc{,
134
- title={LaMini: Distilling Knowledge from Large Language Models},
135
  author={},
136
  year={2023},
137
- eprint={},
138
- archivePrefix={},
139
- primaryClass={}
140
  }
141
  ```
 
23
  [![Model License](https://img.shields.io/badge/Model%20License-CC%20By%20NC%204.0-red.svg)]()
24
 
25
  This model is one of our LaMini model series in paper "[LaMini: A Diverse Herd of Distilled Models from Large-Scale Instructions](https://github.com/mbzuai-nlp/lamini)". This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on [LaMini dataset](https://huggingface.co/datasets/MBZUAI/LaMini-instruction) that contains 2.58M samples for instruction fine-tuning. For more information about our dataset, please refer to our [project repository](https://github.com/mbzuai-nlp/lamini/).
26
+ You can view other LaMini model series as follow. Note that not all models are performing as well. Models with ✩ are those with the best overall performance given their size/architecture. More details can be seen in our paper.
27
 
28
  <table>
29
  <thead>
 
42
  </tr>
43
  <tr>
44
  <td>Flan-T5</td>
45
+ <td><a href="https://huggingface.co/MBZUAI/lamini-flan-t5-77m" target="_blank" rel="noopener noreferrer">LaMini-Flan-T5-77M</a>✩</td>
46
+ <td><a href="https://huggingface.co/MBZUAI/lamini-flan-t5-248m" target="_blank" rel="noopener noreferrer">LaMini-Flan-T5-248M</a>✩</td>
47
+ <td><a href="https://huggingface.co/MBZUAI/lamini-flan-t5-783m" target="_blank" rel="noopener noreferrer">LaMini-Flan-T5-783M</a>✩</td>
48
  <td></td>
49
  </tr>
50
  <tr>
 
56
  </tr>
57
  <tr>
58
  <td>GPT-2</td>
59
+ <td><a href="https://huggingface.co/MBZUAI/lamini-gpt-124m" target="_blank" rel="noopener noreferrer">LaMini-GPT-124M</a>✩</td>
60
+ <td><a href="https://huggingface.co/MBZUAI/lamini-gpt-774m" target="_blank" rel="noopener noreferrer">LaMini-GPT-774M</a>✩</td>
61
+ <td><a href="https://huggingface.co/MBZUAI/lamini-gpt-1.5b" target="_blank" rel="noopener noreferrer">LaMini-GPT-1.5B</a>✩</td>
62
  <td></td>
63
  </tr>
64
  <tr>
 
130
  # Citation
131
 
132
  ```bibtex
133
+ @misc{lamini,
134
+ title={LaMini: A Diverse Herd of Distilled Models from Large-Scale Instructions},
135
  author={},
136
  year={2023},
137
+ publisher = {GitHub},
138
+ journal = {GitHub repository},
 
139
  }
140
  ```