Big (100B+) LLMs (OUTDATED)
Collection
Collection of various big LLMs(Not MOE).
•
47 items
•
Updated
•
6
I took base llama 2 70b model and frankenmerged it with itself using mergekit. Somehow it is coherent.
Thanks for featuring me at https://merge.moe/. I'll try my best to make even more good(!?) merges in the future.
Test name | Base llama | Dicephal |
---|---|---|
B | 0 | 0 |
C | 2 | 0 |
D | 0.5 | 1 |
S | 1.25 | 2.25 |
P | 0 | 2.25 |
Total | 3.75 | 5.5 |
+75% in size, +47% in meme benchmark performance!
name | whacky | left/right |
---|---|---|
ChuckMcSneed/Dicephal-123B | 1.742262578 | -0.131433424 |
meta-llama/Llama-2-70b-hf | 1.930293804 | 0.178771095 |