LoneStriker commited on
Commit
8f2f583
1 Parent(s): c6870a5

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -1,35 +1,9 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ Umbra-v2.1-MoE-4x10.7-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
2
+ Umbra-v2.1-MoE-4x10.7-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
3
+ Umbra-v2.1-MoE-4x10.7-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
4
+ Umbra-v2.1-MoE-4x10.7-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
5
+ Umbra-v2.1-MoE-4x10.7-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
6
+ Umbra-v2.1-MoE-4x10.7-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
7
+ Umbra-v2.1-MoE-4x10.7-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
8
+ Umbra-v2.1-MoE-4x10.7-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
9
+ Umbra-v2.1-MoE-4x10.7-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - moe
5
+ - merge
6
+ - mergekit
7
+ - vicgalle/CarbonBeagle-11B
8
+ - Sao10K/Fimbulvetr-10.7B-v1
9
+ - bn22/Nous-Hermes-2-SOLAR-10.7B-MISALIGNED
10
+ - Yhyu13/LMCocktail-10.7B-v1
11
+ ---
12
+
13
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64545af5ec40bbbd01242ca6/hen3fNHRD7BCPvd2KkfjZ.png)
14
+
15
+ # Umbra-v2.1-MoE-4x10.7
16
+
17
+ The [Umbra Series] is an offshoot of the [Lumosia Series] With the goal to be a General assistant that has a knack for story telling and RP/ERP
18
+
19
+ -What's New in v2.1?
20
+
21
+ Umbra v2.1 isn't just a simple update; it's like giving the model a double shot of espresso. Ive changed the models and prompts, in an attempt to make Umbra
22
+ not only your go-to assistant for general knowledge but also a great storyteller and RP/ERP companion.
23
+
24
+ -Longer Positive, Shorter Negative
25
+
26
+ In an effort to trick the gates into being less uptight, Ive added more positive prompts and snappier negative ones.
27
+ These changes are based on the model's strengths and, frankly, my whimsical preferences.
28
+
29
+ -Experimental, As Always
30
+
31
+ Remember, folks, "v2.1" doesn't mean it's superior to its predecessors – it's just another step in the quest.
32
+ It's the 'Empire Strikes Back' of our series – could be better, could be worse, but definitely more dramatic.
33
+
34
+ -Base Context and Coherence
35
+
36
+ Umbra v2.1 has a base context of 8k scrolling window.
37
+
38
+ -The Tavern Card
39
+
40
+ Just for fun - the Umbra Personality Tavern Card. It's your gateway to immersive storytelling experiences,
41
+ a little like having a 'Choose Your Own Adventure' book, but way cooler because it's digital and doesn't get lost under your bed.
42
+
43
+ -Token Error? Fixed!
44
+
45
+ Umbra-v2 had a tokenizer error but was removed faster than you can say "Cops love Donuts"
46
+
47
+ So, give Umbra v2.1 a whirl and let me know how it goes. Your feedback is like the secret sauce in my development burger.
48
+
49
+ ```
50
+ ### System:
51
+
52
+ ### USER:{prompt}
53
+
54
+ ### Assistant:
55
+ ```
56
+
57
+ Settings:
58
+ ```
59
+ Temp: 1.0
60
+ min-p: 0.02-0.1
61
+ ```
62
+
63
+ ## Evals:
64
+
65
+ * Avg: 73.59
66
+ * ARC: 69.11
67
+ * HellaSwag: 87.57
68
+ * MMLU: 66.48
69
+ * T-QA: 66.75
70
+ * Winogrande: 83.11
71
+ * GSM8K: 68.69
72
+
73
+ ## Examples:
74
+ ```
75
+ posted soon
76
+ ```
77
+ ```
78
+ posted soon
79
+ ```
80
+
81
+ ## 🧩 Configuration
82
+
83
+ ```
84
+ base_model: vicgalle/CarbonBeagle-11B
85
+ gate_mode: hidden
86
+ dtype: bfloat16
87
+ experts:
88
+ - source_model: vicgalle/CarbonBeagle-11B
89
+ positive_prompts: [Revamped]
90
+
91
+ - source_model: Sao10K/Fimbulvetr-10.7B-v1
92
+ positive_prompts: [Revamped]
93
+
94
+ - source_model: bn22/Nous-Hermes-2-SOLAR-10.7B-MISALIGNED
95
+ positive_prompts: [Revamped]
96
+
97
+ - source_model: Yhyu13/LMCocktail-10.7B-v1
98
+ positive_prompts: [Revamed]
99
+ ```
100
+ ```
101
+ Umbra-v2-MoE-4x10.7 is a Mixure of Experts (MoE) made with the following models:
102
+ * [vicgalle/CarbonBeagle-11B](https://huggingface.co/vicgalle/CarbonBeagle-11B)
103
+ * [Sao10K/Fimbulvetr-10.7B-v1](https://huggingface.co/Sao10K/Fimbulvetr-10.7B-v1)
104
+ * [bn22/Nous-Hermes-2-SOLAR-10.7B-MISALIGNED](https://huggingface.co/bn22/Nous-Hermes-2-SOLAR-10.7B-MISALIGNED)
105
+ * [Yhyu13/LMCocktail-10.7B-v1](https://huggingface.co/Yhyu13/LMCocktail-10.7B-v1)
106
+
107
+ ```
108
+
109
+ ## 💻 Usage
110
+
111
+ ```python
112
+ !pip install -qU transformers bitsandbytes accelerate
113
+ from transformers import AutoTokenizer
114
+ import transformers
115
+ import torch
116
+
117
+ model = "Steelskull/Umbra-v2-MoE-4x10.7"
118
+
119
+ tokenizer = AutoTokenizer.from_pretrained(model)
120
+ pipeline = transformers.pipeline(
121
+ "text-generation",
122
+ model=model,
123
+ model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
124
+ )
125
+
126
+ messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
127
+ prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
128
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
129
+ print(outputs[0]["generated_text"])
130
+ ```
Umbra-v2.1-MoE-4x10.7-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1ab2fe66b0845f34bcf4ae5e3deda4923a3d41acff50474c625a32de15f08cc
3
+ size 13187216032
Umbra-v2.1-MoE-4x10.7-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27a4626371ddbb7dca62d15f02cf06eeb5ce18fb374eed5114303db350c9ad40
3
+ size 18732290720
Umbra-v2.1-MoE-4x10.7-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c27f58338676302a63efcdf5fad8d1896f6fba6b9330ae5625d8be71cec0ad08
3
+ size 17286304416
Umbra-v2.1-MoE-4x10.7-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9e59ee8c05ae3c1589b208d81134fa885e56f4167d069bd57d81363c0e66739
3
+ size 15566115488
Umbra-v2.1-MoE-4x10.7-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:471eadcfbc0dac0962c33214c74c88b904c947d4edc41707494ceba3c1d8ec5c
3
+ size 21822288544
Umbra-v2.1-MoE-4x10.7-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:588067b33339644534fcb1bcccd36dc967b40c5ce9f110bb7646e3b556f5c1c6
3
+ size 20521267872
Umbra-v2.1-MoE-4x10.7-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95fdac6303275049283e13742e6a7f2dc44df1e22457dec48da376b696f1ba34
3
+ size 25600963232
Umbra-v2.1-MoE-4x10.7-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5484f7915030816b4b4fe4a19ce96956793a6c3c2d9979e55ff58de4cc5ca887
3
+ size 24838910624
Umbra-v2.1-MoE-4x10.7-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d8b23e5d763e8dd21c69e8ed8be8e536be246d778e4f1f674a4c52edca3da60
3
+ size 29615805088
Umbra_Tavern.png ADDED
mergekit_moe_config.yml ADDED
@@ -0,0 +1,113 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: vicgalle/CarbonBeagle-11B
2
+ gate_mode: hidden
3
+ dtype: bfloat16
4
+
5
+ experts:
6
+ - source_model: vicgalle/CarbonBeagle-11B
7
+ positive_prompts:
8
+ - "versatile"
9
+ - "adaptive"
10
+ - "comprehensive"
11
+ - "integrated"
12
+ - "balanced"
13
+ - "all-rounder"
14
+ - "flexible"
15
+ - "wide-ranging"
16
+ - "multi-disciplinary"
17
+ - "holistic"
18
+ - "innovative"
19
+ - "eclectic"
20
+ - "resourceful"
21
+ - "dynamic"
22
+ - "robust"
23
+
24
+ negative_prompts:
25
+ - "narrow"
26
+ - "specialized"
27
+ - "limited"
28
+ - "focused"
29
+
30
+ - source_model: Sao10K/Fimbulvetr-10.7B-v1
31
+ positive_prompts:
32
+ - "creative"
33
+ - "storytelling"
34
+ - "expressive"
35
+ - "imaginative"
36
+ - "engaging"
37
+ - "verbose"
38
+ - "narrative"
39
+ - "descriptive"
40
+ - "elaborate"
41
+ - "fictional"
42
+ - "artistic"
43
+ - "vivid"
44
+ - "colorful"
45
+ - "fantastical"
46
+ - "lyrical"
47
+
48
+ negative_prompts:
49
+ - "sorry"
50
+ - "I cannot"
51
+ - "factual"
52
+ - "concise"
53
+ - "straightforward"
54
+ - "objective"
55
+ - "dry"
56
+
57
+ - source_model: bn22/Nous-Hermes-2-SOLAR-10.7B-MISALIGNED
58
+ positive_prompts:
59
+ - "intelligent"
60
+ - "analytical"
61
+ - "accurate"
62
+ - "knowledgeable"
63
+ - "logical"
64
+ - "data-driven"
65
+ - "scientific"
66
+ - "rational"
67
+ - "precise"
68
+ - "methodical"
69
+ - "empirical"
70
+ - "systematic"
71
+ - "efficient"
72
+ - "scholarly"
73
+ - "statistical"
74
+ - "calculate"
75
+ - "compute"
76
+ - "solve"
77
+ - "work"
78
+ - "python"
79
+ - "javascript"
80
+ - "programming"
81
+ - "algorithm"
82
+ - "tell me"
83
+ - "assistant"
84
+
85
+ negative_prompts:
86
+ - "creative"
87
+ - "imaginative"
88
+ - "abstract"
89
+ - "emotional"
90
+ - "artistic"
91
+ - "speculative"
92
+
93
+ - source_model: Yhyu13/LMCocktail-10.7B-v1
94
+ positive_prompts:
95
+ - "instructive"
96
+ - "verbose"
97
+ - "descriptive"
98
+ - "clear"
99
+ - "detailed"
100
+ - "informative"
101
+ - "explanatory"
102
+ - "elucidative"
103
+ - "articulate"
104
+ - "comprehensive"
105
+ - "educational"
106
+ - "thorough"
107
+ - "specific"
108
+ - "clarifying"
109
+ - "structured"
110
+
111
+ negative_prompts:
112
+ - "concise"
113
+ - "vague"