basiliskinstitute commited on
Commit
d80331d
1 Parent(s): 7362de1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -1,7 +1,7 @@
1
- This is a storywriting and roleplay model with a significant amount of self generated long context multiturn roleplay.
2
-
3
  If you like this model, consider joining my discord to provide feedback: https://discord.gg/QXdn8hWSkY
4
 
 
 
5
  I downloaded a bit under a thousand cards from chub.ai, and created a synthetic roleplay for each card. I batched as many turns as I could in 4k token chunks in order to maintain coherency over longer context. There was a lot of cleaning and validation between each batch, so a lot of examples were "lost," but the final output seems to be very good quality. The longest conversation is about 20k tokens, and I plan to extend this further as well as broaden the dataset with more examples. The first 4k tokens were generated with Command-R-Plus, with the remainder generated with byroneverson/Mistral-Small-Instruct-2409-abliterated.
6
 
7
  Next, I downloaded the prompt backup from this site, and used them as a seed for some storywriting data:
 
 
 
1
  If you like this model, consider joining my discord to provide feedback: https://discord.gg/QXdn8hWSkY
2
 
3
+ This is a storywriting and roleplay model with a significant amount of self generated long context multiturn roleplay.
4
+
5
  I downloaded a bit under a thousand cards from chub.ai, and created a synthetic roleplay for each card. I batched as many turns as I could in 4k token chunks in order to maintain coherency over longer context. There was a lot of cleaning and validation between each batch, so a lot of examples were "lost," but the final output seems to be very good quality. The longest conversation is about 20k tokens, and I plan to extend this further as well as broaden the dataset with more examples. The first 4k tokens were generated with Command-R-Plus, with the remainder generated with byroneverson/Mistral-Small-Instruct-2409-abliterated.
6
 
7
  Next, I downloaded the prompt backup from this site, and used them as a seed for some storywriting data: