File size: 1,344 Bytes
cb20959
 
0420ce7
 
 
28deb88
 
 
 
 
892b20e
 
db2d4cc
b32d7c4
 
 
db2d4cc
892b20e
 
 
 
 
147ea7c
fec3606
 
147ea7c
fec3606
147ea7c
fec3606
9c9fbd1
 
fec3606
892b20e
4e1b7fe
 
d2a9f77
 
 
 
 
 
4e1b7fe
 
 
1b4c23c
a76d32d
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
---
license: apache-2.0
datasets:
- stingning/ultrachat
- kaist-ai/CoT-Collection
- mesolitica/google-translate-commitpackft
- Wanfq/Explore_Instruct_Rewriting_32k
- Wanfq/Explore_Instruct_Rewriting_10k
- Wanfq/Explore_Instruct_Brainstorming_16k
- xiyuez/red-dot-design-award-product-description
---

# RWKV v4 7B world model 
finetuned with ultrachat , COT and some novel instructions data, commitpackft and so on

use full ultrachat and cot data, about 3B tokens


# Contributor
[@JL-er](https://huggingface.co/JL-er) 
[@Remixa](https://huggingface.co/Remixa)


# Design of experiment
this model lose multi-turn chat ability,cause from using whole ultrachat datasets.

so i continue tuned multi-turn datasets with 2 aspects

1.role play

2.for novel multiturn instruction

# Training details
[wandb.ai](https://wandb.ai/one-/one-rwkv-64k)

# CAses

![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/_1dJo549ldgX6q0JUwC6c.jpeg)

![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/7969wbHaJpBq2n6xvfC7C.jpeg)

# Usage

![image/png](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/cGDF6b4-x_9rcwMdl1KPp.png)

![image/png](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/hUxTVgjLBMcFqxQX9HoxL.png)