About Quantized Models
pinned#14 opened 2 months ago
by
infgrad
The output size when deployed in GCP is 1536 instead of 1024
#23 opened 2 days ago
by
bennegeek
Is this multilingual or bilingual? english and chinese
#22 opened 13 days ago
by
taowang1993
flash attention
#21 opened about 2 months ago
by
Disassemblern
Model loading size on GPU
#20 opened 2 months ago
by
divrajnd
MRL and linear layers
1
#19 opened 2 months ago
by
bobox
Can it output sparse vector?
1
#18 opened 2 months ago
by
kk3dmax
Getting different results for the same examples provided in sample
4
#17 opened 2 months ago
by
sramakintel
Does this model only work on GPU?
1
#16 opened 2 months ago
by
xPurity
Error when loading model KeyError: 'qwen2'
1
#11 opened 2 months ago
by
longluu
Any multi-lingual variant
1
#10 opened 2 months ago
by
prophet123
Can we have it in GGUF F16/32?
2
#9 opened 3 months ago
by
qdrddr
Parameters for peak performances
3
#8 opened 3 months ago
by
cvdbdo
Difference between dunzhang/stella_en_1.5B_v5 and infgrad/stella_en_1.5B_v5?
1
#7 opened 3 months ago
by
gokturkDev
Model max_seq_length
6
#6 opened 3 months ago
by
shuyuej
Could you provide the training data list?
#5 opened 3 months ago
by
Mengyao00
Fix prompt_name typo
1
#4 opened 3 months ago
by
mber
Upload ONNX weights
1
#3 opened 3 months ago
by
Xenova