apepkuss79
commited on
Commit
•
802dcfc
1
Parent(s):
9e80949
Update README.md
Browse files
README.md
CHANGED
@@ -55,13 +55,20 @@ quantized_by: Second State Inc.
|
|
55 |
- Run as LlamaEdge service
|
56 |
|
57 |
```bash
|
58 |
-
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Llama-2-7b-chat-hf-Q5_K_M.gguf
|
|
|
|
|
|
|
|
|
59 |
```
|
60 |
|
61 |
- Run as LlamaEdge command app
|
62 |
|
63 |
```bash
|
64 |
-
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Llama-2-7b-chat-hf-Q5_K_M.gguf
|
|
|
|
|
|
|
65 |
```
|
66 |
|
67 |
## Quantized GGUF Models
|
|
|
55 |
- Run as LlamaEdge service
|
56 |
|
57 |
```bash
|
58 |
+
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Llama-2-7b-chat-hf-Q5_K_M.gguf \
|
59 |
+
llama-api-server.wasm \
|
60 |
+
--prompt-template llama-2-chat \
|
61 |
+
--ctx-size 4096
|
62 |
+
--model-name llama-2-7b-chat
|
63 |
```
|
64 |
|
65 |
- Run as LlamaEdge command app
|
66 |
|
67 |
```bash
|
68 |
+
wasmedge --dir .:. --nn-preload default:GGML:AUTO:Llama-2-7b-chat-hf-Q5_K_M.gguf \
|
69 |
+
llama-chat.wasm \
|
70 |
+
--prompt-template llama-2-chat \
|
71 |
+
--ctx-size 4096
|
72 |
```
|
73 |
|
74 |
## Quantized GGUF Models
|