Update README.md
Browse files
README.md
CHANGED
@@ -193,13 +193,4 @@ print("content:", content)
|
|
193 |
messages.append(
|
194 |
{"role": "assistant", "content": output_text}
|
195 |
)
|
196 |
-
```
|
197 |
-
|
198 |
-
# Running the model on a CPU
|
199 |
-
|
200 |
-
This repo contains gguf versions of `sarvam-m` in both bf16 and q8 precisions. You can use the model on your local machine (without gpu) as explained [here](https://github.com/ggml-org/llama.cpp/tree/master/tools/main).
|
201 |
-
|
202 |
-
Example Command:
|
203 |
-
```
|
204 |
-
./build/bin/llama-cli -i -m /your/folder/path/sarvam-m-q8_0.gguf -c 8192 -t 16
|
205 |
```
|
|
|
193 |
messages.append(
|
194 |
{"role": "assistant", "content": output_text}
|
195 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
196 |
```
|