mohit-sarvam commited on
Commit
c1ecd29
·
verified ·
1 Parent(s): 1431667

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -9
README.md CHANGED
@@ -193,13 +193,4 @@ print("content:", content)
193
  messages.append(
194
  {"role": "assistant", "content": output_text}
195
  )
196
- ```
197
-
198
- # Running the model on a CPU
199
-
200
- This repo contains gguf versions of `sarvam-m` in both bf16 and q8 precisions. You can use the model on your local machine (without gpu) as explained [here](https://github.com/ggml-org/llama.cpp/tree/master/tools/main).
201
-
202
- Example Command:
203
- ```
204
- ./build/bin/llama-cli -i -m /your/folder/path/sarvam-m-q8_0.gguf -c 8192 -t 16
205
  ```
 
193
  messages.append(
194
  {"role": "assistant", "content": output_text}
195
  )
 
 
 
 
 
 
 
 
 
196
  ```