Update README.md
Browse files
README.md
CHANGED
|
@@ -193,13 +193,4 @@ print("content:", content)
|
|
| 193 |
messages.append(
|
| 194 |
{"role": "assistant", "content": output_text}
|
| 195 |
)
|
| 196 |
-
```
|
| 197 |
-
|
| 198 |
-
# Running the model on a CPU
|
| 199 |
-
|
| 200 |
-
This repo contains gguf versions of `sarvam-m` in both bf16 and q8 precisions. You can use the model on your local machine (without gpu) as explained [here](https://github.com/ggml-org/llama.cpp/tree/master/tools/main).
|
| 201 |
-
|
| 202 |
-
Example Command:
|
| 203 |
-
```
|
| 204 |
-
./build/bin/llama-cli -i -m /your/folder/path/sarvam-m-q8_0.gguf -c 8192 -t 16
|
| 205 |
```
|
|
|
|
| 193 |
messages.append(
|
| 194 |
{"role": "assistant", "content": output_text}
|
| 195 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 196 |
```
|