Update README.md
Browse files
README.md
CHANGED
|
@@ -37,7 +37,7 @@ tags:
|
|
| 37 |
|
| 38 |
## Run with LlamaEdge
|
| 39 |
|
| 40 |
-
- LlamaEdge version: [v0.
|
| 41 |
|
| 42 |
- Prompt template
|
| 43 |
|
|
@@ -126,4 +126,4 @@ tags:
|
|
| 126 |
| [Llama-3.2-3B-Instruct-Q8_0.gguf](https://huggingface.co/second-state/Llama-3.2-3B-Instruct-GGUF/blob/main/Llama-3.2-3B-Instruct-Q8_0.gguf) | Q8_0 | 8 | 1.32 GB| very large, extremely low quality loss - not recommended |
|
| 127 |
| [Llama-3.2-3B-Instruct-f16.gguf](https://huggingface.co/second-state/Llama-3.2-3B-Instruct-GGUF/blob/main/Llama-3.2-3B-Instruct-f16.gguf) | f16 | 16 | 2.48 GB| |
|
| 128 |
|
| 129 |
-
*Quantized with llama.cpp
|
|
|
|
| 37 |
|
| 38 |
## Run with LlamaEdge
|
| 39 |
|
| 40 |
+
- LlamaEdge version: [v0.16.5](https://github.com/LlamaEdge/LlamaEdge/releases/tag/0.16.5) and above
|
| 41 |
|
| 42 |
- Prompt template
|
| 43 |
|
|
|
|
| 126 |
| [Llama-3.2-3B-Instruct-Q8_0.gguf](https://huggingface.co/second-state/Llama-3.2-3B-Instruct-GGUF/blob/main/Llama-3.2-3B-Instruct-Q8_0.gguf) | Q8_0 | 8 | 1.32 GB| very large, extremely low quality loss - not recommended |
|
| 127 |
| [Llama-3.2-3B-Instruct-f16.gguf](https://huggingface.co/second-state/Llama-3.2-3B-Instruct-GGUF/blob/main/Llama-3.2-3B-Instruct-f16.gguf) | f16 | 16 | 2.48 GB| |
|
| 128 |
|
| 129 |
+
*Quantized with llama.cpp b4466*
|