Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -3,8 +3,8 @@ from huggingface_hub import hf_hub_download
|
|
| 3 |
from llama_cpp import Llama
|
| 4 |
|
| 5 |
# Define model details
|
| 6 |
-
MODEL_REPO = "
|
| 7 |
-
MODEL_FILE = "
|
| 8 |
|
| 9 |
# Download the quantized model from Hugging Face
|
| 10 |
model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
|
|
|
|
| 3 |
from llama_cpp import Llama
|
| 4 |
|
| 5 |
# Define model details
|
| 6 |
+
MODEL_REPO = "TotoB12/totob-1.5B" # You can swap this for Mistral-7B or another GGUF model
|
| 7 |
+
MODEL_FILE = "totob-1.5B.gguf" # 4-bit quantized model file
|
| 8 |
|
| 9 |
# Download the quantized model from Hugging Face
|
| 10 |
model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
|