Spaces:
Sleeping
Sleeping
Tom
Claude
commited on
Commit
·
5a32df9
1
Parent(s):
c19585a
Revert to Mistral-7B-Instruct-v0.3 for local development
Browse filesReverted from SmolLM3-3B back to Mistral-7B-Instruct-v0.3 which works
correctly for local development. SmolLM3 had issues with JSON parsing
that caused tool calls to be displayed instead of executed.
Mistral-7B works locally but is not available on HuggingFace Inference
API free tier, so production deployment will require account upgrade
or alternative deployment method.
🤖 Generated with [Claude Code](https://claude.com/claude-code)
Co-Authored-By: Claude <[email protected]>
app.py
CHANGED
|
@@ -126,10 +126,9 @@ async def query_mistral_async(message: str, language: str = "en") -> dict:
|
|
| 126 |
{"role": "user", "content": f"Language: {language}\nQuestion: {message}"}
|
| 127 |
]
|
| 128 |
|
| 129 |
-
# Call
|
| 130 |
-
# SmolLM3 is one of the few models available on hf-inference for chat
|
| 131 |
response = client.chat_completion(
|
| 132 |
-
model="
|
| 133 |
messages=messages,
|
| 134 |
max_tokens=500,
|
| 135 |
temperature=0.3
|
|
|
|
| 126 |
{"role": "user", "content": f"Language: {language}\nQuestion: {message}"}
|
| 127 |
]
|
| 128 |
|
| 129 |
+
# Call Mistral via HuggingFace Inference API
|
|
|
|
| 130 |
response = client.chat_completion(
|
| 131 |
+
model="mistralai/Mistral-7B-Instruct-v0.3",
|
| 132 |
messages=messages,
|
| 133 |
max_tokens=500,
|
| 134 |
temperature=0.3
|