File size: 878 Bytes
3786127
 
 
 
891ea24
 
 
 
 
3786127
 
 
 
 
71aed78
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
# --- Core LLM backend (switch here only) ---

# Files written by TTS go here (and get cleaned by our code)
VOICE_AUDIO_DIR=runtime/audio

# Optional cloud keys (not needed for local MVP)
# OPENAI_API_KEY=
# GROQ_API_KEY=

API_BACKEND=sim   # sim | mock | http

# Switch between old rule router and LLM-driven flow (no code changes needed)
ROUTER_MODE=llm

SAY_VOICE=Samantha


# Which backend to use
BACKEND_LLM=llamacpp

# Path where the GGUF model will be saved after download
LLAMACPP_MODEL_PATH=models/qwen2.5-1.5b-instruct-q4_k_m.gguf

# HF repo to download the model from
HF_MODEL_REPO=Qwen/Qwen2.5-1.5B-Instruct-GGUF
HF_MODEL_FILE=qwen2.5-1.5b-instruct-q4_k_m.gguf

# llama.cpp runtime knobs
N_CTX=4096
N_THREADS=4
N_GPU_LAYERS=0

# Audio + misc
ASR_DEVICE=cpu
TTS_ENGINE=piper
PIPER_MODEL=models/piper/en_US-amy-medium.onnx
PIPER_BIN=piper

IS_HF_SPACE=true
DEBUG=false