Spaces:
Running
Running
add logging for llm output
Browse files
main.py
CHANGED
|
@@ -365,9 +365,13 @@ async def search_assistant(query: SearchQueryModel, api_key: str = Depends(verif
|
|
| 365 |
raise HTTPException(status_code=500, detail="Failed to fetch search data")
|
| 366 |
|
| 367 |
def process_response():
|
|
|
|
|
|
|
| 368 |
for content in chat_with_llama_stream(messages, model=query.model_id):
|
|
|
|
| 369 |
yield content
|
| 370 |
logger.info(f"Completed search assistant response for query: {query.query}")
|
|
|
|
| 371 |
|
| 372 |
return StreamingResponse(process_response(), media_type="text/event-stream")
|
| 373 |
|
|
|
|
| 365 |
raise HTTPException(status_code=500, detail="Failed to fetch search data")
|
| 366 |
|
| 367 |
def process_response():
|
| 368 |
+
logger.info(f"Generating response using LLM: {messages}")
|
| 369 |
+
full_response = ""
|
| 370 |
for content in chat_with_llama_stream(messages, model=query.model_id):
|
| 371 |
+
full_response+=content
|
| 372 |
yield content
|
| 373 |
logger.info(f"Completed search assistant response for query: {query.query}")
|
| 374 |
+
logger.info(f"LLM Response: {full_response}")
|
| 375 |
|
| 376 |
return StreamingResponse(process_response(), media_type="text/event-stream")
|
| 377 |
|