Spaces:
Running
on
A10G
Running
on
A10G
Update app.py
Browse files
app.py
CHANGED
|
@@ -87,7 +87,9 @@ def start_vllm_service() -> bool:
|
|
| 87 |
[
|
| 88 |
"vllm", "serve", VLLM_MODEL_NAME,
|
| 89 |
"--port", str(VLLM_PORT),
|
| 90 |
-
"--host", VLLM_HOST
|
|
|
|
|
|
|
| 91 |
],
|
| 92 |
text=True
|
| 93 |
)
|
|
@@ -485,7 +487,7 @@ def parse_arguments():
|
|
| 485 |
parser.add_argument("--openai_api_source", default="we_provided", choices=["we_provided", "user_provided"], help="Source of OpenAI API key.")
|
| 486 |
return parser.parse_args()
|
| 487 |
|
| 488 |
-
|
| 489 |
def solve_problem_gradio(user_query, max_steps=10, max_time=60, llm_model_engine=None, enabled_tools=None):
|
| 490 |
"""
|
| 491 |
Wrapper function to connect the solver to Gradio.
|
|
|
|
| 87 |
[
|
| 88 |
"vllm", "serve", VLLM_MODEL_NAME,
|
| 89 |
"--port", str(VLLM_PORT),
|
| 90 |
+
"--host", VLLM_HOST,
|
| 91 |
+
"--tensor-parallel-size", "1",
|
| 92 |
+
"--gpu-memory-utilization", "0.95"
|
| 93 |
],
|
| 94 |
text=True
|
| 95 |
)
|
|
|
|
| 487 |
parser.add_argument("--openai_api_source", default="we_provided", choices=["we_provided", "user_provided"], help="Source of OpenAI API key.")
|
| 488 |
return parser.parse_args()
|
| 489 |
|
| 490 |
+
@spaces.GPU(duration=300)
|
| 491 |
def solve_problem_gradio(user_query, max_steps=10, max_time=60, llm_model_engine=None, enabled_tools=None):
|
| 492 |
"""
|
| 493 |
Wrapper function to connect the solver to Gradio.
|