Zheng Han
traphix
·
AI & ML interests
None yet
Organizations
None yet
Error on 4 x L40s
➕
2
1
#4 opened 2 months ago
by
traphix
I got ValueError
👀
2
10
#3 opened 2 months ago
by
spow12
How to run this model via vllm?
11
#2 opened 3 months ago
by
traphix
FP8 please
👀
➕
16
8
#18 opened 3 months ago
by
aliquis-pe
vllm v0.10.2 error
❤️
1
#2 opened 3 months ago
by
traphix
VLLM compatibility?
9
#1 opened 3 months ago
by
aidendle94
Instruct or Thinking?
#1 opened 3 months ago
by
traphix
Could you share your quantization code?
1
#1 opened 3 months ago
by
traphix
Instruct or thinking?
#1 opened 3 months ago
by
traphix
Any plans to quantize Qwen3-235B-A22B-Instruct-2507?
#1 opened 3 months ago
by
traphix
Qwen3-235B-A22B-Instruct-2507, int4-w4a16 or awq? Which one has better accuracy recovery?
1
#1 opened 4 months ago
by
traphix
Does vllm 0.7.3 support this model?
1
#10 opened 9 months ago
by
traphix
Any plans to quantize Qwen3-235B-A22B-Instruct-2507?
#1 opened 4 months ago
by
traphix
Is 4 x H20 96G sufficient to run this model?
2
#2 opened 6 months ago
by
milongwong
Remove vLLM FP8 Limitation
10
#2 opened 7 months ago
by
simon-mo
Error running on A100?
2
#4 opened 7 months ago
by
traphix
Any plans for int8 quantized.w8a8?
#5 opened 7 months ago
by
traphix
How about int8 quantization?
#3 opened 7 months ago
by
traphix
How many RAM in GBs when quantizing Qwen3-235B-A22B?
#2 opened 7 months ago
by
traphix
Where are the safetensors?
👀
1
1
#1 opened 7 months ago
by
traphix