-
-
-
-
-
-
Inference Providers
Active filters: int4
jojo1899/Llama-2-13b-chat-hf-ov-int4
Text Generation
• Updated
• 4
jojo1899/Mistral-7B-Instruct-v0.2-ov-int4
Text Generation
• Updated
• 3
model-scope/glm-4-9b-chat-GPTQ-Int4
Text Generation
• 9B • Updated
• 82
• 6
ModelCloud/Mistral-Nemo-Instruct-2407-gptq-4bit
Text Generation
• 12B • Updated
• 102
• 5
ModelCloud/Meta-Llama-3.1-8B-Instruct-gptq-4bit
Text Generation
• 8B • Updated
• 161
• 4
ModelCloud/Meta-Llama-3.1-8B-gptq-4bit
Text Generation
• 8B • Updated
• 151
ModelCloud/Meta-Llama-3.1-70B-Instruct-gptq-4bit
Text Generation
• 71B • Updated
• 75
• 4
ModelCloud/Mistral-Large-Instruct-2407-gptq-4bit
Text Generation
• 123B • Updated
• 58
• 1
RedHatAI/Meta-Llama-3.1-8B-Instruct-quantized.w4a16
Text Generation
• 8B • Updated
• 34.6k
• 30
angeloc1/llama3dot1SimilarProcesses4
Text Generation
• 8B • Updated
• 4
angeloc1/llama3dot1DifferentProcesses4
Text Generation
• 8B • Updated
• 6
ModelCloud/Meta-Llama-3.1-405B-Instruct-gptq-4bit
Text Generation
• 410B • Updated
• 3
• 2
RedHatAI/Meta-Llama-3.1-70B-Instruct-quantized.w4a16
Text Generation
• 71B • Updated
• 115k
• 32
ModelCloud/EXAONE-3.0-7.8B-Instruct-gptq-4bit
8B • Updated
• 3
RedHatAI/Meta-Llama-3.1-405B-Instruct-quantized.w4a16
Text Generation
• 58B • Updated
• 94
• 12
angeloc1/llama3dot1FoodDel4v05
Text Generation
• 8B • Updated
• 1
zzzmahesh/Meta-Llama-3-8B-Instruct-quantized.w4a4
Text Generation
• 8B • Updated
• 43
• 1
ModelCloud/GRIN-MoE-gptq-4bit
42B • Updated
• 4
• 6
joshmiller656/Llama3.2-1B-AWQ-INT4
1B • Updated
• 2
Advantech-EIOT/intel_llama-3.1-8b-instruct
Updated
RedHatAI/Qwen2.5-7B-quantized.w4a16
Text Generation
• 2B • Updated
• 203
joshmiller656/Llama-3.1-Nemotron-70B-Instruct-AWQ-INT4
Text Generation
• 71B • Updated
• 8
• 3
ModelCloud/Llama-3.2-1B-Instruct-gptqmodel-4bit-vortex-v1
Text Generation
• 1B • Updated
• 66
• 2
jojo1899/llama-3_1-8b-instruct-ov-int4
ModelCloud/Llama-3.2-1B-Instruct-gptqmodel-4bit-vortex-v2
Text Generation
• 1B • Updated
• 3
ModelCloud/Llama-3.2-3B-Instruct-gptqmodel-4bit-vortex-v3
Text Generation
• 4B • Updated
• 115
• 5
tclf90/qwen2.5-72b-instruct-gptq-int4
Text Generation
• 73B • Updated
• 96
• 2
ModelCloud/Llama-3.2-1B-Instruct-gptqmodel-4bit-vortex-v2.5
Text Generation
• 1B • Updated
• 646
• 5
jojo1899/Phi-3.5-mini-instruct-ov-int4
ModelCloud/Qwen2.5-Coder-32B-Instruct-gptqmodel-4bit-vortex-v1
Text Generation
• 33B • Updated
• 128
• 16