-
-
-
-
-
-
Inference Providers
Active filters:
int4
ModelCloud/Mistral-Large-Instruct-2407-gptq-4bit
Text Generation
•
123B
•
Updated
•
9
•
1
RedHatAI/Meta-Llama-3.1-8B-Instruct-quantized.w4a16
Text Generation
•
8B
•
Updated
•
17.9k
•
30
angeloc1/llama3dot1SimilarProcesses4
Text Generation
•
8B
•
Updated
•
10
angeloc1/llama3dot1DifferentProcesses4
Text Generation
•
8B
•
Updated
•
14
ModelCloud/Meta-Llama-3.1-405B-Instruct-gptq-4bit
Text Generation
•
410B
•
Updated
•
8
•
2
RedHatAI/Meta-Llama-3.1-70B-Instruct-quantized.w4a16
Text Generation
•
71B
•
Updated
•
1.46k
•
32
ModelCloud/EXAONE-3.0-7.8B-Instruct-gptq-4bit
8B
•
Updated
•
7
•
3
RedHatAI/Meta-Llama-3.1-405B-Instruct-quantized.w4a16
Text Generation
•
58B
•
Updated
•
41
•
12
angeloc1/llama3dot1FoodDel4v05
Text Generation
•
8B
•
Updated
•
11
zzzmahesh/Meta-Llama-3-8B-Instruct-quantized.w4a4
Text Generation
•
8B
•
Updated
•
10
•
1
ModelCloud/GRIN-MoE-gptq-4bit
42B
•
Updated
•
7
•
6
joshmiller656/Llama3.2-1B-AWQ-INT4
1B
•
Updated
•
12
Advantech-EIOT/intel_llama-3.1-8b-instruct
Updated
•
11
RedHatAI/Qwen2.5-7B-quantized.w4a16
Text Generation
•
2B
•
Updated
•
176
joshmiller656/Llama-3.1-Nemotron-70B-Instruct-AWQ-INT4
Text Generation
•
71B
•
Updated
•
40
•
3
ModelCloud/Llama-3.2-1B-Instruct-gptqmodel-4bit-vortex-v1
Text Generation
•
1B
•
Updated
•
168
•
2
jojo1899/llama-3_1-8b-instruct-ov-int4
ModelCloud/Llama-3.2-1B-Instruct-gptqmodel-4bit-vortex-v2
Text Generation
•
1B
•
Updated
•
14
•
3
ModelCloud/Llama-3.2-3B-Instruct-gptqmodel-4bit-vortex-v3
Text Generation
•
4B
•
Updated
•
67
•
5
tclf90/qwen2.5-72b-instruct-gptq-int4
Text Generation
•
73B
•
Updated
•
76
•
2
ModelCloud/Llama-3.2-1B-Instruct-gptqmodel-4bit-vortex-v2.5
Text Generation
•
1B
•
Updated
•
284
•
5
jojo1899/Phi-3.5-mini-instruct-ov-int4
ModelCloud/Qwen2.5-Coder-32B-Instruct-gptqmodel-4bit-vortex-v1
Text Generation
•
33B
•
Updated
•
103
•
16
RedHatAI/Sparse-Llama-3.1-8B-evolcodealpaca-2of4-FP8-dynamic
Text Generation
•
8B
•
Updated
•
14
RedHatAI/Sparse-Llama-3.1-8B-evolcodealpaca-2of4-quantized.w4a16
Text Generation
•
2B
•
Updated
•
24
ModelCloud/QwQ-32B-Preview-gptqmodel-4bit-vortex-v1
Text Generation
•
33B
•
Updated
•
99
•
51
ModelCloud/QwQ-32B-Preview-gptqmodel-4bit-vortex-v2
Text Generation
•
33B
•
Updated
•
35
•
16
ModelCloud/QwQ-32B-Preview-gptqmodel-4bit-vortex-v3
Text Generation
•
33B
•
Updated
•
24
•
14
ModelCloud/Falcon3-10B-Instruct-gptqmodel-4bit-vortex-v1
Text Generation
•
10B
•
Updated
•
11
•
3
RedHatAI/Llama-3.3-70B-Instruct-quantized.w4a16
Text Generation
•
11B
•
Updated
•
1.47k
•
3