-
-
-
-
-
-
Inference Providers
Active filters:
W4A16
ModelCloud/QwQ-32B-Preview-gptqmodel-4bit-vortex-v2
Text Generation
•
33B
•
Updated
•
20
•
16
ModelCloud/QwQ-32B-Preview-gptqmodel-4bit-vortex-v3
Text Generation
•
33B
•
Updated
•
18
•
14
ModelCloud/Falcon3-10B-Instruct-gptqmodel-4bit-vortex-v1
Text Generation
•
10B
•
Updated
•
7
•
3
ModelCloud/Qwen2.5-0.5B-Instruct-gptqmodel-w4a16
Text Generation
•
0.5B
•
Updated
•
10
•
1
ModelCloud/DeepSeek-R1-Distill-Qwen-7B-gptqmodel-4bit-vortex-v1
Text Generation
•
8B
•
Updated
•
24
•
5
ModelCloud/DeepSeek-R1-Distill-Qwen-7B-gptqmodel-4bit-vortex-v2
Text Generation
•
8B
•
Updated
•
233
•
7
RedHatAI/phi-4-quantized.w4a16
Text Generation
•
3B
•
Updated
•
277
•
4
RedHatAI/Mistral-Small-3.1-24B-Instruct-2503-quantized.w4a16
Image-Text-to-Text
•
5B
•
Updated
•
5.1k
•
10
RedHatAI/Llama-4-Scout-17B-16E-Instruct-quantized.w4a16
Image-Text-to-Text
•
20B
•
Updated
•
148k
•
12
pyrymikko/nomic-embed-code-W4A16-AWQ
1B
•
Updated
•
213k
tcclaviger/Minimax-M2-Thrift-GPTQ-W4A16-AMD
Text Generation
•
24B
•
Updated
•
7
•
1
TevunahAi/granite-34b-code-instruct-8k-Ultra-Hybrid
Text Generation
•
11B
•
Updated
•
8
TevunahAi/Llama-3.1-70B-Instruct-Ultra-Hybrid
Text Generation
•
22B
•
Updated
•
26