-
-
-
-
-
-
Inference Providers
Active filters:
w4a16
RedHatAI/gemma-3-12b-it-quantized.w4a16
Image-Text-to-Text
•
4B
•
Updated
•
5.42k
•
2
RedHatAI/gemma-3-27b-it-quantized.w4a16
Image-Text-to-Text
•
7B
•
Updated
•
27.4k
•
12
abdou-u/MNLP_M3_quantized_dpo_mcqa_model
Multiple Choice
•
0.6B
•
Updated
•
1
RedHatAI/gemma-3-1b-it-quantized.w4a16
Text Generation
•
0.7B
•
Updated
•
150
adamrb/mpt-30b-chat-w4a16-gptq
30B
•
Updated
RedHatAI/SmolLM3-3B-quantized.w4a16
0.9B
•
Updated
•
1
ramblingpolymath/Qwen3-30B-A3B-Instruct-2507-W4A16
Text Generation
•
5B
•
Updated
•
23
ramblingpolymath/Qwen3-Coder-30B-A3B-Instruct-W4A16
Text Generation
•
5B
•
Updated
•
133
•
2
ramblingpolymath/Qwen3-30B-A3B-Thinking-2507-W4A16
Text Generation
•
5B
•
Updated
•
25
•
3
TheHouseOfTheDude/Behemoth-R1-123B-v2_Compressed-Tensors
Text Generation
•
Updated
•
1
TheHouseOfTheDude/Behemoth-X-123B-v2_Compressed-Tensors
Text Generation
•
Updated
•
3
TheHouseOfTheDude/GLM-Steam-106B-A12B-v1_Compressed-Tensors
Text Generation
•
Updated
TheHouseOfTheDude/L3.3-Animus-V10.0_Compressed-Tensors
Text Generation
•
Updated
TheHouseOfTheDude/Behemoth-ReduX-123B-v1_Compressed-Tensors
Text Generation
•
Updated
TheHouseOfTheDude/Qwen3-Next-80B-A3B-Instruct_Compressed-Tensors
Text Generation
•
Updated
•
9
TheHouseOfTheDude/Fallen-Command-A-111B-v1_Compresses-Tensors
Text Generation
•
Updated
TheHouseOfTheDude/Behemoth-ReduX-123B-v1.1_Compressed-Tensors
Text Generation
•
Updated
TheHouseOfTheDude/L3.3-70B-Animus-V12.0_Compressed-Tensors
Text Generation
•
Updated
•
1
TheHouseOfTheDude/Behemoth-X-123B-v2.1_Compressed-Tensors
Text Generation
•
Updated
•
1
ModelCloud/GLM-4.6-GPTQMODEL-W4A16-v1
Text Generation
•
357B
•
Updated
•
3
ModelCloud/GLM-4.6-GPTQMODEL-W4A16-v2
Text Generation
•
357B
•
Updated
•
3
•
1
ModelCloud/GLM-4.6-REAP-268B-A32B-GPTQMODEL-W4A16
Text Generation
•
269B
•
Updated
•
47
•
2
ModelCloud/MiniMax-M2-GPTQMODEL-W4A16
Text Generation
•
229B
•
Updated
•
58
•
3
tcclaviger/Qwen3-42B-A3B-2507-Thinking-TOTAL-RECALL-v2-Medium-MASTER-CODER
Text Generation
•
6B
•
Updated
•
14
ModelCloud/Marin-32B-Base-GPTQMODEL-W4A16
Text Generation
•
33B
•
Updated
•
4
•
1
ModelCloud/Marin-32B-Base-GPTQMODEL-AWQ-W4A16
Text Generation
•
33B
•
Updated
•
5
•
1
tcclaviger/Qwen3-Coder-42B-A3B-Instruct-TOTAL-RECALL-MASTER-CODER-M-512k-ctx-W4A16
Text Generation
•
6B
•
Updated
•
299
•
2
TheHouseOfTheDude/Legion-V2.1-LLaMa-70B_CompressedTensors
Text Generation
•
Updated
ModelCloud/Granite-4.0-H-1B-GPTQMODEL-W4A16
Text Generation
•
1B
•
Updated
•
3
ModelCloud/Granite-4.0-H-350M-GPTQMODEL-W4A16
Text Generation
•
0.3B
•
Updated
•
22