-
-
-
-
-
-
Inference Providers
Active filters:
2-bit
sergiones/Mistral-quantized-2b
Text Generation
•
0.7B
•
Updated
•
10
CallMcMargin/Qwen2.5-14B-Instruct-1M-abliterated-mlx-bf16-affine-qgroup32-q2
Text Generation
•
15B
•
Updated
•
5
Infatoshi/Qwen3-Next-80B-A3B-Thinking-EXL3-2.0bpw
Text Generation
•
Updated
•
1
saviochow/GLM-4.6-mlx-2Bit
Text Generation
•
353B
•
Updated
•
76
garrison/Precog-123B-v1-mlx-2Bit
123B
•
Updated
•
6
garrison/Precog-24B-v1-mlx-2Bit
24B
•
Updated
•
4
MaziyarPanahi/VibeThinker-1.5B-GGUF
Text Generation
•
2B
•
Updated
•
1.13k
•
35
MaziyarPanahi/AesCoder-4B-GGUF
Text Generation
•
4B
•
Updated
•
57
•
1
MaziyarPanahi/MiniCPM4.1-8B-GGUF
Text Generation
•
8B
•
Updated
•
39
MaziyarPanahi/MiroThinker-v1.0-8B-GGUF
Text Generation
•
8B
•
Updated
•
61
MaziyarPanahi/MiroThinker-v1.0-30B-GGUF
Text Generation
•
31B
•
Updated
•
38
MaziyarPanahi/Apertus-8B-Instruct-2509-GGUF
Text Generation
•
8B
•
Updated
•
118
garrison/GLM-4.5-Air-REAP-82B-A12B-mlx-2Bit
Text Generation
•
82B
•
Updated
•
47
MaziyarPanahi/Qwen3-4B-Thinking-2507-GGUF
Text Generation
•
4B
•
Updated
•
63.2k
•
2
MaziyarPanahi/Qwen3-30B-A3B-Thinking-2507-GGUF
Text Generation
•
31B
•
Updated
•
60
garrison/Snowpiercer-15B-v4-mlx-2Bit
1B
•
Updated
•
9
garrison/GLM-4.5-Air-Derestricted-mlx-2Bit
Text Generation
•
107B
•
Updated
•
26
garrison/Olmo-3-32B-Think-mlx-2Bit
Text Generation
•
32B
•
Updated
•
6
ncls-p/INTELLECT-3-mlx-2Bit
Text Generation
•
107B
•
Updated
•
10
•
1
fifrio/Qwen3-1.7B-gptq-2bit-calibration-Chinese
2B
•
Updated
•
1
MaziyarPanahi/NVIDIA-Nemotron-Nano-12B-v2-GGUF
Text Generation
•
12B
•
Updated
•
63.9k
MaziyarPanahi/Olmo-3-32B-Think-GGUF
Text Generation
•
32B
•
Updated
•
20
MaziyarPanahi/Olmo-3-7B-Think-GGUF
Text Generation
•
7B
•
Updated
•
19
MaziyarPanahi/Olmo-3-7B-Instruct-GGUF
Text Generation
•
7B
•
Updated
•
16
MaziyarPanahi/NVIDIA-Nemotron-Nano-9B-v2-GGUF
Text Generation
•
9B
•
Updated
•
843
•
3
introvoyz041/Apriel-1.5-15b-Thinker-2bit-MLX-mlx-4Bit
Image-Text-to-Text
•
1B
•
Updated
•
21
MaziyarPanahi/Ministral-3-3B-Reasoning-2512-GGUF
Text Generation
•
3B
•
Updated
•
124k
•
3
MaziyarPanahi/Ministral-3-8B-Reasoning-2512-GGUF
Text Generation
•
8B
•
Updated
•
98
•
1
MaziyarPanahi/Trinity-Nano-Preview-GGUF
Text Generation
•
6B
•
Updated
•
61
•
1
MaziyarPanahi/Trinity-Mini-GGUF
Text Generation
•
26B
•
Updated
•
59.8k
•
1