Inference Providers
Active filters: cuda
Text Generation
• 8B • Updated • 13.8k
• 290
prism-ml/Bonsai-1.7B-gguf
Text Generation
• 2B • Updated • 2.6k
• 23
Text Generation
• 4B • Updated • 1.96k
• 12
Multilingual-Multimodal-NLP/IndustrialCoder
Text Generation
• 32B • Updated • 1.52k
• 50
mradermacher/IndustrialCoder-i1-GGUF
32B • Updated • 2.29k
• 4
Multilingual-Multimodal-NLP/IndustrialCoder-Thinking
Text Generation
• 32B • Updated • 52
• 3
ussoewwin/Flash-Attention-2_for_Windows
Multilingual-Multimodal-NLP/IndustrialCoder-Base
Text Generation
• 32B • Updated • 81
• 2
mradermacher/IndustrialCoder-Thinking-GGUF
32B • Updated • 660
• 1
Text Generation
• Updated • 6
• 23
CalderaAI/13B-Ouroboros-GPTQ4bit-128g-CUDA
Text Generation
• Updated • 6
marcorez8/llama-cpp-python-windows-blackwell-cuda
ValiantLabs/Qwen3-8B-ShiningValiant3
Text Generation
• 8B • Updated • 10
• 3
mradermacher/Qwen3-8B-ShiningValiant3-GGUF
8B • Updated • 923
• 2
mradermacher/Qwen3-8B-ShiningValiant3-i1-GGUF
8B • Updated • 383
• 2
ValiantLabs/Qwen3-1.7B-ShiningValiant3
Text Generation
• 2B • Updated • 63
• 5
mradermacher/Qwen3-1.7B-ShiningValiant3-GGUF
2B • Updated • 82
mradermacher/Qwen3-1.7B-ShiningValiant3-i1-GGUF
2B • Updated • 146
ValiantLabs/Qwen3-4B-ShiningValiant3
Text Generation
• 4B • Updated • 13
• 7
sequelbox/Qwen3-8B-PlumEsper
Text Generation
• 8B • Updated • 4
sequelbox/Qwen3-4B-PlumEsper
Text Generation
• 4B • Updated • 3
mradermacher/Qwen3-Shining-Lucy-CODER-3.5B-Brainstorm20x-e32-GGUF
3B • Updated • 178
• 2
mradermacher/Qwen3-Shining-Lucy-CODER-2.4B-mix2-GGUF
2B • Updated • 95
mradermacher/Qwen3-Shining-Lucy-CODER-2.4B-GGUF
2B • Updated • 53
mradermacher/Qwen3-Shining-Lucy-CODER-2.4B-mix2-i1-GGUF
2B • Updated • 62
• 1
mradermacher/Qwen3-Shining-Lucy-CODER-2.4B-i1-GGUF
2B • Updated • 141
mradermacher/Qwen3-Shining-Lucy-CODER-3.5B-Brainstorm20x-e32-i1-GGUF
3B • Updated • 326
• 1
mradermacher/Qwen3-Shining-Valiant-Instruct-Fast-CODER-Reasoning-2.4B-GGUF
2B • Updated • 62
mradermacher/Qwen3-Shining-Valiant-Instruct-Fast-CODER-Reasoning-2.4B-i1-GGUF
2B • Updated • 65