-
-
-
-
-
-
Inference Providers
Active filters:
modelopt
nvidia/Qwen3-Next-80B-A3B-Instruct-NVFP4
Text Generation
•
Updated
•
12k
•
8
nvidia/Qwen3-Next-80B-A3B-Thinking-NVFP4
Text Generation
•
Updated
•
10.8k
•
14
Elias-Schwegler/IQuest-Coder-V1-40B-Loop-Instruct-NVFP4
Text Generation
•
20B
•
Updated
•
369
•
5
nvidia/Qwen2.5-VL-7B-Instruct-NVFP4
Text Generation
•
5B
•
Updated
•
2.3k
•
10
Text Generation
•
5B
•
Updated
•
3.96k
•
10
Text Generation
•
17B
•
Updated
•
5.32k
•
5
Tengyunw/MiniMax-M2.1-NVFP4
Text Generation
•
115B
•
Updated
•
85
•
4
nvidia/Qwen3-30B-A3B-NVFP4
Text Generation
•
16B
•
Updated
•
27.7k
•
19
nvidia/Llama-3.1-8B-Instruct-NVFP4
5B
•
Updated
•
18k
•
6
Text Generation
•
177B
•
Updated
•
4.06k
•
7
lukealonso/MiniMax-M2.1-NVFP4
115B
•
Updated
•
7.14k
•
16
nvidia/Llama-4-Scout-17B-16E-Instruct-NVFP4
56B
•
Updated
•
28.9k
•
18
nvidia/Llama-4-Maverick-17B-128E-Instruct-FP8
402B
•
Updated
•
142
•
11
nvidia/Llama-4-Scout-17B-16E-Instruct-FP8
109B
•
Updated
•
225k
•
9
ishan24/test_modelopt_quant
nvidia/Llama-4-Maverick-17B-128E-Eagle3
Updated
•
63
•
9
jiangchengchengNLP/L3.3-MS-Nevoria-70b-FP8
Text Generation
•
71B
•
Updated
•
2
NVFP4/Qwen3-30B-A3B-Instruct-2507-FP4
Text Generation
•
16B
•
Updated
•
668
•
11
NVFP4/Qwen3-Coder-30B-A3B-Instruct-FP4
Text Generation
•
16B
•
Updated
•
1.36k
•
3
gesong2077/Qwen3-32B-NVFP4
19B
•
Updated
•
4
•
1
54B
•
Updated
•
8
nvidia/Phi-4-multimodal-instruct-NVFP4
4B
•
Updated
•
1.14k
•
6
nvidia/Phi-4-multimodal-instruct-FP8
6B
•
Updated
•
1.82k
•
4
nvidia/Phi-4-reasoning-plus-FP8
15B
•
Updated
•
497
•
3
nvidia/Phi-4-reasoning-plus-NVFP4
8B
•
Updated
•
531
•
6
Text Generation
•
8B
•
Updated
•
4.19k
•
3
Text Generation
•
8B
•
Updated
•
3k
•
5
Text Generation
•
15B
•
Updated
•
2.09k
•
2
nvidia/Qwen2.5-VL-7B-Instruct-FP8
Text Generation
•
8B
•
Updated
•
435
•
7
nuphoto-ian/Qwen3-8B-QAT-NVFP4
5B
•
Updated