Llama-4-Maverick-17B-128E-Instruct

Open Advisor →
401.6B

VRAM Requirements

VRAM requirements for Llama-4-Maverick-17B-128E-Instruct at different quantization levels
QuantizationVRAM Required
FP16897.6 GB
Q8448.8 GB
Q6336.6 GB
Q4224.4 GB

Compatible GPUs

No compatible GPUs in database.

Llama-4-Maverick-17B-128E-Instruct FAQ