A10 vs L40

Open Advisor →

Side-by-side GPU comparison: specs, memory, compute performance, and live cloud pricing.

Verdict

L40 has 24 GB more VRAM, making it better suited for large models and long context windows. For compute-bound workloads like training, L40 delivers 2.9× higher FP16 throughput. At $0.99/hr vs $1.29/hr, L40 is the more cost-efficient choice for inference. L40 supports a broader range of models (624 vs 566 from this catalog), giving more flexibility.

Specifications

A10L40
VRAM24 GB48 GB
VRAM TypeGDDR6GDDR6
Memory Bandwidth0.6 TB/s0.9 TB/s
FP16 Performance63 TFLOPS181 TFLOPS
ManufacturerNVIDIANVIDIA
FP8 SupportNoYes
FP4 SupportNoNo

Price / Performance

Based on cheapest single-GPU on-demand pricing. Lower $/TFLOP = better compute value; lower $/GB = better memory value.

A10L40
$/hr (cheapest)$1.29$0.99✓ best
$/TFLOP (compute value)$0.0206$0.0055✓ best
$/GB VRAM (memory value)$0.0537$0.0206✓ best

Cloud Pricing

Cheapest on-demand price per provider (single GPU).

A10

ProviderOn-demandSpotRent
Lambda$1.29/hr
Microsoft Azure$3.20/hr$0.80/hr

L40

ProviderOn-demandSpotRent
RunPod$0.99/hr$0.50/hrRent

Model Compatibility

Models from the catalog that fit on each GPU, grouped by required precision.

A10 (566 models)

L40 (624 models)

You might also compare…

Pricing data refreshed hourly · Last updated April 11, 2026 · Browse all comparisons