Products
Features
Marketplace
Pricing
Partners
Company
Contact Sales
Vultr Docs
Latest Content
Documentation
Inference Cookbook
Model Library
Search...
K
Inference Cookbook
Cookbook for CUDA
Benchmarks
Export PDF
Focus Mode
Text Size
16px
A
A
Benchmarks
Updated on 12 March, 2026
Benchmark methodology and performance results for LLM inference workloads on NVIDIA HGX B200 GPUs.
Methodology
How we measure LLM inference performance on NVIDIA HGX B200 GPUs.
Results Overview
Consolidated benchmark results for all five models on NVIDIA HGX B200 GPUs.