Vultr DocsLatest Content


Frequently Asked Questions (FAQs) About Vultr Serverless Inference

Updated on 10 September, 2025

Frequently asked questions and answers about Vultrs products, services, and platform features.


These are the frequently asked questions for Vultr Serverless Inference.

Can I run inference workloads for models other than large language models on Vultr serverless inference?

Currently, Vultr Serverless Inference is optimized for running inference workloads on large language models like Mistral-7B-v0.3, Deepseek-R1, Llama-3.1-70B-Instruct-FP8, and Qwen2.5-32B-Instruct. Support for other types of models may be added in the future, but for now, the focus is on language model inference.

Comments

No comments yet.