Frequently Asked Questions (FAQs)

Updated on September 23, 2024

These are the frequently asked questions for Vultr Serverless Inference.

Can I run inference workloads for models other than large language models on Vultr serverless inference?

Currently, Vultr Serverless Inference is optimized for running inference workloads on large language models like Mistral-7B-v0.3, Deepseek-R1, Llama-3.1-70B-Instruct-FP8, and Qwen2.5-32B-Instruct. Support for other types of models may be added in the future, but for now, the focus is on language model inference.

Comments

No comments yet.