Frequently Asked Questions (FAQs)

Updated on November 27, 2024

These are the frequently asked questions for Vultr Serverless Inference.

Can I run inference workloads for models other than large language models on Vultr serverless inference?

Currently, Vultr Serverless Inference is optimized for running inference workloads on large language models like Mixtral 8x7B, Mistral 7B, and Meta Llama 2 70B. Support for other types of models may be added in the future, but for now, the focus is on language model inference.