Serverless Inference offers a Prompt tab in the customer portal for testing and evaluating inference workloads before full deployment.
Yes. You can evaluate inference workloads using the Prompt tab within the Vultr Serverless Inference section of the Vultr Customer Portal. This interface lets you enter sample prompts, select the desired model, and configure parameters such as maximum tokens, temperature, and top-p. By testing with representative inputs, you can observe the model’s behavior, response quality, and output characteristics while avoiding unnecessary consumption or costs when scaling up to production workloads.