Vultr Serverless Inference prompt endpoint allows users to send a single prompt to an AI model for generating responses. This service supports interactive and dynamic AI interactions, enabling users to obtain specific outputs based on their prompts and integrate these responses into their applications effectively. By using this feature, you can enhance your application's responsiveness and adaptability to user inputs, providing more personalized and relevant outputs.
Follow this guide to utilize the prompt endpoint on your Vultr account using the Vultr Customer Portal.
Navigate to Products, click Serverless, and then click Inference.
Click your target inference service to open its management page.
Open the Prompt page.
Select a preferred model.
Provide values of Max Tokens, Seed, Temperature, Top-k and Top-p.
Provide a prompt and click on Prompt
Click Reset to provide a new prompt.