How to Use the Embeddings Endpoint in Vultr Serverless Inference

Updated on November 27, 2024

Vultr Serverless Inference embeddings endpoint transforms text into vector representations using advanced AI models. This service allows users to integrate semantic understanding into their applications, facilitating tasks such as similarity search and natural language processing by converting text into structured, machine-readable formats. By leveraging this feature, you can enhance your application's ability to interpret and process natural language, enabling more sophisticated and accurate AI-driven functionalities.

Follow this guide to utilize the embeddings endpoint on your Vultr account using the Vultr Customer Portal.

  • Vultr Customer Portal
  1. Navigate to Products, click Serverless, and then click Inference.

    Serverless Inference option in products menu

  2. Click your target inference service to open its management page.

    Selection of a target serverless inference service

  3. Open the Embeddings page.

    Button to open the embeddings endpoint page

  4. Select a preferred model.

    Button to select a preferred embeddings generation model

  5. Selecting an Encoding Format.

    Button to select a preffered encoding format

  6. Provide an input.

    Field to provide input for embeddings generation

  7. Click Generate to create embeddings.

    Button to generate embeddings of the provided input

  8. Click Reset to create embeddings for a new input.

    Button to reset the input field for a new input