Embedding Generation

Text Embedding Models on Sentry Block

Sentry Block supports high-performance text embedding models for a wide range of language-processing applications, including search, semantic understanding, clustering, and classification.


Available Models

  • UAE-Large-V1 A versatile, general-purpose model designed to generate high-accuracy embeddings across various topics and use cases.

  • BGE Large EN v1.5 Fine-tuned specifically for English, this model offers enhanced performance for tasks requiring deeper semantic representation.


Using the Models

Through the API Endpoint

You can generate embeddings programmatically using Sentry Block’s API. Below are quick-start examples in multiple languages.

⚠️ Remember to use your Sentry API key. See the [API Reference] for more details on authentication and error handling.


cURL Example

bashCopyEditcurl -X POST "https://inference.sentryblock.com/v1/embeddings" \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $SENTRY_API_KEY" \
--data-raw '{
  "model": "WhereIsAI/UAE-Large-V1",
  "input": [ 
    "Bananas are berries, but strawberries are not, according to botanical classifications.",
    "The Eiffel Tower in Paris was originally intended to be a temporary structure."
  ]
}'

Python Example


JavaScript Example


Model Mapping

Use the following model identifiers when making API calls:

  • UAE-Large-V1WhereIsAI/UAE-Large-V1

  • BGE Large EN v1.5BAAI/bge-large-en-v1.5


Response Example

A successful embedding response will return a JSON object with the following structure:

🔍 Note: Each input string receives its own embedding. In the example above, two input sentences were provided, and two indexed embeddings were returned accordingly.

Last updated