Text Embeddings Inference
Overview
Managing application via Apolo CLI
apolo app-template get text-embeddings-inference > tei.yaml# Application template configuration for: text-embeddings-inference
# Fill in the values below to configure your application.
# To use values from another app, use the following format:
# my_param:
# type: "app-instance-ref"
# instance_id: "<app-instance-id>"
# path: "<path-from-get-values-response>"
template_name: text-embeddings-inference
template_version: v25.7.0
input:
# Select the resource preset used per service replica.
preset:
# The name of the preset.
name: ''
# Enable access to your application over the internet using HTTPS.
ingress_http:
enabled: true
# Hugging Face Model Configuration.
model:
# The name of the Hugging Face model.
model_hf_name: ''
# The Hugging Face API token.
hf_token:
# Configure extra arguments to pass to the server (see TEI doc, e.g. --max-client-batch-size=1024).
server_extra_args:
- ''
# Additional environment variables to inject into the container. These will override any existing environment variables with the same name.
extra_env_vars:
-
# Specify the name of the environment variable to inject into the container.
name: ''
# Specify the value of the environment variable.
value:
Usage
References:
Last updated
Was this helpful?