GPT OSS
Overview
Managing application via Apolo CLI
apolo app-template get gpt-inference > gpt-oss.yaml# Application template configuration for: gpt-inference
# Fill in the values below to configure your application.
# To use values from another app, use the following format:
# my_param:
# type: "app-instance-ref"
# instance_id: "<app-instance-id>"
# path: "<path-from-get-values-response>"
# yaml-language-server: $schema=https://api.dev.apolo.us/apis/apps/v2/templates/gpt-inference/v25.7.1/schema
template_name: gpt-inference
template_version: v25.7.1
input:
# Apolo Secret Configuration.
hf_token:
key: ''
# Enable or disable autoscaling for the LLM.
autoscaling_enabled: false
size: gpt-oss-20b
llm_class: gpt-oss
Usage
References:
Last updated
Was this helpful?