File size: 1,058 Bytes
2e888e9 f5aec1b 2e888e9 bf96358 2e888e9 bf96358 2e888e9 f5aec1b 2e888e9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 |
# LiteLLM Proxy config for Replicate
# Save as config.yaml and run: litellm --config ./config.yaml --debug
model_list:
# Meta Llama 3 Instruct (8B) via Replicate
- model_name: openai/gpt-5-nano
litellm_params:
model: replicate/openai/gpt-5-nano
api_key: os.environ/REPLICATE_API_KEY
# optional common params
temperature: 0.7
max_tokens: 1024
# Llama-2-70B chat with a pinned version ID (replace the example hash with the current one you use)
- model_name: openai/gpt-5-mini
litellm_params:
model: replicate/openai/gpt-5-mini
api_key: os.environ/REPLICATE_API_KEY
# Call a Replicate *deployment* you own (replace owner/name)
- model_name: openai/gpt-5
litellm_params:
model: replicate/openai/gpt-5
api_key: os.environ/REPLICATE_API_KEY
# Optional: proxy/router tuning
router_settings:
timeout: 600 # Replicate can cold-boot, so give it time
num_retries: 1
# (Optional) enable simple logging callbacks, etc.
# litellm_settings:
# success_callback: ["generic"]
|