File size: 1,058 Bytes
2e888e9
 
 
f5aec1b
2e888e9
 
 
 
 
 
 
 
 
 
 
bf96358
2e888e9
 
 
 
 
bf96358
2e888e9
 
 
 
 
 
 
f5aec1b
2e888e9
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
# LiteLLM Proxy config for Replicate
# Save as config.yaml and run:  litellm --config ./config.yaml --debug

model_list:
  # Meta Llama 3 Instruct (8B) via Replicate
  - model_name: openai/gpt-5-nano
    litellm_params:
      model: replicate/openai/gpt-5-nano
      api_key: os.environ/REPLICATE_API_KEY
      # optional common params
      temperature: 0.7
      max_tokens: 1024

  # Llama-2-70B chat with a pinned version ID (replace the example hash with the current one you use)
  - model_name: openai/gpt-5-mini
    litellm_params:
      model: replicate/openai/gpt-5-mini
      api_key: os.environ/REPLICATE_API_KEY

  # Call a Replicate *deployment* you own (replace owner/name)
  - model_name: openai/gpt-5
    litellm_params:
      model: replicate/openai/gpt-5
      api_key: os.environ/REPLICATE_API_KEY

# Optional: proxy/router tuning
router_settings:
  timeout: 600           # Replicate can cold-boot, so give it time
  num_retries: 1

# (Optional) enable simple logging callbacks, etc.
# litellm_settings:
#   success_callback: ["generic"]