# LiteLLM Proxy config for Replicate # Save as config.yaml and run: litellm --config ./config.yaml --debug model_list: # Meta Llama 3 Instruct (8B) via Replicate - model_name: openai/gpt-5-nano litellm_params: model: replicate/openai/gpt-5-nano api_key: os.environ/REPLICATE_API_KEY # optional common params temperature: 0.7 max_tokens: 1024 # Llama-2-70B chat with a pinned version ID (replace the example hash with the current one you use) - model_name: openai/gpt-5-mini litellm_params: model: replicate/openai/gpt-5-mini api_key: os.environ/REPLICATE_API_KEY # Call a Replicate *deployment* you own (replace owner/name) - model_name: openai/gpt-5 litellm_params: model: replicate/openai/gpt-5 api_key: os.environ/REPLICATE_API_KEY # Optional: proxy/router tuning router_settings: timeout: 600 # Replicate can cold-boot, so give it time num_retries: 1 # (Optional) enable simple logging callbacks, etc. # litellm_settings: # success_callback: ["generic"]