""" Central configuration for models and providers This file centralizes all configurations related to models and providers used in the application. """ # Definition of preferred providers, used in get_available_model_provider.py # PREFERRED_PROVIDERS = ["sambanova", "novita"] # increased number of providers PREFERRED_PROVIDERS = ["fireworks-ai", "cerebras", "sambanova", "together", "nebius", "novita", "hyperbolic", "cohere", "hf-inference"] # Default models to evaluate for evaluation DEFAULT_EVALUATION_MODELS = [ "Qwen/Qwen3-30B-A3B", "Qwen/Qwen3-235B-A22B", "meta-llama/Llama-4-Maverick-17B-128E-Instruct", "meta-llama/Llama-4-Scout-17B-16E-Instruct", "deepseek-ai/DeepSeek-V3-0324", "google/gemma-3-27b-it", "mistralai/Mistral-Small-3.1-24B-Instruct-2503" ] # DEFAULT_EVALUATION_MODELS = [ # "Qwen/QwQ-32B", # "Qwen/Qwen2.5-72B-Instruct", # "Qwen/Qwen2.5-32B-Instruct", # "meta-llama/Llama-3.1-8B-Instruct", # "meta-llama/Llama-3.3-70B-Instruct", # "deepseek-ai/DeepSeek-R1-Distill-Llama-70B", # "mistralai/Mistral-Small-24B-Instruct-2501", # ] # Modèles alternatifs à utiliser si le modèle par défaut n'est pas disponible ALTERNATIVE_BENCHMARK_MODELS = [ "deepseek-ai/DeepSeek-R1-Distill-Llama-70B", "meta-llama/Llama-3.3-70B-Instruct", "meta-llama/Llama-3.1-8B-Instruct", "Qwen/Qwen2.5-72B-Instruct", "mistralai/Mistral-Small-24B-Instruct-2501", # Modèles open-source qui peuvent fonctionner sans authentification "HuggingFaceH4/zephyr-7b-beta", "TinyLlama/TinyLlama-1.1B-Chat-v1.0", "microsoft/phi-2", ] # Required model for create_bench_config_file.py (only one default model) DEFAULT_BENCHMARK_MODEL = "Qwen/Qwen3-30B-A3B" # DEFAULT_BENCHMARK_MODEL = "Qwen/Qwen2.5-32B-Instruct" # Models by roles for benchmark configuration # All roles use the default model except chunking BENCHMARK_MODEL_ROLES = { "ingestion": [DEFAULT_BENCHMARK_MODEL], "summarization": [DEFAULT_BENCHMARK_MODEL], "chunking": ["intfloat/multilingual-e5-large-instruct"], "single_shot_question_generation": [DEFAULT_BENCHMARK_MODEL], "multi_hop_question_generation": [DEFAULT_BENCHMARK_MODEL], } # Default evaluation timeout (in seconds) DEFAULT_EVALUATION_TIMEOUT = 60.0 # Default benchmark timeout (in seconds) DEFAULT_BENCHMARK_TIMEOUT = 300.0