Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -23,6 +23,10 @@ models_info = {
|
|
23 |
"8B": {"name": "meta-llama/Llama-3-8B", "languages": ["en"]},
|
24 |
"3.2-3B": {"name": "meta-llama/Llama-3.2-3B", "languages": ["en", "de", "fr", "it", "pt", "hi", "es", "th"]},
|
25 |
},
|
|
|
|
|
|
|
|
|
26 |
},
|
27 |
"Mistral AI": {
|
28 |
"Mistral": {
|
@@ -58,6 +62,7 @@ model_parameters = {
|
|
58 |
"google/gemma-2-2b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
59 |
"google/gemma-2-7b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
60 |
"croissantllm/CroissantLLMBase": {"temperature": 0.8, "top_p": 0.92, "top_k": 50}
|
|
|
61 |
}
|
62 |
|
63 |
# Variables globales
|
|
|
23 |
"8B": {"name": "meta-llama/Llama-3-8B", "languages": ["en"]},
|
24 |
"3.2-3B": {"name": "meta-llama/Llama-3.2-3B", "languages": ["en", "de", "fr", "it", "pt", "hi", "es", "th"]},
|
25 |
},
|
26 |
+
"Lucie": {
|
27 |
+
"Lucie (Llama 3.1)": {"name": "OpenLLM-France/Lucie-7B", "languages": ["fr"]},
|
28 |
+
|
29 |
+
},
|
30 |
},
|
31 |
"Mistral AI": {
|
32 |
"Mistral": {
|
|
|
62 |
"google/gemma-2-2b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
63 |
"google/gemma-2-7b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
64 |
"croissantllm/CroissantLLMBase": {"temperature": 0.8, "top_p": 0.92, "top_k": 50}
|
65 |
+
"OpenLLM-France/Lucie-7B": {"temperature": 0.7, "top_p": 0.9, "top_k": 50}
|
66 |
}
|
67 |
|
68 |
# Variables globales
|