Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -10,12 +10,6 @@ app = Flask(__name__)
|
|
10 |
token = os.environ.get("TOKEN")
|
11 |
genai.configure(api_key=token)
|
12 |
|
13 |
-
generation_config = {
|
14 |
-
"temperature": 1,
|
15 |
-
"top_p": 0.95,
|
16 |
-
"top_k": 64,
|
17 |
-
"max_output_tokens": 8192,
|
18 |
-
}
|
19 |
|
20 |
safety_settings = [
|
21 |
{"category": "HARM_CATEGORY_HARASSMENT", "threshold": "BLOCK_NONE"},
|
@@ -27,8 +21,7 @@ safety_settings = [
|
|
27 |
mm = """ resous cet exercice. tu répondras en détaillant au maximum ton procédé de calcul. réponse attendue uniquement en Latex"""
|
28 |
|
29 |
model = genai.GenerativeModel(
|
30 |
-
model_name="gemini-
|
31 |
-
generation_config=generation_config,
|
32 |
safety_settings=safety_settings
|
33 |
)
|
34 |
|
|
|
10 |
token = os.environ.get("TOKEN")
|
11 |
genai.configure(api_key=token)
|
12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
safety_settings = [
|
15 |
{"category": "HARM_CATEGORY_HARASSMENT", "threshold": "BLOCK_NONE"},
|
|
|
21 |
mm = """ resous cet exercice. tu répondras en détaillant au maximum ton procédé de calcul. réponse attendue uniquement en Latex"""
|
22 |
|
23 |
model = genai.GenerativeModel(
|
24 |
+
model_name="gemini-exp-1206",
|
|
|
25 |
safety_settings=safety_settings
|
26 |
)
|
27 |
|