Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -54,7 +54,6 @@ def process_files(files, model_type):
|
|
54 |
def chat_submit_func(message, files, chat_history, model, temperature, top_p, max_tokens, api_key):
|
55 |
print(model)
|
56 |
client = genai.Client(api_key=api_key)
|
57 |
-
gen_model = client.models.get(model=model)
|
58 |
|
59 |
# Prepare inputs
|
60 |
if model_types[model] == "text" and files:
|
@@ -73,7 +72,7 @@ def chat_submit_func(message, files, chat_history, model, temperature, top_p, ma
|
|
73 |
}
|
74 |
|
75 |
try:
|
76 |
-
response =
|
77 |
response_text = ""
|
78 |
response_images = []
|
79 |
|
@@ -107,7 +106,6 @@ def chat_submit_func(message, files, chat_history, model, temperature, top_p, ma
|
|
107 |
def single_submit_func(prompt, files, model, temperature, top_p, max_tokens, api_key):
|
108 |
print(model)
|
109 |
client = genai.Client(api_key=api_key)
|
110 |
-
gen_model = client.models.get(model=model)
|
111 |
|
112 |
# Prepare inputs
|
113 |
if model_types[model] == "text" and files:
|
@@ -127,7 +125,7 @@ def single_submit_func(prompt, files, model, temperature, top_p, max_tokens, api
|
|
127 |
}
|
128 |
|
129 |
try:
|
130 |
-
response =
|
131 |
response_text = warning
|
132 |
response_images = []
|
133 |
|
|
|
54 |
def chat_submit_func(message, files, chat_history, model, temperature, top_p, max_tokens, api_key):
|
55 |
print(model)
|
56 |
client = genai.Client(api_key=api_key)
|
|
|
57 |
|
58 |
# Prepare inputs
|
59 |
if model_types[model] == "text" and files:
|
|
|
72 |
}
|
73 |
|
74 |
try:
|
75 |
+
response = client.models.generate_content(inputs, model=model, generation_config=generation_config)
|
76 |
response_text = ""
|
77 |
response_images = []
|
78 |
|
|
|
106 |
def single_submit_func(prompt, files, model, temperature, top_p, max_tokens, api_key):
|
107 |
print(model)
|
108 |
client = genai.Client(api_key=api_key)
|
|
|
109 |
|
110 |
# Prepare inputs
|
111 |
if model_types[model] == "text" and files:
|
|
|
125 |
}
|
126 |
|
127 |
try:
|
128 |
+
response = client.models.generate_content(inputs, model=model, generation_config=generation_config)
|
129 |
response_text = warning
|
130 |
response_images = []
|
131 |
|