Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -165,7 +165,7 @@ def llm_chat_response(text, image_base64=None):
|
|
165 |
|
166 |
logger.info("Initializing InferenceClient...")
|
167 |
client = InferenceClient(
|
168 |
-
provider="
|
169 |
api_key=HF_TOKEN
|
170 |
)
|
171 |
|
@@ -199,7 +199,7 @@ def llm_chat_response(text, image_base64=None):
|
|
199 |
logger.info("Sending request to model...")
|
200 |
try:
|
201 |
completion = client.chat.completions.create(
|
202 |
-
model="
|
203 |
messages=messages,
|
204 |
max_tokens=500
|
205 |
)
|
|
|
165 |
|
166 |
logger.info("Initializing InferenceClient...")
|
167 |
client = InferenceClient(
|
168 |
+
provider="cerebras", # Using cerebras as in this example
|
169 |
api_key=HF_TOKEN
|
170 |
)
|
171 |
|
|
|
199 |
logger.info("Sending request to model...")
|
200 |
try:
|
201 |
completion = client.chat.completions.create(
|
202 |
+
model="meta-llama/Llama-4-Scout-17B-16E-Instruct",
|
203 |
messages=messages,
|
204 |
max_tokens=500
|
205 |
)
|