debisoft commited on
Commit
63a2ee7
·
1 Parent(s): 921294b

Switch to Databricks Model Serving

Browse files
Files changed (1) hide show
  1. app.py +18 -10
app.py CHANGED
@@ -1,4 +1,6 @@
1
- import openai
 
 
2
  import os
3
  import gradio as gr
4
  import json
@@ -6,16 +8,22 @@ from dotenv import load_dotenv, find_dotenv
6
  _ = load_dotenv(find_dotenv())
7
 
8
 
9
- openai.api_key = os.getenv('OPENAI_API_KEY')
 
10
 
11
- def get_completion(prompt, model="gpt-3.5-turbo"):
12
- messages = [{"role": "user", "content": prompt}]
13
- response = openai.ChatCompletion.create(
14
- model=model,
15
- messages=messages,
16
- temperature=0, # this is the degree of randomness of the model's output
17
- )
18
- return response.choices[0].message["content"]
 
 
 
 
 
19
 
20
  def greet(input):
21
  prompt = f"""
 
1
+ import numpy as np
2
+ import pandas as pd
3
+ import requests
4
  import os
5
  import gradio as gr
6
  import json
 
8
  _ = load_dotenv(find_dotenv())
9
 
10
 
11
+ databricks_token = os.getenv('DATABRICKS_TOKEN')
12
+ model_uri = "https://dbc-eb788f31-6c73.cloud.databricks.com/serving-endpoints/Mpt-7b-tester/invocations"
13
 
14
+ def score_model(model_uri, databricks_token, data):
15
+ headers = {
16
+ "Authorization": f"Bearer {databricks_token}",
17
+ "Content-Type": "application/json",
18
+ }
19
+ data_json = json.dumps({"dataframe_split": {"index": [0], "columns": ["prompt", "temperature", "max_tokens"], "data": [[data, 0.5, 500]]}})
20
+ response = requests.request(method='POST', headers=headers, url=model_uri, json=data_json)
21
+ if response.status_code != 200:
22
+ raise Exception(f"Request failed with status {response.status_code}, {response.text}")
23
+ return response.json()
24
+
25
+ def get_completion(prompt):
26
+ return score_model(model_uri, databricks_token, prompt)
27
 
28
  def greet(input):
29
  prompt = f"""