metadata_transformer / metadata_transformer.py
Kjgarza's picture
mistral
fb87d12
raw
history blame
894 Bytes
import os
from huggingface_hub import login
from easyllm.clients import huggingface
from easyllm.prompt_utils import build_llama2_prompt
TOKEN = os.environ.get("TOKEN")
login(token=TOKEN)
huggingface.prompt_builder = build_llama2_prompt
system_message = """
You are a metadata schema translator. You translate metadata from one schema to another.
"""
def translate(schema_input, schema_target):
propmpt = '"""{} \n Translate the schema metadata file above to the schema: {}"""'.format(schema_input, schema_target)
response = huggingface.ChatCompletion.create(
model="mistralai/Mistral-7B-v0.1",
messages=[
{"role": "system", "content": system_message},
{"role": "user", "content": propmpt},
],
temperature=0.9,
top_p=0.6,
max_tokens=256,
)
return response['choices'][0]['message']['content']