| import os | |
| import argparse | |
| import json | |
| from elm.infer_elm import generate_elm_responses | |
| parser = argparse.ArgumentParser(description='run prompts with elm model.') | |
| parser.add_argument('elm_model_path', help='Path to the elm_model_path') | |
| def get_prompt_config_file(elm_model_path): | |
| return os.path.join(elm_model_path, "example_prompts.json") | |
| def run(elm_model_path: str): | |
| prompt_config_file = get_prompt_config_file(elm_model_path) | |
| with open(prompt_config_file, "r") as f: | |
| prompt_info = json.load(f) | |
| prompts = [prompt_info["template"].format(input=input) for input in prompt_info["inputs"]] | |
| print(f"Loaded prompts from: {prompt_config_file}") | |
| generate_elm_responses(elm_model_path, prompts, verbose=True) | |
| if __name__ == "__main__": | |
| args = parser.parse_args() | |
| run(args.elm_model_path) | 

