| import os |
| import argparse |
| import json |
| from elm.infer_elm import generate_elm_responses |
|
|
| parser = argparse.ArgumentParser(description='run prompts with elm model.') |
| parser.add_argument('elm_model_path', help='Path to the elm_model_path') |
|
|
|
|
| def get_prompt_config_file(elm_model_path): |
| return os.path.join(elm_model_path, "example_prompts.json") |
|
|
| def run(elm_model_path: str): |
| prompt_config_file = get_prompt_config_file(elm_model_path) |
| |
| with open(prompt_config_file, "r") as f: |
| prompt_info = json.load(f) |
| prompts = [prompt_info["template"].format(input=input) for input in prompt_info["inputs"]] |
| print(f"Loaded prompts from: {prompt_config_file}") |
| generate_elm_responses(elm_model_path, prompts, verbose=True) |
| |
| if __name__ == "__main__": |
| args = parser.parse_args() |
| run(args.elm_model_path) |