Update README.md
Browse files
README.md
CHANGED
|
@@ -72,9 +72,9 @@ config = PeftConfig.from_pretrained('VityaVitalich/TaxoLLaMA')
|
|
| 72 |
# Do not forget your token for Llama2 models
|
| 73 |
model = LlamaForCausalLM.from_pretrained(config.base_model_name_or_path, load_in_4bit=True, torch_dtype=torch.bfloat16)
|
| 74 |
tokenizer = LlamaTokenizer.from_pretrained(config.base_model_name_or_path)
|
| 75 |
-
inference_model = PeftModel.from_pretrained(model, 'VityaVitalich/TaxoLLaMA'
|
| 76 |
|
| 77 |
-
processed_term =
|
| 78 |
|
| 79 |
system_prompt = """<s>[INST] <<SYS>> You are a helpfull assistant. List all the possible words divided with a coma. Your answer should not include anything except the words divided by a coma<</SYS>>"""
|
| 80 |
processed_term = system_prompt + '\n' + processed_term + '[/INST]'
|
|
|
|
| 72 |
# Do not forget your token for Llama2 models
|
| 73 |
model = LlamaForCausalLM.from_pretrained(config.base_model_name_or_path, load_in_4bit=True, torch_dtype=torch.bfloat16)
|
| 74 |
tokenizer = LlamaTokenizer.from_pretrained(config.base_model_name_or_path)
|
| 75 |
+
inference_model = PeftModel.from_pretrained(model, 'VityaVitalich/TaxoLLaMA')
|
| 76 |
|
| 77 |
+
processed_term = "hyponym: tiger | hypernyms:"
|
| 78 |
|
| 79 |
system_prompt = """<s>[INST] <<SYS>> You are a helpfull assistant. List all the possible words divided with a coma. Your answer should not include anything except the words divided by a coma<</SYS>>"""
|
| 80 |
processed_term = system_prompt + '\n' + processed_term + '[/INST]'
|