Mistral_Chat / getting_started_with_mistral_7b_instruct.py
ahmtsari's picture
Upload getting_started_with_mistral_7b_instruct.py
f440d97 verified
# -*- coding: utf-8 -*-
"""Getting-Started-with-Mistral-7b-Instruct.ipynb
Automatically generated by Colaboratory.
Original file is located at
https://colab.research.google.com/drive/1ckGJcooxH_jiohgmb7PIKrFsDxHoUPq2
"""
!pip install -q git+https://github.com/huggingface/transformers
model_path="/kaggle/input/mistral/pytorch/7b-instruct-v0.1-hf/1"
from transformers import AutoTokenizer
tokenizer=AutoTokenizer.from_pretrained(model_path)
from transformers import AutoModelForCausalLM
import torch
model = AutoModelForCausalLM.from_pretrained(
model_path,
torch_dtype = torch.bfloat16,
device_map = "auto",
trust_remote_code = True
)
messages = [{
"role":"user",
"content": "Can you tell us 3 cities to visit in Turkey"
}]
tokenizer.apply_chat_template(messages, tokenize=False)
model_inputs = tokenizer.apply_chat_template(messages,
return_tensors = "pt")
model_inputs
generated_ids = model.generate(
model_inputs,
max_new_tokens = 1000,
do_sample = True,
)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
messages = [{
"role": "user",
"content": "Act as a gourmet chef. I have a friend coming over who is a vegetarian. \
I want to impress my friend with a special vegetarian dish. \
What do you recommend? \
Give me two options, along with the whole recipe for each"
}]
model_inputs = tokenizer.apply_chat_template(messages,return_tensors = "pt")
generated_ids = model.generate(
model_inputs,
max_new_tokens = 1000,
do_sample = True,
)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
messages = [{
"role": "user", "content": "How many helicopters can a human eat in one sitting?"
}]
model_inputs = tokenizer.apply_chat_template(messages,return_tensors = "pt")
generated_ids = model.generate(
model_inputs,
max_new_tokens = 1000,
do_sample = True,
)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
messages = [
{"role": "user", "content": "What fascinates you about AI?"},
{"role": "assistant", "content": "I'm fascinated \
by AI's data analysis and prediction abilities. \
It has the potential to revolutionize industries and improve problem-solving."},
{"role": "user", "content": "Should people be afraid of AI?"}
]
model_inputs = tokenizer.apply_chat_template(messages,return_tensors = "pt")
generated_ids = model.generate(
model_inputs,
max_new_tokens = 1000,
do_sample = True,
)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
messages = [
{"role": "user", "content": "Türkiye'de en fazla ziyaret edilen 3 şehir hangidir?"},
]
model_inputs = tokenizer.apply_chat_template(messages,return_tensors = "pt")
generated_ids = model.generate(
model_inputs,
max_new_tokens = 1000,
do_sample = True,
)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
"""Let's connect [YouTube](http://youtube.com/tirendazacademy) | [Medium](http://tirendazacademy.medium.com) | [X](http://x.com/tirendazacademy) | [Linkedin](https://www.linkedin.com/in/tirendaz-academy)"""