| ``` | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| model_name = "shellchat-v1" | |
| model = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True).to("cuda") | |
| tokenizer = AutoTokenizer.from_pretrained(model_name) | |
| query = "hello world!" | |
| history = [] | |
| response = model.chat(query, history, tokenizer) | |
| ``` |