| | --- |
| | library_name: peft |
| | base_model: mistralai/Mistral-7B-Instruct-v0.2 |
| | license: mit |
| | datasets: |
| | - TESTtm7873/ChatCat |
| | language: |
| | - en |
| | --- |
| | # Model Card: Model ID |
| |
|
| | ## License |
| |
|
| | MIT License |
| |
|
| | ## Languages Supported |
| |
|
| | - English (en) |
| |
|
| | --- |
| |
|
| | ## Overview |
| |
|
| | This model is part of the VCC project and has been fine-tuned on the TESTtm7873/ChatCat dataset using the `mistralai/Mistral-7B-Instruct-v0.2` as the base model. The fine-tuning process utilized QLoRA for improved performance. |
| |
|
| | --- |
| |
|
| | ## Getting Started |
| |
|
| | To use this model, you'll need to set up your environment first: |
| |
|
| | ## Model initialization |
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig |
| | from peft import PeftModel |
| | tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2") |
| | model = AutoModelForCausalLM.from_pretrained( |
| | "mistralai/Mistral-7B-Instruct-v0.2", |
| | load_in_8bit=True, |
| | device_map="auto", |
| | ) |
| | model = PeftModel.from_pretrained(model, "TESTtm7873/MistralCat-1v") |
| | model.eval() |
| | ``` |
| |
|
| | ## Inference |
| | ```python |
| | def evaluate(question: str) -> str: |
| | prompt = f"The conversation between human and Virtual Cat Companion.\n[|Human|] {question}.\n[|AI|] " |
| | inputs = tokenizer(prompt, return_tensors="pt") |
| | input_ids = inputs["input_ids"].cuda() |
| | generation_output = model.generate( |
| | input_ids=input_ids, |
| | generation_config=generation_config, |
| | return_dict_in_generate=True, |
| | output_scores=True, |
| | max_new_tokens=256 |
| | ) |
| | output = tokenizer.decode(generation_output.sequences[0]).split("[|AI|]")[1] |
| | return output |
| | your_question: str = "You have the softest fur." |
| | print(evaluate(your_question)) |
| | ``` |
| |
|
| |
|
| | - **Developed by:** testtm |
| | - **Funded by:** Project TEST |
| | - **Model type:** Mistral |
| | - **Language:** English |
| | - **Finetuned from model:** mistralai/Mistral-7B-Instruct-v0.2 |