| | --- |
| | base_model: [] |
| | tags: |
| | - mergekit |
| | - merge |
| | license: apache-2.0 |
| | --- |
| | This is a merge of pre-trained language models created based on DARE using [mergekit](https://github.com/cg123/mergekit). |
| |
|
| | More descriptions of the model will be added soon. |
| |
|
| | ### **Loading the Model** |
| |
|
| | Use the following Python code to load the model: |
| |
|
| | ```python |
| | import torch |
| | from transformers import MistralForCausalLM, AutoTokenizer |
| | |
| | model = MistralForCausalLM.from_pretrained("vanillaOVO/Beagle_Turdus", device_map="auto") |
| | tokenizer = AutoTokenizer.from_pretrained("vanillaOVO/Beagle_Turdus") |
| | ``` |
| |
|
| | ### **Generating Text** |
| |
|
| | To generate text, use the following Python code: |
| |
|
| | ```python |
| | text = "Large language models are " |
| | inputs = tokenizer(text, return_tensors="pt") |
| | |
| | outputs = model.generate(**inputs, max_new_tokens=256) |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| | ``` |
| |
|