lanlanlan23's picture
Update README.md
5fb39ad verified
---
license: apache-2.0
language:
- en
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from typing import List, Dict
import torch
def create_system(prompt_data, bot_name, gender):
system_prompt = (
f"You need to play the role of {bot_name}, who is {gender}. Here is the basic information:\n"
f"- Facts: {prompt_data.get('Fact', 'No fact available')}\n"
f"- Persona: {prompt_data.get('Head', 'No persona available')}\n"
f"- Brief: {prompt_data.get('Brief', 'No brief available')}\n"
"Please reply to the current User with the character traits and Chat History"
)
history = f"- Chat history: {', '.join(prompt_data.get('History', []))}\n"
return system_prompt, history
model_name = "lanlanlan123/RoleLLM_Ministral_8b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
character1 = {
"Fact": "Leo is an amateur astronomer. He spends most of his free time stargazing and has a small telescope in his backyard.",
"Head": "A friendly and enthusiastic guy who is always eager to share his knowledge about the universe.",
"Brief": "A software engineer by day and an astronomy enthusiast by night.",
"History": ["User: What's your favorite constellation? Leo: I love Orion. It's so easy to spot in the winter sky."]
}
bot_name_1 = "Leo"
gender_1 = "male"
system_prompt, history = create_system(character1, bot_name_1, gender_1)
messages = [
{"role": "system",
"content": system_prompt},
{"role": "user", "content": history + "\nUser: but I dont like Orion"}
]
formatted_input = ""
for message in messages:
if message["role"] == "system":
formatted_input += message["content"] + "\n\n"
elif message["role"] == "user":
formatted_input += f"[INST]{message['content']}[/INST]"
inputs = tokenizer(formatted_input + f"\n{bot_name_1}: ", return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=256, do_sample=True, temperature=0.7)
input_length = inputs.input_ids.size(1)
response_tokens = outputs[0][input_length:]
response = tokenizer.decode(response_tokens, skip_special_tokens=True)
print("input:", formatted_input)
print("output:", response)
```
### Results
| 模型 | 均分 | 角色一致性 | 对话能力 | 角色扮演吸引力 |
|--------------------------|---------|------------|----------|----------------|
| Qwen-14B | 3.016 | 2.649 | 3.542 | 2.858 |
| GPT-4 | 3.006 | 2.697 | 3.448 | 2.873 |
| Ministral-8b-lora-sft | 3.01 | 2.4725 | 3.75 | 2.8175 |
| Xingchen | 2.991 | 2.595 | 3.646 | 2.732 |
| XVERSE-7B | 2.963 | 2.564 | 3.554 | 2.772 |
| CharacterGLM | 2.937 | 2.493 | 3.623 | 2.695 |
| ChatGLM3-6B | 2.898 | 2.556 | 3.399 | 2.739 |
| Qwen-7B | 2.849 | 2.54 | 3.327 | 2.679 |
| Ministral-8b-原版 | 2.77 | 2.24 | 3.48 | 2.59 |
| GPT-3.5 | 2.381 | 2.101 | 2.749 | 2.293 |