MAIRK commited on
Commit
706a225
·
verified ·
1 Parent(s): 53f447c

Upload README.md

Browse files

---
tags:
- llama
- text-generation
- conversational
- chatbot
license: mit
language:
- zh
- en
datasets:
- custom
metrics:
- name: C‑Eval EM
value: 68.3
- name: GPT4Bot‑Bench F1
value: 72.1
- name: SelfChat Similarity
value: 0.87
pipeline_tag: text-generation
model-index:
- name: MAIRK/abab
results: []
---

# My LLaMA 2 Chatbot (7B)

This repository contains a **fine‑tuned bilingual conversational model** based on **LLaMA 2 7B**, built for Chinese and English dialogue tasks.

...

Files changed (1) hide show
  1. README.md +30 -57
README.md CHANGED
@@ -1,57 +1,30 @@
1
- # My LLaMA 2 Chatbot (7B)
2
-
3
- This repository contains a **fine-tuned bilingual conversational model** based on **LLaMA 2 7B**, built for Chinese and English dialogue tasks.
4
-
5
- ## 🚀 Model Overview
6
-
7
- - **Base Model**: LLaMA 2 7B
8
- - **Languages**: Chinese (zh), English (en)
9
- - **Model Size**: 7 billion parameters
10
- - **Training Data**:
11
- - Chinese dialogue (Weibo, Zhihu)
12
- - English dialogue (Reddit, StackExchange)
13
- - Domain Q&A (IT, healthcare, finance)
14
-
15
- ## 📦 Files
16
-
17
- | File | Description |
18
- |------------------------|------------------------------------------|
19
- | `config.json` | Model architecture config |
20
- | `pytorch_model.bin` | Model weights |
21
- | `tokenizer.model` | SentencePiece tokenizer model |
22
- | `tokenizer_config.json`| Tokenizer configuration |
23
- | `special_tokens_map.json` | Token ID mapping for special tokens |
24
- | `generation_config.json` | Optional generation settings |
25
- | `README.md` | Model card & usage |
26
-
27
- ## ✨ Usage Example
28
-
29
- ```python
30
- from transformers import AutoTokenizer, AutoModelForCausalLM
31
-
32
- model = AutoModelForCausalLM.from_pretrained("your-username/my-llama2-chatbot", torch_dtype="auto")
33
- tokenizer = AutoTokenizer.from_pretrained("your-username/my-llama2-chatbot")
34
-
35
- prompt = "你好,请问你是谁?"
36
- inputs = tokenizer(prompt, return_tensors="pt")
37
- out = model.generate(**inputs, max_new_tokens=128)
38
- print(tokenizer.decode(out[0], skip_special_tokens=True))
39
- ```
40
-
41
- ## 🔐 License
42
-
43
- This model is released under the MIT license. If using Meta's weights, you must agree to [LLaMA 2 license terms](https://ai.meta.com/resources/models-and-libraries/llama-downloads/).
44
-
45
- ## 📈 Metrics
46
-
47
- - **C-Eval EM**: 68.3%
48
- - **GPT4Bot-Bench F1**: 72.1%
49
- - **SelfChat Similarity**: 0.87
50
-
51
- ## 🤗 Hosted on Hugging Face
52
-
53
- See: [https://huggingface.co/your-username/my-llama2-chatbot](https://huggingface.co/your-username/my-llama2-chatbot)
54
-
55
- ---
56
-
57
- For support, improvements or questions, please open an issue or pull request.
 
1
+ ---
2
+ tags:
3
+ - llama
4
+ - text-generation
5
+ - conversational
6
+ - chatbot
7
+ license: mit
8
+ language:
9
+ - zh
10
+ - en
11
+ datasets:
12
+ - custom
13
+ metrics:
14
+ - name: C‑Eval EM
15
+ value: 68.3
16
+ - name: GPT4Bot‑Bench F1
17
+ value: 72.1
18
+ - name: SelfChat Similarity
19
+ value: 0.87
20
+ pipeline_tag: text-generation
21
+ model-index:
22
+ - name: MAIRK/abab
23
+ results: []
24
+ ---
25
+
26
+ # My LLaMA 2 Chatbot (7B)
27
+
28
+ This repository contains a **fine‑tuned bilingual conversational model** based on **LLaMA 2 7B**, built for Chinese and English dialogue tasks.
29
+
30
+ ...