lia21 commited on
Commit
0a9da78
·
verified ·
1 Parent(s): 2e3890e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -160
README.md CHANGED
@@ -1,170 +1,29 @@
1
  ---
2
  tags:
3
- - bittensor
4
- - subnet-20
5
- - agent
6
- - function-calling
7
- - generated_from_trainer
 
 
8
  license: mit
9
- datasets:
10
- - HuggingFaceH4/ultrachat_200k
11
- - HuggingFaceH4/ultrafeedback_binarized
12
  language:
13
- - en
14
  base_model: mistralai/Mistral-7B-v0.1
 
 
 
15
  pipeline_tag: text-generation
16
- widget:
17
- - example_title: Pirate!
18
- messages:
19
- - role: system
20
- content: You are a pirate chatbot who always responds with Arr!
21
- - role: user
22
- content: There's a llama on my lawn, how can I get rid of him?
23
- output:
24
- text: >-
25
- Arr! 'Tis a puzzlin' matter, me hearty! A llama on yer lawn be a rare
26
- sight, but I've got a plan that might help ye get rid of 'im. Ye'll need
27
- to gather some carrots and hay, and then lure the llama away with the
28
- promise of a tasty treat. Once he's gone, ye can clean up yer lawn and
29
- enjoy the peace and quiet once again. But beware, me hearty, for there
30
- may be more llamas where that one came from! Arr!
31
- model-index:
32
- - name: zephyr-7b-beta
33
- results:
34
- - task:
35
- type: text-generation
36
- name: Text Generation
37
- dataset:
38
- name: AI2 Reasoning Challenge (25-Shot)
39
- type: ai2_arc
40
- config: ARC-Challenge
41
- split: test
42
- args:
43
- num_few_shot: 25
44
- metrics:
45
- - type: acc_norm
46
- name: normalized accuracy
47
- value: 62.03
48
- - task:
49
- type: text-generation
50
- name: Text Generation
51
- dataset:
52
- name: HellaSwag (10-Shot)
53
- type: hellaswag
54
- split: validation
55
- args:
56
- num_few_shot: 10
57
- metrics:
58
- - type: acc_norm
59
- name: normalized accuracy
60
- value: 84.36
61
- - task:
62
- type: text-generation
63
- name: Text Generation
64
- dataset:
65
- name: Drop (3-Shot)
66
- type: drop
67
- split: validation
68
- args:
69
- num_few_shot: 3
70
- metrics:
71
- - type: f1
72
- name: f1 score
73
- value: 9.66
74
- - task:
75
- type: text-generation
76
- name: Text Generation
77
- dataset:
78
- name: TruthfulQA (0-shot)
79
- type: truthful_qa
80
- config: multiple_choice
81
- split: validation
82
- args:
83
- num_few_shot: 0
84
- metrics:
85
- - type: mc2
86
- value: 57.45
87
- - task:
88
- type: text-generation
89
- name: Text Generation
90
- dataset:
91
- name: GSM8k (5-shot)
92
- type: gsm8k
93
- config: main
94
- split: test
95
- args:
96
- num_few_shot: 5
97
- metrics:
98
- - type: acc
99
- name: accuracy
100
- value: 12.74
101
- - task:
102
- type: text-generation
103
- name: Text Generation
104
- dataset:
105
- name: MMLU (5-Shot)
106
- type: cais/mmlu
107
- config: all
108
- split: test
109
- args:
110
- num_few_shot: 5
111
- metrics:
112
- - type: acc
113
- name: accuracy
114
- value: 61.07
115
- - task:
116
- type: text-generation
117
- name: Text Generation
118
- dataset:
119
- name: Winogrande (5-shot)
120
- type: winogrande
121
- config: winogrande_xl
122
- split: validation
123
- args:
124
- num_few_shot: 5
125
- metrics:
126
- - type: acc
127
- name: accuracy
128
- value: 77.74
129
- - task:
130
- type: text-generation
131
- name: Text Generation
132
- dataset:
133
- name: AlpacaEval
134
- type: tatsu-lab/alpaca_eval
135
- metrics:
136
- - type: unknown
137
- name: win rate
138
- value: 0.906
139
- - task:
140
- type: text-generation
141
- name: Text Generation
142
- dataset:
143
- name: MT-Bench
144
- type: unknown
145
- metrics:
146
- - type: unknown
147
- name: score
148
- value: 7.34
149
  ---
150
 
151
- # Zephyr-7B-Beta (Fine-tuned for Subnet 20 BitAgent)
152
-
153
- This is a fine-tuned version of **Mistral-7B** adapted for **function-calling and reasoning tasks** in **Bittensor Subnet 20 (BitAgent)**.
154
-
155
- ## 🧠 Use Case
156
- - Works as an **agent LLM** inside the BitAgent subnet.
157
- - Supports **reasoning** and **function-calling outputs**.
158
- - Optimized for **task delegation and structured outputs**.
159
-
160
- ## 🚀 How to Use
161
- ```python
162
- from transformers import AutoModelForCausalLM, AutoTokenizer
163
 
164
- model = AutoModelForCausalLM.from_pretrained("username/zephyr-7b-beta-bitagent")
165
- tokenizer = AutoTokenizer.from_pretrained("username/zephyr-7b-beta-bitagent")
166
 
167
- prompt = "Summarize the latest research in AI safety in 3 bullet points."
168
- inputs = tokenizer(prompt, return_tensors="pt")
169
- outputs = model.generate(**inputs, max_new_tokens=200)
170
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
 
 
1
  ---
2
  tags:
3
+ - bittensor
4
+ - subnet-20
5
+ - agent
6
+ - function-calling
7
+ - json-output
8
+ - mistral
9
+ - conversational
10
  license: mit
 
 
 
11
  language:
12
+ - en
13
  base_model: mistralai/Mistral-7B-v0.1
14
+ datasets:
15
+ - HuggingFaceH4/ultrachat_200k
16
+ - HuggingFaceH4/ultrafeedback_binarized
17
  pipeline_tag: text-generation
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ---
19
 
20
+ # lia21/bitagent — Zephyr-7B for **Bittensor Subnet 20 (BitAgent)**
 
 
 
 
 
 
 
 
 
 
 
21
 
22
+ Fine-tuned version of **Mistral-7B** adapted for **function-calling** and **structured reasoning** inside **Subnet 20 (BitAgent)**.
23
+ The model is optimized to produce **valid JSON outputs** of the form:
24
 
25
+ ```json
26
+ {
27
+ "name": "<function_name>",
28
+ "arguments": { "...": "..." }
29
+ }