EthanLiu1991 commited on
Commit
743890e
·
verified ·
1 Parent(s): e987231

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. README.md +108 -0
  2. biomedgpt.safetensors +3 -0
README.md ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ tags:
4
+ - merge
5
+ - mergekit
6
+ - lazymergekit
7
+ - automerger
8
+ ---
9
+
10
+ ## 🧩 Configuration
11
+ ```yaml
12
+ #slices:
13
+ # - sources:
14
+ # - model: liminerity/M7-7b
15
+ # layer_range: [0, 32]
16
+ # - model: AurelPx/Percival_01-7b-slerp
17
+ # layer_range: [0, 32]
18
+ #merge_method: slerp
19
+ #base_model: liminerity/M7-7b
20
+ #parameters:
21
+ # t:
22
+ # - filter: self_attn
23
+ # value: [0.729086620552417, 0.4742644222549576, 0.47065411083849984, 0.9373988134098882, 0.6820526568624088]
24
+ # - filter: mlp
25
+ # value: [0.270913379447583, 0.5257355777450424, 0.06260118659011182, 0.06260118659011182, 0.3179473431375912]
26
+ # - value: 0.8480269455484635
27
+ #dtype: bfloat16
28
+ #random_seed: 0
29
+
30
+ #slices:
31
+ # - sources:
32
+ # - model: psmathur/orca_mini_v3_13b
33
+ # layer_range: [0, 40]
34
+ # - model: garage-bAInd/Platypus2-7b
35
+ # layer_range: [0, 32]
36
+ #merge_method: slerp
37
+ #base_model: psmathur/orca_mini_v3_13b
38
+ #parameters:
39
+ # t:
40
+ # - filter: self_attn
41
+ # value: [0.729086620552417, 0.4742644222549576, 0.47065411083849984, 0.9373988134098882, 0.6820526568624088]
42
+ # - filter: mlp
43
+ # value: [0.270913379447583, 0.5257355777450424, 0.5293458891615002, 0.06260118659011182, 0.3179473431375912]
44
+ # - value: 0.8480269455484635
45
+ #dtype: float16
46
+ #random_seed: 0
47
+
48
+ #slices:
49
+ # - sources:
50
+ # - model: psmathur/orca_mini_v3_13b
51
+ # parameters:
52
+ # density: [1, 0.7, 0.1] # density gradient
53
+ # weight: 1.0
54
+ # - model: garage-bAInd/Platypus2-13B
55
+ # parameters:
56
+ # density: 0.5
57
+ # weight: [0, 0.3, 0.7, 1] # weight gradient
58
+ # - model: WizardLM/WizardMath-13B-V1.0
59
+ # parameters:
60
+ # density: 0.33
61
+ # weight:
62
+ # - filter: mlp
63
+ # value: 0.5
64
+ # - value: 0
65
+ #merge_method: ties
66
+ #base_model: TheBloke/Llama-2-13B-fp16
67
+ #parameters:
68
+ # normalize: true
69
+ # int8_mask: true
70
+ #dtype: float16
71
+ #random_seed: 0
72
+
73
+ base_model: mlabonne/AlphaMonarch-7B
74
+ experts:
75
+ - source_model: mlabonne/AlphaMonarch-7B
76
+ positive_prompts:
77
+ - "chat"
78
+ - "assistant"
79
+ - "tell me"
80
+ - "explain"
81
+ - "I want"
82
+ - source_model: beowolx/CodeNinja-1.0-OpenChat-7B
83
+ positive_prompts:
84
+ - "code"
85
+ - "python"
86
+ - "javascript"
87
+ - "programming"
88
+ - "algorithm"
89
+ ```
90
+ ## 💻 Usage
91
+ ```python
92
+ !pip install -qU transformers accelerate
93
+ from transformers import AutoTokenizer
94
+ import transformers
95
+ import torch
96
+ model = "EthanLiu1991/BioMedGPT"
97
+ messages = [{"role": "user", "content": "What is a large language model?"}]
98
+ tokenizer = AutoTokenizer.from_pretrained(model)
99
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
100
+ pipeline = transformers.pipeline(
101
+ "text-generation",
102
+ model=model,
103
+ torch_dtype=torch.float16,
104
+ device_map="auto",
105
+ )
106
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
107
+ print(outputs[0]["generated_text"])
108
+ ```
biomedgpt.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:146b78b8eb19ea6de4355cc11fb502157f6e4c63b24850925b5fc7378c38ef4d
3
+ size 19214580994