Text Generation
Safetensors
English
Chinese
medical

Improve language tag

#2
by lbourdois - opened
Files changed (1) hide show
  1. README.md +90 -79
README.md CHANGED
@@ -1,80 +1,91 @@
1
- ---
2
- license: apache-2.0
3
- datasets:
4
- - FreedomIntelligence/medical-o1-reasoning-SFT
5
- - FreedomIntelligence/medical-o1-verifiable-problem
6
- language:
7
- - en
8
- - zh
9
- base_model:
10
- - Qwen/Qwen2.5-7B-Instruct
11
- pipeline_tag: text-generation
12
- tags:
13
- - medical
14
- ---
15
-
16
- <div align="center">
17
- <h1>
18
- HuatuoGPT-o1-7B
19
- </h1>
20
- </div>
21
-
22
- <div align="center">
23
- <a href="https://github.com/FreedomIntelligence/HuatuoGPT-o1" target="_blank">GitHub</a> | <a href="https://arxiv.org/pdf/2412.18925" target="_blank">Paper</a>
24
- </div>
25
-
26
- # <span>Introduction</span>
27
- **HuatuoGPT-o1** is a medical LLM designed for advanced medical reasoning. It generates a complex thought process, reflecting and refining its reasoning, before providing a final response.
28
-
29
- For more information, visit our GitHub repository:
30
- [https://github.com/FreedomIntelligence/HuatuoGPT-o1](https://github.com/FreedomIntelligence/HuatuoGPT-o1).
31
-
32
- # <span>Model Info</span>
33
- | | Backbone | Supported Languages | Link |
34
- | -------------------- | ------------ | ----- | --------------------------------------------------------------------- |
35
- | **HuatuoGPT-o1-8B** | LLaMA-3.1-8B | English | [HF Link](https://huggingface.co/FreedomIntelligence/HuatuoGPT-o1-8B) |
36
- | **HuatuoGPT-o1-70B** | LLaMA-3.1-70B | English | [HF Link](https://huggingface.co/FreedomIntelligence/HuatuoGPT-o1-70B) |
37
- | **HuatuoGPT-o1-7B** | Qwen2.5-7B | English & Chinese | [HF Link](https://huggingface.co/FreedomIntelligence/HuatuoGPT-o1-7B) |
38
- | **HuatuoGPT-o1-72B** | Qwen2.5-72B | English & Chinese | [HF Link](https://huggingface.co/FreedomIntelligence/HuatuoGPT-o1-72B) |
39
-
40
-
41
-
42
- # <span>Usage</span>
43
- You can use HuatuoGPT-o1-7B in the same way as `Qwen2.5-7B-Instruct`. You can deploy it with tools like [vllm](https://github.com/vllm-project/vllm) or [Sglang](https://github.com/sgl-project/sglang), or perform direct inference:
44
- ```python
45
- from transformers import AutoModelForCausalLM, AutoTokenizer
46
-
47
- model = AutoModelForCausalLM.from_pretrained("FreedomIntelligence/HuatuoGPT-o1-7B",torch_dtype="auto",device_map="auto")
48
- tokenizer = AutoTokenizer.from_pretrained("FreedomIntelligence/HuatuoGPT-o1-7B")
49
-
50
- input_text = "How to stop a cough?"
51
- messages = [{"role": "user", "content": input_text}]
52
-
53
- inputs = tokenizer(tokenizer.apply_chat_template(messages, tokenize=False,add_generation_prompt=True
54
- ), return_tensors="pt").to(model.device)
55
- outputs = model.generate(**inputs, max_new_tokens=2048)
56
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
57
- ```
58
-
59
- HuatuoGPT-o1 adopts a *thinks-before-it-answers* approach, with outputs formatted as:
60
-
61
- ```
62
- ## Thinking
63
- [Reasoning process]
64
-
65
- ## Final Response
66
- [Output]
67
- ```
68
-
69
- # <span>📖 Citation</span>
70
- ```
71
- @misc{chen2024huatuogpto1medicalcomplexreasoning,
72
- title={HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs},
73
- author={Junying Chen and Zhenyang Cai and Ke Ji and Xidong Wang and Wanlong Liu and Rongsheng Wang and Jianye Hou and Benyou Wang},
74
- year={2024},
75
- eprint={2412.18925},
76
- archivePrefix={arXiv},
77
- primaryClass={cs.CL},
78
- url={https://arxiv.org/abs/2412.18925},
79
- }
 
 
 
 
 
 
 
 
 
 
 
80
  ```
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - FreedomIntelligence/medical-o1-reasoning-SFT
5
+ - FreedomIntelligence/medical-o1-verifiable-problem
6
+ language:
7
+ - zho
8
+ - eng
9
+ - fra
10
+ - spa
11
+ - por
12
+ - deu
13
+ - ita
14
+ - rus
15
+ - jpn
16
+ - kor
17
+ - vie
18
+ - tha
19
+ - ara
20
+ base_model:
21
+ - Qwen/Qwen2.5-7B-Instruct
22
+ pipeline_tag: text-generation
23
+ tags:
24
+ - medical
25
+ ---
26
+
27
+ <div align="center">
28
+ <h1>
29
+ HuatuoGPT-o1-7B
30
+ </h1>
31
+ </div>
32
+
33
+ <div align="center">
34
+ <a href="https://github.com/FreedomIntelligence/HuatuoGPT-o1" target="_blank">GitHub</a> | <a href="https://arxiv.org/pdf/2412.18925" target="_blank">Paper</a>
35
+ </div>
36
+
37
+ # <span>Introduction</span>
38
+ **HuatuoGPT-o1** is a medical LLM designed for advanced medical reasoning. It generates a complex thought process, reflecting and refining its reasoning, before providing a final response.
39
+
40
+ For more information, visit our GitHub repository:
41
+ [https://github.com/FreedomIntelligence/HuatuoGPT-o1](https://github.com/FreedomIntelligence/HuatuoGPT-o1).
42
+
43
+ # <span>Model Info</span>
44
+ | | Backbone | Supported Languages | Link |
45
+ | -------------------- | ------------ | ----- | --------------------------------------------------------------------- |
46
+ | **HuatuoGPT-o1-8B** | LLaMA-3.1-8B | English | [HF Link](https://huggingface.co/FreedomIntelligence/HuatuoGPT-o1-8B) |
47
+ | **HuatuoGPT-o1-70B** | LLaMA-3.1-70B | English | [HF Link](https://huggingface.co/FreedomIntelligence/HuatuoGPT-o1-70B) |
48
+ | **HuatuoGPT-o1-7B** | Qwen2.5-7B | English & Chinese | [HF Link](https://huggingface.co/FreedomIntelligence/HuatuoGPT-o1-7B) |
49
+ | **HuatuoGPT-o1-72B** | Qwen2.5-72B | English & Chinese | [HF Link](https://huggingface.co/FreedomIntelligence/HuatuoGPT-o1-72B) |
50
+
51
+
52
+
53
+ # <span>Usage</span>
54
+ You can use HuatuoGPT-o1-7B in the same way as `Qwen2.5-7B-Instruct`. You can deploy it with tools like [vllm](https://github.com/vllm-project/vllm) or [Sglang](https://github.com/sgl-project/sglang), or perform direct inference:
55
+ ```python
56
+ from transformers import AutoModelForCausalLM, AutoTokenizer
57
+
58
+ model = AutoModelForCausalLM.from_pretrained("FreedomIntelligence/HuatuoGPT-o1-7B",torch_dtype="auto",device_map="auto")
59
+ tokenizer = AutoTokenizer.from_pretrained("FreedomIntelligence/HuatuoGPT-o1-7B")
60
+
61
+ input_text = "How to stop a cough?"
62
+ messages = [{"role": "user", "content": input_text}]
63
+
64
+ inputs = tokenizer(tokenizer.apply_chat_template(messages, tokenize=False,add_generation_prompt=True
65
+ ), return_tensors="pt").to(model.device)
66
+ outputs = model.generate(**inputs, max_new_tokens=2048)
67
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
68
+ ```
69
+
70
+ HuatuoGPT-o1 adopts a *thinks-before-it-answers* approach, with outputs formatted as:
71
+
72
+ ```
73
+ ## Thinking
74
+ [Reasoning process]
75
+
76
+ ## Final Response
77
+ [Output]
78
+ ```
79
+
80
+ # <span>📖 Citation</span>
81
+ ```
82
+ @misc{chen2024huatuogpto1medicalcomplexreasoning,
83
+ title={HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs},
84
+ author={Junying Chen and Zhenyang Cai and Ke Ji and Xidong Wang and Wanlong Liu and Rongsheng Wang and Jianye Hou and Benyou Wang},
85
+ year={2024},
86
+ eprint={2412.18925},
87
+ archivePrefix={arXiv},
88
+ primaryClass={cs.CL},
89
+ url={https://arxiv.org/abs/2412.18925},
90
+ }
91
  ```