alphaaico commited on
Commit
d81de5d
·
verified ·
1 Parent(s): ef5b584

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +99 -7
README.md CHANGED
@@ -1,21 +1,113 @@
1
  ---
2
- base_model: unsloth/llama-3.2-1b-instruct-unsloth-bnb-4bit
3
  tags:
4
  - text-generation-inference
5
  - transformers
6
  - unsloth
7
  - llama
 
 
8
  license: apache-2.0
9
  language:
10
  - en
11
  ---
12
 
13
- # Uploaded finetuned model
 
 
 
 
 
 
14
 
15
- - **Developed by:** alphaaico
16
- - **License:** apache-2.0
17
- - **Finetuned from model :** unsloth/llama-3.2-1b-instruct-unsloth-bnb-4bit
18
 
19
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: meta-llama/Llama-3.2-1B-Instruct
3
  tags:
4
  - text-generation-inference
5
  - transformers
6
  - unsloth
7
  - llama
8
+ - gguf
9
+ - chain-of-thought
10
  license: apache-2.0
11
  language:
12
  - en
13
  ---
14
 
15
+ <div align="center">
16
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/669777597cb32718c20d97e9/4emWK_PB-RrifIbrCUjE8.png"
17
+ alt="Title card"
18
+ style="width: 500px;
19
+ height: auto;
20
+ object-position: center top;">
21
+ </div>
22
 
23
+ **Website - https://www.alphaai.biz**
 
 
24
 
25
+ # Model Name: Medical-Guide-COT-llama3.2-1B
26
 
27
+ **Developed by:** Alpha AI
28
+ **License:** apache-2.0
29
+ **Finetuned from model:** meta-llama/Llama-3.2-1B-Instruct
30
+ **Formats available:** Float16 (safetensors + GGUF-f16), GGUF quantized (q4\_k\_m, q5\_k\_m, q8\_0)
31
+
32
+ ## Overview
33
+
34
+ **Medical-Guide-COT-llama3.2-1B** is a lightweight yet powerful medical reasoning model designed to produce explicit **Chain of Thought (CoT)** reasoning with `<think>...</think>` tags for transparency and clarity. Built for interpretability and performance, this model excels in structured medical question answering.
35
+
36
+ * **Finetuning Objective:** Supervised fine-tuning (SFT) on medical QA datasets with enforced reasoning chains.
37
+ * **Instruction format:** Adheres to Llama 3.2 Instruct prompting standards.
38
+ * **Deployment flexibility:** Offers multiple GGUF quantized variants for local, edge, or efficient inference environments.
39
+
40
+ ## Training Data
41
+
42
+ * **Public sources:** PubMedQA, MedMCQA, USMLE-type questions (filtered)
43
+ * **Proprietary augmentation:** Alpha AI's curated "Clinical-Cases-CoT" dataset with physician-authored reasoning chains
44
+ * **Sample size:** 42,000 examples (approx. 60% public / 40% private)
45
+ * **Token structure:**
46
+
47
+ ```
48
+ <think>
49
+ Step-by-step clinical reasoning...
50
+ </think>
51
+ Final answer.
52
+ ```
53
+
54
+ ## Model Specifications
55
+
56
+ | Attribute | Value |
57
+ | -------------- | ----------------------------------------- |
58
+ | Base Model | meta-llama/Llama-3.2-1B-Instruct |
59
+ | Model Type | Causal Language Model |
60
+ | Finetuned By | Alpha AI |
61
+ | Precision | Float16, GGUF q4\_k\_m / q5\_k\_m / q8\_0 |
62
+ | Context Length | 8,192 tokens |
63
+ | Language | English |
64
+
65
+ ## Intended Use
66
+
67
+ * **Medical Education:** Transparent QA for students (USMLE/PLAB prep)
68
+ * **Prototype Decision Support:** Clear reasoning steps before answers
69
+ * **Research on COT Safety:** Evaluation of model interpretability and hallucination control
70
+
71
+ ## Example Usage
72
+
73
+ ```python
74
+ from transformers import AutoModelForCausalLM, AutoTokenizer
75
+
76
+ model_id = "alpha-ai/Medical-Guide-COT-llama3.2-1B"
77
+ model = AutoModelForCausalLM.from_pretrained(model_id, trust_remote_code=True)
78
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
79
+
80
+ prompt = """### Question:
81
+ A 65-year-old male presents with sudden chest pain radiating to the back. Most likely diagnosis?
82
+ ### Answer:
83
+ """
84
+ inputs = tokenizer(prompt, return_tensors="pt")
85
+ outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7, top_p=0.9)
86
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
87
+ ```
88
+
89
+ **Expected Output Format:**
90
+
91
+ ```text
92
+ <think>
93
+ Sudden tearing chest pain suggests aortic dissection.
94
+ Hypertension is a key risk factor. Location of pain supports Stanford Type A.
95
+ </think>
96
+ Acute aortic dissection (Stanford Type A)
97
+ ```
98
+
99
+ ## Limitations & Usage Warnings
100
+
101
+ * **Not a clinical diagnostic tool.** Use only for research or educational purposes.
102
+ * **Bias & Hallucination Risk.** Outputs must be validated by qualified professionals.
103
+ * **Sensitive Content.** Model not trained on PHI but care should be taken with input prompts.
104
+
105
+ ## License
106
+
107
+ Distributed under the **Apache-2.0** license.
108
+
109
+ ## Acknowledgments
110
+
111
+ Thanks to Meta AI for Llama-3.2, the creators of open medical QA datasets, and the Alpha AI medical advisory board for domain alignment and data verification.
112
+
113
+ **Website:** [https://www.alphaai.biz](https://www.alphaai.biz)