Update README.md
Browse files
README.md
CHANGED
|
@@ -57,9 +57,9 @@ print(pipe(prompt, max_new_tokens=2048, do_sample=True)[0]["generated_text"])
|
|
| 57 |
|
| 58 |
## Methodology
|
| 59 |
|
| 60 |
-
We
|
| 61 |
|
| 62 |
-
Prompts
|
| 63 |
|
| 64 |
Clusters tagged psychology or philosophy were retained for LoRA finetuning (rank=8, alpha=16, max length=2048, epoch=1, batch size=16).
|
| 65 |
|
|
|
|
| 57 |
|
| 58 |
## Methodology
|
| 59 |
|
| 60 |
+
We performed domain filtering on [Dolphin R1](https://huggingface.co/datasets/cognitivecomputations/dolphin-r1) and [General Reasoning](https://huggingface.co/datasets/GeneralReasoning/GeneralThought-430K).
|
| 61 |
|
| 62 |
+
Prompts were embedded, clustered with k-means (k=20 000) and majority-voted for domain labels using [Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B), following the [Intelligent Internet pipeline](https://huggingface.co/Intelligent-Internet/II-Medical-8B-1706).
|
| 63 |
|
| 64 |
Clusters tagged psychology or philosophy were retained for LoRA finetuning (rank=8, alpha=16, max length=2048, epoch=1, batch size=16).
|
| 65 |
|