File size: 2,023 Bytes
6c46c52 36d972c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 |
---
license: cc-by-4.0
language:
- en
base_model:
- Qwen/Qwen3-8B
pipeline_tag: text-generation
---
# 𧬠Thoth
**Thoth** is a large language model for **biological experimental protocol generation**, designed to transform scientific knowledge into **accurate, logically ordered, and executable wet-lab procedures**.
- π **Paper**: *Unleashing Scientific Reasoning for Bio-experimental Protocol Generation via Structured Component-based Reward Mechanism* (ICLR 2026)
- π **GitHub**: https://github.com/manglu097/Thoth
- π€ **Dataset**: https://huggingface.co/datasets/manglu3935/SciRecipe
---
## π Model Overview
- **Base model**: Qwen3-8B
- **Parameters**: 8B
- **Training data**: SciRecipe (12K+ expert-curated biological protocols across 27 subfields)
- **Primary task**: End-to-end biological experimental protocol generation
Thoth follows a **Sketch-and-Fill** reasoning paradigm and is optimized using a **Structured Component-based Reward (SCORE)** mechanism, enforcing step ordering, granularity control, and semantic consistency.
---
## π§ Output Format
```
<think> reasoning and planning </think>
<key> structured machine-readable steps </key>
<orc> natural language protocol </orc>
<note> optional safety notes </note>
```
---
## π Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("manglu3935/Thoth")
model = AutoModelForCausalLM.from_pretrained("manglu3935/Thoth")
```
---
## β οΈ Intended Use
For research on scientific reasoning and experimental protocol generation.
Generated protocols must be reviewed by qualified domain experts before laboratory use.
---
## π Citation
```bibtex
@article{sun2025unleashing,
title={Unleashing Scientific Reasoning for Bio-experimental Protocol Generation via Structured Component-based Reward Mechanism},
author={Sun, Haoran and Jiang, Yankai and Tang, Zhenyu and others},
journal={arXiv preprint arXiv:2510.15600},
year={2025}
}
```
|