Spaces:
Running
Running
Commit
·
971caa1
1
Parent(s):
0bf1172
update notebook readme
Browse files
notebooks_tutorials/06_NTv3_generative_training.ipynb
CHANGED
|
@@ -7,9 +7,9 @@
|
|
| 7 |
"id": "15bb8c23"
|
| 8 |
},
|
| 9 |
"source": [
|
| 10 |
-
"# 🧬
|
| 11 |
"\n",
|
| 12 |
-
"This notebook demonstrates how to **fine-tune NTv3** for controllable enhancer sequence generation using **Masked Diffusion Language Modeling (MDLM)
|
| 13 |
"\n",
|
| 14 |
"---\n",
|
| 15 |
"\n",
|
|
|
|
| 7 |
"id": "15bb8c23"
|
| 8 |
},
|
| 9 |
"source": [
|
| 10 |
+
"# 🧬 Training a discrete diffusion model on biological sequences\n",
|
| 11 |
"\n",
|
| 12 |
+
"This notebook demonstrates how to **fine-tune NTv3** for controllable enhancer sequence generation using **Masked Diffusion Language Modeling (MDLM)** ([paper](https://arxiv.org/pdf/2406.07524)).\n",
|
| 13 |
"\n",
|
| 14 |
"---\n",
|
| 15 |
"\n",
|
notebooks_tutorials/07_enhancer_generation.ipynb
CHANGED
|
@@ -6,15 +6,15 @@
|
|
| 6 |
"id": "E06SHsd4Zf82"
|
| 7 |
},
|
| 8 |
"source": [
|
| 9 |
-
"# 🧬
|
| 10 |
"\n",
|
| 11 |
-
"This notebook demonstrates how to use **NTv3
|
| 12 |
"\n",
|
| 13 |
"> 📝 **Note for Google Colab users:** This notebook is compatible with Colab! For faster generation, enable GPU: Runtime → Change runtime type → GPU (T4 or better recommended). \n",
|
| 14 |
"\n",
|
| 15 |
"## 📋 Overview\n",
|
| 16 |
"\n",
|
| 17 |
-
"
|
| 18 |
"- **Promoter context** (DSCP or RpS12 backbone)\n",
|
| 19 |
"- **Activity level** (bins 0-4, from lowest to highest)\n",
|
| 20 |
"\n",
|
|
@@ -223,7 +223,7 @@
|
|
| 223 |
},
|
| 224 |
"outputs": [],
|
| 225 |
"source": [
|
| 226 |
-
"# Load NTv3
|
| 227 |
"repo_id = \"InstaDeepAI/NTv3_generative\"\n",
|
| 228 |
"\n",
|
| 229 |
"tokenizer = AutoTokenizer.from_pretrained(repo_id, trust_remote_code=True)\n",
|
|
|
|
| 6 |
"id": "E06SHsd4Zf82"
|
| 7 |
},
|
| 8 |
"source": [
|
| 9 |
+
"# 🧬 Controllable Enhancer Sequence Generation\n",
|
| 10 |
"\n",
|
| 11 |
+
"This notebook demonstrates how to use **NTv3** as a diffusion-based model for generating *de novo* enhancer sequences with controllable activity levels.\n",
|
| 12 |
"\n",
|
| 13 |
"> 📝 **Note for Google Colab users:** This notebook is compatible with Colab! For faster generation, enable GPU: Runtime → Change runtime type → GPU (T4 or better recommended). \n",
|
| 14 |
"\n",
|
| 15 |
"## 📋 Overview\n",
|
| 16 |
"\n",
|
| 17 |
+
"NTv uses **Masked Diffusion Language Modeling (MDLM)** to generate enhancer sequences conditioned on:\n",
|
| 18 |
"- **Promoter context** (DSCP or RpS12 backbone)\n",
|
| 19 |
"- **Activity level** (bins 0-4, from lowest to highest)\n",
|
| 20 |
"\n",
|
|
|
|
| 223 |
},
|
| 224 |
"outputs": [],
|
| 225 |
"source": [
|
| 226 |
+
"# Load NTv3 generative model from HuggingFace\n",
|
| 227 |
"repo_id = \"InstaDeepAI/NTv3_generative\"\n",
|
| 228 |
"\n",
|
| 229 |
"tokenizer = AutoTokenizer.from_pretrained(repo_id, trust_remote_code=True)\n",
|
tabs/home.html
CHANGED
|
@@ -88,7 +88,7 @@
|
|
| 88 |
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/03_fine_tuning_posttrained_model_biwig.ipynb" target="_blank" rel="noopener noreferrer">🎯 03 — Fine-tune a post-trained model on bigwig tracks</a></li>
|
| 89 |
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/04_fine_tuning_pretrained_model_annotation.ipynb" target="_blank" rel="noopener noreferrer">🏷️ 04 — Fine-tune a pre-trained model on annotations</a></li>
|
| 90 |
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/05_model_interpretation.ipynb" target="_blank" rel="noopener noreferrer">🔍 05 — Model interpretation</a></li>
|
| 91 |
-
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/06_NTv3_generative_training.ipynb" target="_blank" rel="noopener noreferrer">🧪 06 —
|
| 92 |
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/07_enhancer_generation.ipynb" target="_blank" rel="noopener noreferrer">🪰 07 — Generating enhancer sequences</a></li>
|
| 93 |
</ul>
|
| 94 |
</div>
|
|
|
|
| 88 |
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/03_fine_tuning_posttrained_model_biwig.ipynb" target="_blank" rel="noopener noreferrer">🎯 03 — Fine-tune a post-trained model on bigwig tracks</a></li>
|
| 89 |
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/04_fine_tuning_pretrained_model_annotation.ipynb" target="_blank" rel="noopener noreferrer">🏷️ 04 — Fine-tune a pre-trained model on annotations</a></li>
|
| 90 |
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/05_model_interpretation.ipynb" target="_blank" rel="noopener noreferrer">🔍 05 — Model interpretation</a></li>
|
| 91 |
+
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/06_NTv3_generative_training.ipynb" target="_blank" rel="noopener noreferrer">🧪 06 — Fine-tuning NTv3 into a diffusion model</a></li>
|
| 92 |
<li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/07_enhancer_generation.ipynb" target="_blank" rel="noopener noreferrer">🪰 07 — Generating enhancer sequences</a></li>
|
| 93 |
</ul>
|
| 94 |
</div>
|