bernardo-de-almeida commited on
Commit
971caa1
·
1 Parent(s): 0bf1172

update notebook readme

Browse files
notebooks_tutorials/06_NTv3_generative_training.ipynb CHANGED
@@ -7,9 +7,9 @@
7
  "id": "15bb8c23"
8
  },
9
  "source": [
10
- "# 🧬 NTv3-Generative: Training Tutorial\n",
11
  "\n",
12
- "This notebook demonstrates how to **fine-tune NTv3** for controllable enhancer sequence generation using **Masked Diffusion Language Modeling (MDLM)**.\n",
13
  "\n",
14
  "---\n",
15
  "\n",
 
7
  "id": "15bb8c23"
8
  },
9
  "source": [
10
+ "# 🧬 Training a discrete diffusion model on biological sequences\n",
11
  "\n",
12
+ "This notebook demonstrates how to **fine-tune NTv3** for controllable enhancer sequence generation using **Masked Diffusion Language Modeling (MDLM)** ([paper](https://arxiv.org/pdf/2406.07524)).\n",
13
  "\n",
14
  "---\n",
15
  "\n",
notebooks_tutorials/07_enhancer_generation.ipynb CHANGED
@@ -6,15 +6,15 @@
6
  "id": "E06SHsd4Zf82"
7
  },
8
  "source": [
9
- "# 🧬 NTv3-Generative: Controllable Enhancer Sequence Generation\n",
10
  "\n",
11
- "This notebook demonstrates how to use **NTv3-Generative**, a diffusion-based model for generating *de novo* enhancer sequences with controllable activity levels.\n",
12
  "\n",
13
  "> 📝 **Note for Google Colab users:** This notebook is compatible with Colab! For faster generation, enable GPU: Runtime → Change runtime type → GPU (T4 or better recommended). \n",
14
  "\n",
15
  "## 📋 Overview\n",
16
  "\n",
17
- "NTv3-Generative uses **Masked Diffusion Language Modeling (MDLM)** to generate enhancer sequences conditioned on:\n",
18
  "- **Promoter context** (DSCP or RpS12 backbone)\n",
19
  "- **Activity level** (bins 0-4, from lowest to highest)\n",
20
  "\n",
@@ -223,7 +223,7 @@
223
  },
224
  "outputs": [],
225
  "source": [
226
- "# Load NTv3-Generative model from HuggingFace\n",
227
  "repo_id = \"InstaDeepAI/NTv3_generative\"\n",
228
  "\n",
229
  "tokenizer = AutoTokenizer.from_pretrained(repo_id, trust_remote_code=True)\n",
 
6
  "id": "E06SHsd4Zf82"
7
  },
8
  "source": [
9
+ "# 🧬 Controllable Enhancer Sequence Generation\n",
10
  "\n",
11
+ "This notebook demonstrates how to use **NTv3** as a diffusion-based model for generating *de novo* enhancer sequences with controllable activity levels.\n",
12
  "\n",
13
  "> 📝 **Note for Google Colab users:** This notebook is compatible with Colab! For faster generation, enable GPU: Runtime → Change runtime type → GPU (T4 or better recommended). \n",
14
  "\n",
15
  "## 📋 Overview\n",
16
  "\n",
17
+ "NTv uses **Masked Diffusion Language Modeling (MDLM)** to generate enhancer sequences conditioned on:\n",
18
  "- **Promoter context** (DSCP or RpS12 backbone)\n",
19
  "- **Activity level** (bins 0-4, from lowest to highest)\n",
20
  "\n",
 
223
  },
224
  "outputs": [],
225
  "source": [
226
+ "# Load NTv3 generative model from HuggingFace\n",
227
  "repo_id = \"InstaDeepAI/NTv3_generative\"\n",
228
  "\n",
229
  "tokenizer = AutoTokenizer.from_pretrained(repo_id, trust_remote_code=True)\n",
tabs/home.html CHANGED
@@ -88,7 +88,7 @@
88
  <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/03_fine_tuning_posttrained_model_biwig.ipynb" target="_blank" rel="noopener noreferrer">🎯 03 — Fine-tune a post-trained model on bigwig tracks</a></li>
89
  <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/04_fine_tuning_pretrained_model_annotation.ipynb" target="_blank" rel="noopener noreferrer">🏷️ 04 — Fine-tune a pre-trained model on annotations</a></li>
90
  <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/05_model_interpretation.ipynb" target="_blank" rel="noopener noreferrer">🔍 05 — Model interpretation</a></li>
91
- <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/06_NTv3_generative_training.ipynb" target="_blank" rel="noopener noreferrer">🧪 06 — Training NTv3-generative</a></li>
92
  <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/07_enhancer_generation.ipynb" target="_blank" rel="noopener noreferrer">🪰 07 — Generating enhancer sequences</a></li>
93
  </ul>
94
  </div>
 
88
  <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/03_fine_tuning_posttrained_model_biwig.ipynb" target="_blank" rel="noopener noreferrer">🎯 03 — Fine-tune a post-trained model on bigwig tracks</a></li>
89
  <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/04_fine_tuning_pretrained_model_annotation.ipynb" target="_blank" rel="noopener noreferrer">🏷️ 04 — Fine-tune a pre-trained model on annotations</a></li>
90
  <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/05_model_interpretation.ipynb" target="_blank" rel="noopener noreferrer">🔍 05 — Model interpretation</a></li>
91
+ <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/06_NTv3_generative_training.ipynb" target="_blank" rel="noopener noreferrer">🧪 06 — Fine-tuning NTv3 into a diffusion model</a></li>
92
  <li><a href="https://huggingface.co/spaces/InstaDeepAI/ntv3/blob/main/notebooks_tutorials/07_enhancer_generation.ipynb" target="_blank" rel="noopener noreferrer">🪰 07 — Generating enhancer sequences</a></li>
93
  </ul>
94
  </div>