Update README.md
Browse files
README.md
CHANGED
|
@@ -1,121 +1,128 @@
|
|
| 1 |
-
|
| 2 |
-
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
β
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
β β’
|
| 27 |
-
β β’
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
-
|
| 54 |
-
|
| 55 |
-
##
|
| 56 |
-
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
|
| 61 |
-
|
| 62 |
-
|
| 63 |
-
|
| 64 |
-
|
| 65 |
-
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
|
| 73 |
-
|
| 74 |
-
|
| 75 |
-
|
| 76 |
-
|
| 77 |
-
|
| 78 |
-
|
| 79 |
-
|
| 80 |
-
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
|
| 84 |
-
|
| 85 |
-
|
| 86 |
-
|
| 87 |
-
|
| 88 |
-
|
| 89 |
-
|
| 90 |
-
|
| 91 |
-
|
| 92 |
-
|
| 93 |
-
|
| 94 |
-
|
| 95 |
-
|
| 96 |
-
|
| 97 |
-
)
|
| 98 |
-
|
| 99 |
-
|
| 100 |
-
|
| 101 |
-
|
| 102 |
-
|
| 103 |
-
|
| 104 |
-
|
| 105 |
-
|
| 106 |
-
|
| 107 |
-
|
| 108 |
-
|
| 109 |
-
|
| 110 |
-
|
| 111 |
-
|
| 112 |
-
|
| 113 |
-
|
|
| 114 |
-
|
| 115 |
-
|
| 116 |
-
|
| 117 |
-
-
|
| 118 |
-
-
|
| 119 |
-
|
| 120 |
-
|
| 121 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model:
|
| 3 |
+
- microsoft/deberta-v3-base
|
| 4 |
+
pipeline_tag: token-classification
|
| 5 |
+
tags:
|
| 6 |
+
- links
|
| 7 |
+
---
|
| 8 |
+
# Link Anchor Detection Model
|
| 9 |
+
|
| 10 |
+
A fine-tuned DeBERTa v3 model that predicts which words in text should be hyperlinks. Trained on 10,273 pages scraped from [The Keyword](https://blog.google/) (Google's official blog), where editorial linking decisions serve as ground truth labels.
|
| 11 |
+
|
| 12 |
+
## How It Works
|
| 13 |
+
|
| 14 |
+
Given raw text, the model performs token-level binary classification β each token is labeled `LINK` or `O` (not a link). This identifies anchor text candidates: words that a human editor would likely hyperlink.
|
| 15 |
+
|
| 16 |
+
## Pipeline
|
| 17 |
+
|
| 18 |
+
```
|
| 19 |
+
sitemap.xml (10,274 URLs from blog.google)
|
| 20 |
+
β
|
| 21 |
+
βΌ
|
| 22 |
+
scrape.py βββΊ scraped.db (SQLite, 10,273 pages with markdown + inline links)
|
| 23 |
+
β
|
| 24 |
+
βΌ
|
| 25 |
+
_prep.py βββΊ train_windows.jsonl / val_windows.jsonl
|
| 26 |
+
β β’ Strip markdown, annotate link spans as [LINK_START]...[LINK_END]
|
| 27 |
+
β β’ Tokenize with DeBERTa, align labels to tokens
|
| 28 |
+
β β’ Sliding windows (512 tokens, stride 128)
|
| 29 |
+
β β’ 90/10 doc-level split
|
| 30 |
+
βΌ
|
| 31 |
+
train.py βββΊ model_link_token_cls/
|
| 32 |
+
β β’ Fine-tune microsoft/mdeberta-v3-base
|
| 33 |
+
β β’ Weighted cross-entropy (~25x for minority class)
|
| 34 |
+
β β’ 3 epochs, lr 2e-5, batch 16
|
| 35 |
+
βΌ
|
| 36 |
+
app.py βββΊ Streamlit UI
|
| 37 |
+
β’ Sliding-window inference (handles any text length)
|
| 38 |
+
β’ Word-level highlighting with confidence scores
|
| 39 |
+
```
|
| 40 |
+
|
| 41 |
+
## Data
|
| 42 |
+
|
| 43 |
+
Source: [blog.google](https://blog.google/) sitemap (The Keyword β Google's product and technology blog).
|
| 44 |
+
|
| 45 |
+
| Metric | Value |
|
| 46 |
+
|---|---|
|
| 47 |
+
| Pages scraped | 10,273 |
|
| 48 |
+
| Total tokens | 8.2M |
|
| 49 |
+
| Link tokens | 286,799 (3.48%) |
|
| 50 |
+
| Training windows | 21,264 |
|
| 51 |
+
| Validation windows | 2,402 |
|
| 52 |
+
|
| 53 |
+
The class imbalance (96.5% non-link vs 3.5% link) is handled with weighted cross-entropy loss during training.
|
| 54 |
+
|
| 55 |
+
## Model
|
| 56 |
+
|
| 57 |
+
- **Base**: `microsoft/mdeberta-v3-base` (DebertaV2ForTokenClassification)
|
| 58 |
+
- **Labels**: `O` (0), `LINK` (1)
|
| 59 |
+
- **Max position**: 512 tokens
|
| 60 |
+
- **Parameters**: 12 layers, 768 hidden, 12 attention heads
|
| 61 |
+
|
| 62 |
+
### Evaluation Results
|
| 63 |
+
|
| 64 |
+
| Metric | Value |
|
| 65 |
+
|---|---|
|
| 66 |
+
| Accuracy | 95.6% |
|
| 67 |
+
| Precision | 42.4% |
|
| 68 |
+
| Recall | 79.5% |
|
| 69 |
+
| F1 | 0.553 |
|
| 70 |
+
|
| 71 |
+
High recall means the model catches most link-worthy text. Lower precision reflects the inherent ambiguity β many words *could* be linked, so "false positives" are often reasonable candidates.
|
| 72 |
+
|
| 73 |
+
## Usage
|
| 74 |
+
|
| 75 |
+
### Streamlit App
|
| 76 |
+
|
| 77 |
+
```bash
|
| 78 |
+
streamlit run app.py
|
| 79 |
+
```
|
| 80 |
+
|
| 81 |
+
Paste text, adjust the confidence threshold, and see predicted link anchors highlighted in green.
|
| 82 |
+
|
| 83 |
+
### Python
|
| 84 |
+
|
| 85 |
+
```python
|
| 86 |
+
from transformers import AutoTokenizer, AutoModelForTokenClassification
|
| 87 |
+
import torch
|
| 88 |
+
import torch.nn.functional as F
|
| 89 |
+
|
| 90 |
+
tokenizer = AutoTokenizer.from_pretrained("model_link_token_cls")
|
| 91 |
+
model = AutoModelForTokenClassification.from_pretrained("model_link_token_cls")
|
| 92 |
+
model.eval()
|
| 93 |
+
|
| 94 |
+
text = "Google announced new features for Search and Gmail today."
|
| 95 |
+
enc = tokenizer(text, return_tensors="pt", return_offsets_mapping=True)
|
| 96 |
+
with torch.no_grad():
|
| 97 |
+
logits = model(input_ids=enc["input_ids"], attention_mask=enc["attention_mask"]).logits
|
| 98 |
+
probs = F.softmax(logits, dim=-1)[0, :, 1] # P(LINK) per token
|
| 99 |
+
|
| 100 |
+
for token, offset, p in zip(
|
| 101 |
+
tokenizer.convert_ids_to_tokens(enc["input_ids"][0]),
|
| 102 |
+
enc["offset_mapping"][0],
|
| 103 |
+
probs
|
| 104 |
+
):
|
| 105 |
+
if offset[0] == offset[1]:
|
| 106 |
+
continue # skip special tokens
|
| 107 |
+
if p > 0.5:
|
| 108 |
+
print(f" LINK: {text[offset[0]:offset[1]]} ({p:.2%})")
|
| 109 |
+
```
|
| 110 |
+
|
| 111 |
+
## Scripts
|
| 112 |
+
|
| 113 |
+
| File | Purpose |
|
| 114 |
+
|---|---|
|
| 115 |
+
| `scrape.py` | Async Playwright scraper; reads sitemap.xml, saves to SQLite + markdown files |
|
| 116 |
+
| `_prep.py` | Cleans markdown, annotates link spans, tokenizes, creates sliding windows |
|
| 117 |
+
| `train.py` | Fine-tunes DeBERTa with weighted loss, W&B tracking |
|
| 118 |
+
| `app.py` | Streamlit inference app with sliding-window support |
|
| 119 |
+
| `_count.py` | Token length analysis utility |
|
| 120 |
+
| `_detok.py` | Token ID decoder (Streamlit) |
|
| 121 |
+
|
| 122 |
+
## Requirements
|
| 123 |
+
|
| 124 |
+
- Python 3.8+
|
| 125 |
+
- PyTorch
|
| 126 |
+
- Transformers
|
| 127 |
+
- Playwright (for scraping)
|
| 128 |
+
- Streamlit (for inference app)
|