Update README.md
Browse filesAdded content embedding section
README.md
CHANGED
|
@@ -41,6 +41,28 @@ Dataset Statistics:
|
|
| 41 |
<!-- | May 2025 | 5,833,993 | 87,752,862 | 1 | 30.08 | 1,581,282 | 17,683 | 2.6e-06 | -->
|
| 42 |
<!-- | June 2025 | 9,974,275 | 152,449,542 | 1 | 30.57 | 3,381,364 | 25,447 | 1.5e06 | -->
|
| 43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
### Resources
|
| 45 |
|
| 46 |
<!-- Provide the basic links for the dataset. -->
|
|
|
|
| 41 |
<!-- | May 2025 | 5,833,993 | 87,752,862 | 1 | 30.08 | 1,581,282 | 17,683 | 2.6e-06 | -->
|
| 42 |
<!-- | June 2025 | 9,974,275 | 152,449,542 | 1 | 30.57 | 3,381,364 | 25,447 | 1.5e06 | -->
|
| 43 |
|
| 44 |
+
**Content Embedding:**
|
| 45 |
+
|
| 46 |
+
Domain-level content embeddings are generated using multiple LLM-based embedding models with varying LLM-model sizes and embedding dimensions.
|
| 47 |
+
The embeddings are intended to support feature initialization for downstream GNN models.
|
| 48 |
+
For each domain, the textual content is first extracted and then encoded into dense vector representations using the selected embedding model.
|
| 49 |
+
|
| 50 |
+
The dataset is organized by month under the content_embeddings directory.
|
| 51 |
+
|
| 52 |
+
Each pickled file stores a dictionary:
|
| 53 |
+
```
|
| 54 |
+
{ domain1:[[page_url1, embedding_vector1],[page_url2, embedding_vector2], ...],
|
| 55 |
+
domain2:[[page_url1, embedding_vector1],[page_url2, embedding_vector2], ...],
|
| 56 |
+
...
|
| 57 |
+
}
|
| 58 |
+
```
|
| 59 |
+
|
| 60 |
+
| Month | Embedding-model | Emb-dim| Total-files-size|
|
| 61 |
+
| -- | -- | -- | -- |
|
| 62 |
+
| October 2024 | embeddinggemma-300m | 256 | 30GB|
|
| 63 |
+
| November 2024 | embeddinggemma-300m | 256 | 30GB|
|
| 64 |
+
| December 2024 | embeddinggemma-300m | 256 | 30GB|
|
| 65 |
+
|
| 66 |
### Resources
|
| 67 |
|
| 68 |
<!-- Provide the basic links for the dataset. -->
|