File size: 3,423 Bytes
138b150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
---
license: cc-by-4.0
pretty_name: MGnify Protein Catalogues
size_categories:
  - unknown
task_categories:
  - other
language:
  - en
tags:
  - biology
  - proteins
  - sequences
  - fasta
  - mgnify
  - metagenomics
---

# MGnify Protein Catalogues

Normalized FASTA shards of MGnify protein cluster catalogues (mgy_clusters and mgy_proteins partitions).

Processed and uploaded by the [MegaData](https://github.com/) post-download pipeline
(internal repo). Original source: <https://www.ebi.ac.uk/metagenomics/>.

> Note: `shard-000000.fasta.zst` per source is a 100-record sample left over from validation passes and is **not part of this dataset**. Only `shard-000001.fasta.zst` and higher contain full-clean data.

## Statistics

| | |
|---|---|
| Source files | 26 |
| Shards | 3,148 |
| Compressed shard bytes | 319.27 GiB (342,815,994,580) |
| Normalized table files | 28 |
| Compressed table bytes | 2.17 TiB (2,380,980,172,257) |

## Layout

```
.
├── tables/<source_slug>.jsonl         # normalized table rows (one JSON object per line)
└── sequences/<source_slug>/shard-NNNNNN.fasta.zst
```

> `metadata/`, `manifests/`, and `_MANIFEST.json` are intentionally not shipped here — the on-disk versions were sample-only. Recompute per-shard record/residue counts downstream by streaming the FASTA records out of the shipped shards.

`<source_slug>` corresponds 1:1 with an upstream source archive; e.g.
`sequence_uniprotkb_uniprot_sprot.fasta.gz`.

## Loading

Stream every shard of one source (replace `<source_slug>` with the directory of
interest under `sequences/`):

```bash
hf download LiteFold/Mgnify --repo-type dataset \
  --include 'sequences/<source_slug>/shard-*.fasta.zst' \
  --local-dir ./mgnify_proteins
zstd -dc ./mgnify_proteins/sequences/<source_slug>/shard-*.fasta.zst | head
```

Programmatic streaming with [`zstandard`](https://pypi.org/project/zstandard/):

```python
from huggingface_hub import snapshot_download
from pathlib import Path
import zstandard as zstd

local = snapshot_download(
    repo_id="LiteFold/Mgnify",
    repo_type="dataset",
    allow_patterns=["sequences/*/shard-*.fasta.zst"],
)

dctx = zstd.ZstdDecompressor()
for shard in sorted(Path(local).rglob("shard-*.fasta.zst")):
    with shard.open("rb") as f, dctx.stream_reader(f) as reader:
        buf = b""
        while chunk := reader.read(1 << 20):
            buf += chunk
            *lines, buf = buf.split(b"\n")
            for line in lines:
                ...  # naive splitter; swap in your FASTA parser
```

## License

CC BY 4.0 (EMBL-EBI MGnify).

## Citation

> Mitchell AL, et al. MGnify: the microbiome analysis resource in 2020. Nucleic Acids Research, 48(D1):D570-D578, 2020.

## Caveats

- `shard-000000.fasta.zst` per source is a 100-record sample left over from validation passes and is **excluded** from this upload. Only `shard-000001.fasta.zst` and onward are shipped.
- Per-source `manifests/*.json` and `metadata/*.records.jsonl` are **not shipped** because their on-disk contents are sample-only (100 records per source). They can be regenerated downstream by streaming the shipped shards (each shard is a concatenation of FASTA records; counting `^>` headers gives per-shard record counts).

## Provenance

Built from the local manifest entry `mgnify_proteins` of `manifests/atlas_download_plan.json`.
Pipeline source: `megadata-post normalize --dataset mgnify_proteins`.