--- license: cc-by-4.0 pretty_name: MGnify Protein Catalogues size_categories: - unknown task_categories: - other language: - en tags: - biology - proteins - sequences - fasta - mgnify - metagenomics --- # MGnify Protein Catalogues Normalized FASTA shards of MGnify protein cluster catalogues (mgy_clusters and mgy_proteins partitions). Processed and uploaded by the [MegaData](https://github.com/) post-download pipeline (internal repo). Original source: . > Note: `shard-000000.fasta.zst` per source is a 100-record sample left over from validation passes and is **not part of this dataset**. Only `shard-000001.fasta.zst` and higher contain full-clean data. ## Statistics | | | |---|---| | Source files | 26 | | Shards | 3,148 | | Compressed shard bytes | 319.27 GiB (342,815,994,580) | | Normalized table files | 28 | | Compressed table bytes | 2.17 TiB (2,380,980,172,257) | ## Layout ``` . ├── tables/.jsonl # normalized table rows (one JSON object per line) └── sequences//shard-NNNNNN.fasta.zst ``` > `metadata/`, `manifests/`, and `_MANIFEST.json` are intentionally not shipped here — the on-disk versions were sample-only. Recompute per-shard record/residue counts downstream by streaming the FASTA records out of the shipped shards. `` corresponds 1:1 with an upstream source archive; e.g. `sequence_uniprotkb_uniprot_sprot.fasta.gz`. ## Loading Stream every shard of one source (replace `` with the directory of interest under `sequences/`): ```bash hf download LiteFold/Mgnify --repo-type dataset \ --include 'sequences//shard-*.fasta.zst' \ --local-dir ./mgnify_proteins zstd -dc ./mgnify_proteins/sequences//shard-*.fasta.zst | head ``` Programmatic streaming with [`zstandard`](https://pypi.org/project/zstandard/): ```python from huggingface_hub import snapshot_download from pathlib import Path import zstandard as zstd local = snapshot_download( repo_id="LiteFold/Mgnify", repo_type="dataset", allow_patterns=["sequences/*/shard-*.fasta.zst"], ) dctx = zstd.ZstdDecompressor() for shard in sorted(Path(local).rglob("shard-*.fasta.zst")): with shard.open("rb") as f, dctx.stream_reader(f) as reader: buf = b"" while chunk := reader.read(1 << 20): buf += chunk *lines, buf = buf.split(b"\n") for line in lines: ... # naive splitter; swap in your FASTA parser ``` ## License CC BY 4.0 (EMBL-EBI MGnify). ## Citation > Mitchell AL, et al. MGnify: the microbiome analysis resource in 2020. Nucleic Acids Research, 48(D1):D570-D578, 2020. ## Caveats - `shard-000000.fasta.zst` per source is a 100-record sample left over from validation passes and is **excluded** from this upload. Only `shard-000001.fasta.zst` and onward are shipped. - Per-source `manifests/*.json` and `metadata/*.records.jsonl` are **not shipped** because their on-disk contents are sample-only (100 records per source). They can be regenerated downstream by streaming the shipped shards (each shard is a concatenation of FASTA records; counting `^>` headers gives per-shard record counts). ## Provenance Built from the local manifest entry `mgnify_proteins` of `manifests/atlas_download_plan.json`. Pipeline source: `megadata-post normalize --dataset mgnify_proteins`.