--- license: apache-2.0 configs: - config_name: default data_files: - split: train path: data/train-* - split: valid path: data/valid-* - split: test path: data/test-* dataset_info: features: - name: sequence dtype: string - name: length dtype: int64 splits: - name: train num_bytes: 12071713186 num_examples: 41546293 - name: valid num_bytes: 24293086 num_examples: 82929 - name: test num_bytes: 19981814 num_examples: 48941 download_size: 11690105266 dataset_size: 12115988086 --- # Uniref50: Uniref Sequences clustered at 50% sequence identity - ~40M Protein Sequences. - Split into train val and test. # Usage ``` from datasets import load_dataset # Step 1: Load the dataset from HuggingFace Hub dataset = load_dataset("zhangzhi/Uniref50") # Step 2: Access a specific split (e.g., "train", "validation", "test") train_split = dataset["train"] print(f"Number of sequences in the train split: {len(train_split)}") ```