test-tokenizers / README.md
christopher's picture
Upload dataset
8872ef7 verified
metadata
dataset_info:
  - config_name: metadata
    features:
      - name: _id
        dtype: string
      - name: id
        dtype: string
      - name: created_at
        dtype: timestamp[us, tz=UTC]
      - name: downloads_all_time
        dtype: int64
      - name: downloads
        dtype: int64
      - name: likes
        dtype: int64
      - name: trending_score
        dtype: float64
      - name: tokenizer_id
        dtype: int64
      - name: hash
        dtype: string
    splits:
      - name: batch_0
        num_bytes: 13156
        num_examples: 100
      - name: batch_1
        num_bytes: 12911
        num_examples: 100
      - name: batch_2
        num_bytes: 13009
        num_examples: 100
      - name: batch_3
        num_bytes: 13092
        num_examples: 100
      - name: batch_4
        num_bytes: 13282
        num_examples: 100
    download_size: 65425
    dataset_size: 65450
  - config_name: tokenizers
    features:
      - name: tokenizer_id
        dtype: int64
      - name: hash
        dtype: string
      - name: vocab_hash
        dtype: string
      - name: vocab_size
        dtype: int64
      - name: model_type
        dtype: string
      - name: num_merges
        dtype: int64
      - name: has_normalizer
        dtype: bool
      - name: has_pre_tokenizer
        dtype: bool
      - name: has_post_processor
        dtype: bool
      - name: has_decoder
        dtype: bool
      - name: num_added_tokens
        dtype: int64
      - name: normalizer_type
        dtype: string
      - name: pre_tokenizer_type
        dtype: string
      - name: decoder_type
        dtype: string
      - name: normalizer_types
        list: string
      - name: pre_tokenizer_types
        list: string
      - name: decoder_types
        list: string
      - name: version
        dtype: string
      - name: added_tokens
        dtype: string
      - name: normalizer
        dtype: string
      - name: pre_tokenizer
        dtype: string
      - name: post_processor
        dtype: string
      - name: decoder
        dtype: string
      - name: model
        dtype: string
    splits:
      - name: train
        num_bytes: 640975277
        num_examples: 195
    download_size: 359764431
    dataset_size: 640975277
configs:
  - config_name: metadata
    data_files:
      - split: batch_0
        path: metadata/batch_0-*
      - split: batch_1
        path: metadata/batch_1-*
      - split: batch_2
        path: metadata/batch_2-*
      - split: batch_3
        path: metadata/batch_3-*
      - split: batch_4
        path: metadata/batch_4-*
  - config_name: tokenizers
    data_files:
      - split: train
        path: tokenizers/train-*