tokenizer-metadata / README.md
christopher's picture
Upload dataset
c932ee6 verified
---
dataset_info:
features:
- name: _id
dtype: large_string
- name: id
dtype: large_string
- name: created_at
dtype: timestamp[us, tz=UTC]
- name: downloads_all_time
dtype: int64
- name: downloads
dtype: int64
- name: likes
dtype: int64
- name: trending_score
dtype: float64
- name: hash
dtype: large_string
- name: vocab_hash
dtype: large_string
- name: vocab_size
dtype: int64
- name: model_type
dtype: large_string
- name: num_merges
dtype: float64
- name: has_normalizer
dtype: bool
- name: has_pre_tokenizer
dtype: bool
- name: has_post_processor
dtype: bool
- name: has_decoder
dtype: bool
- name: num_added_tokens
dtype: int64
- name: normalizer_type
dtype: large_string
- name: pre_tokenizer_type
dtype: large_string
- name: decoder_type
dtype: large_string
- name: normalizer_types
list: string
- name: pre_tokenizer_types
list: string
- name: decoder_types
list: string
- name: tokenizer_id
dtype: int64
splits:
- name: train
num_bytes: 135916939
num_examples: 423650
download_size: 50377312
dataset_size: 135916939
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---