from datasets import load_dataset from transformers import AutoTokenizer def main(): ds = load_dataset("takara-ai/micropajama", split="train") tok = AutoTokenizer.from_pretrained("BAAI/bge-large-en-v1.5") lens = ds.map(lambda b: {"len": [len(x) for x in tok(b["text"], add_special_tokens=False).input_ids]}, batched=True, remove_columns=ds.column_names) print(sum(lens["len"])) if __name__ == "__main__": main()