from datasets import load_dataset import numpy as np import unicodedata dataset = load_dataset("Fizzarolli/wattpad2", "default", split="train") from lingua import LanguageDetectorBuilder detector = LanguageDetectorBuilder.from_all_languages().with_preloaded_language_models().build() def fix_language_column(example): res = detector.compute_language_confidence_values(unicodedata.normalize('NFKD', example["description"])) example["language"] = res[0].language.iso_code_639_3.name confidences = list(map(lambda x: x.value, res)) example["language_confidence"] = np.max(confidences) return example dataset = dataset.map(fix_language_column) dataset.push_to_hub("Fizzarolli/wattpad2", "default")