--- license: apache-2.0 datasets: - OmniAICreator/Japanese-Novels-23M - NilanE/ParallelFiction-Ja_En-100k - globis-university/aozorabunko-clean - joujiboi/Galgame-VisualNovel-Reupload - CC100 - AnimeText language: - ja pipeline_tag: fill-mask library_name: transformers --- # Custom Japanese BERT (4-layer) This model is a tiny Japanese BERT model with 4 layers, optimized for speed. ## Model Background - **Architecture:** BERT (4 layers, 256 hidden size, 4 heads, 1024 FFN) - **Distillation:** Distilled from a fine-tuned version of `tohoku-nlp/bert-base-japanese-char-v2`. - **Initialization:** The student model was randomly initialized. - **Tokenizer:** Japanese Character-level tokenizer, shared with the teacher.