checkpoint = "google/mt5-small" tokenizer = MT5Tokenizer.from_pretrained(checkpoint, legacy=False) model = MT5ForConditionalGeneration.from_pretarined(checkpoint)

This model is katakana-oriented. CharacTER score = 0.1677

Training

  • Trainer: Seq2seqTrainer
  • Epochs: 200
  • Optimizer: AdamW
  • Learning Rate: 2e-5
  • Weight decay: 0.01
  • Warm-ups: 0.05*total_steps
  • Scheduler: cosine
Downloads last month
29
Safetensors
Model size
0.3B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for kinlas/Gunma-to-hyojun-converter

Quantizations
1 model