tokenizer-262k / tokenizer.model
timpal0l's picture
Upload 262k vocab SentencePiece BPE tokenizer (trained on 175GB)
f1d2f3d verified
This file is stored with Xet . It is too big to display, but you can still download it.

Large File Pointer Details

( Raw pointer file )
SHA256:
ac214401227404105817141c21c6693e3b17c878465a42765d2c345310e003ce
Pointer size:
132 Bytes
·
Size of remote file:
4.88 MB
·
Xet hash:
cf3c8d57733c14529cc6e6eb1f6a49d7cf3e66c2dd236694b1008dd73685b082

Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.