Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
stefan-it
/
ModernBERT-large-tokenizer-fix
like
2
Fill-Mask
Transformers
PyTorch
ONNX
Safetensors
English
modernbert
masked-lm
long-context
arxiv:
2412.13663
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
Deploy
Use this model
main
ModernBERT-large-tokenizer-fix
7.86 GB
7 contributors
History:
32 commits
stefan-it
docs: revert fix
3327b69
verified
7 months ago
onnx
Upload ONNX weights (#1)
about 1 year ago
.gitattributes
1.52 kB
initial commit
about 1 year ago
README.md
9.02 kB
docs: revert fix
7 months ago
config.json
1.19 kB
Bump `max_position_embeddings` to 8192
about 1 year ago
model.safetensors
1.58 GB
xet
Purge duplicate "decoder.weight", rely on tied weights instead
about 1 year ago
pytorch_model.bin
1.58 GB
xet
Purge duplicate "decoder.weight", rely on tied weights instead
about 1 year ago
special_tokens_map.json
694 Bytes
Also update tokenizer/special_tokens_map
about 1 year ago
tokenizer.json
2.13 MB
Also update tokenizer/special_tokens_map
about 1 year ago
tokenizer_config.json
20.8 kB
fix: also use `add_prefix_space = True` in tokenizer config
9 months ago