SUTRA: Scalable Multilingual Language Model Architecture
Paper β’ 2405.06694 β’ Published β’ 38
How to use TWO/sutra-mlt256-v2 with Transformers:
# Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("TWO/sutra-mlt256-v2", dtype="auto")This is the model card of a π€ transformers model that has been pushed on the Hub. This model card has been automatically generated.
BibTeX:
@misc{bendale2023sutra,
author = {Abhijit Bendale and Michael Sapienza and Steven Ripplinger and Simon Gibbs and Jaewon Lee and Pranav Mistry},
title = {SUTRA: Scalable Multilingual Language Model Architecture},
howpublished = {arXiv preprint arXiv:2405.06694},
year = {2024}
}