Instructions to use Qsevent77/xlm-roberta-flash-implementation with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use Qsevent77/xlm-roberta-flash-implementation with Transformers:
# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("Qsevent77/xlm-roberta-flash-implementation", dtype="auto") - Notebooks
- Google Colab
- Kaggle
Core implementation of Jina XLM-RoBERTa
This implementation is adapted from XLM-Roberta. In contrast to the original implementation, this model uses Rotary positional encodings and supports flash-attention 2.
Models that use this implementation
Converting weights
Weights from an original XLMRoberta model can be converted using the convert_roberta_weights_to_flash.py script in the model repository.
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support