Instructions to use radna/mini_intern_chat_triton with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use radna/mini_intern_chat_triton with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("visual-question-answering", model="radna/mini_intern_chat_triton", trust_remote_code=True)# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("radna/mini_intern_chat_triton", trust_remote_code=True, dtype="auto") - Notebooks
- Google Colab
- Kaggle
Commit History
Update modeling_internvl_chat.py 664bae3 verified
Update modeling_internvl_chat.py 01488b9 verified
Update modeling_internvl_chat.py 05b2c8b verified
remove formatting e065dd1
HGB commited on
add bert padding + modify internVit classes 9c8bb9e
HGB commited on
Initial commit for testing triton flash attention e9fd7b3
HGB commited on