Instructions to use vishalgimhan/uber-assistant with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use vishalgimhan/uber-assistant with Transformers:
# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("vishalgimhan/uber-assistant", dtype="auto") - Notebooks
- Google Colab
- Kaggle
File size: 325 Bytes
a499caf | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | {
"backend": "tokenizers",
"bos_token": "<|begin_of_text|>",
"clean_up_tokenization_spaces": true,
"eos_token": "<|eot_id|>",
"is_local": false,
"model_input_names": [
"input_ids",
"attention_mask"
],
"model_max_length": 131072,
"pad_token": "<|eot_id|>",
"tokenizer_class": "TokenizersBackend"
}
|