General
This is a variation of the SHARE-4B-Base model that where context length was extended to 8192 tokens with 5000 documents. For all relevant details on the model, plese check the SHARE-4B-Base model card: https://huggingface.co/Joaoffg/SHARE-4B-Base-2604
Recommended use
We recommend using this version of the model exclusively for tasks that require a longer context length, and might show lower performance on short context tasks. Therefore, we advise use of the latter for most us cases.
- Downloads last month
- 3
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support