General

This is a variation of the SHARE-4B-Base model that where context length was extended to 8192 tokens with 5000 documents. For all relevant details on the model, plese check the SHARE-4B-Base model card: https://huggingface.co/Joaoffg/SHARE-4B-Base-2604

Recommended use

We recommend using this version of the model exclusively for tasks that require a longer context length, and might show lower performance on short context tasks. Therefore, we advise use of the latter for most us cases.

Downloads last month
3
Safetensors
Model size
2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support