Update README.md
Browse files
README.md
CHANGED
|
@@ -10,43 +10,47 @@ pipeline_tag: feature-extraction
|
|
| 10 |
|
| 11 |
# ARC-Encoder models
|
| 12 |
|
| 13 |
-
This page houses `ARC8-Encoder_Mistral` from three different versions of pretrained ARC-Encoders. Architectures and methods to train them are described in the paper *ARC-Encoder: learning compressed text representations for large language models* available [here](https://arxiv.org/abs/2510.20535).
|
|
|
|
| 14 |
|
| 15 |
-
##
|
| 16 |
-
First, use the following code to load the released models and format the folders accurately in your `<TMP_PATH>`. You just need to perform it once per model:
|
| 17 |
-
```python
|
| 18 |
-
from embed_llm.models.augmented_model import load_and_save_released_models
|
| 19 |
-
|
| 20 |
-
# Example for ARC8_Encoder_Mistral, other options include "ARC8_Encoder_Llama" or "ARC8_Encoder_multi"
|
| 21 |
-
load_and_save_released_models("ARC8_Encoder_Mistral", hf_token="<YOUR_HF_TOKEN>")
|
| 22 |
-
```
|
| 23 |
-
*Remark:* This code snippet loads the model from Hugging Face and then creates the appropriate folder at `<TMP_PATH>` containing the checkpoint and additional necessary files to perform finetuning or evaluation with this codebase. To reduce the occupied memory space, you can then delete the model from your Hugging Face cache.
|
| 24 |
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
All the encoders released here are trained on web crawl filtered using [Dactory](https://github.com/kyutai-labs/dactory) based on a [Llama3.2-3B](https://github.com/meta-llama/llama-cookbook) base backbone. It consists in two ARC-Encoder specifically trained for one decoder and one for two decoders in the same time:
|
| 28 |
- `ARC8-Encoder_Llama`, trained on 2.6B tokens on [Llama3.1-8B](https://github.com/meta-llama/llama-cookbook) base specifically with a pooling factor of 8.
|
| 29 |
-
- `ARC8-Encoder_Mistral`, trained on 2.6B tokens on [Mistral-7B](https://
|
| 30 |
- `ARC8-Encoder_multi`, trained by sampling among the two decoders with a pooling factor of 8.
|
| 31 |
|
| 32 |
-
### Uses
|
| 33 |
|
| 34 |
-
As described in the [paper](https://arxiv.org/abs/2510.20535), the pretrained ARC-Encoders can be fine-tuned to perform various downstream tasks.
|
| 35 |
You can also adapt an ARC-Encoder to a new pooling factor (PF) by fine-tuning it on the desired PF.
|
| 36 |
For optimal results, we recommend fine-tuning toward a lower PF than the one used during pretraining.
|
| 37 |
To reproduce the results presented in the paper, you can use our released fine-tuning dataset, [ARC_finetuning](https://huggingface.co/datasets/kyutai/ARC_finetuning).
|
| 38 |
|
| 39 |
-
### Licensing
|
| 40 |
|
| 41 |
-
ARC-Encoders are licensed under the CC-BY 4.0 license.
|
| 42 |
|
| 43 |
Terms of use: As the released models are pretrained from Llama3.2 3B backbone, ARC-Encoders are subject to the Llama Terms of Use found at [Llama license](https://www.llama.com/license/).
|
| 44 |
|
| 45 |
-
##
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
|
| 47 |
-
If you use one of these models, please cite:
|
| 48 |
|
| 49 |
-
```bibtex
|
| 50 |
@misc{pilchen2025arcencoderlearningcompressedtext,
|
| 51 |
title={ARC-Encoder: learning compressed text representations for large language models},
|
| 52 |
author={Hippolyte Pilchen and Edouard Grave and Patrick Pérez},
|
|
|
|
| 10 |
|
| 11 |
# ARC-Encoder models
|
| 12 |
|
| 13 |
+
This page houses `ARC8-Encoder_Mistral` from three different versions of pretrained ARC-Encoders. Architectures and methods to train them are described in the paper *ARC-Encoder: learning compressed text representations for large language models* available [here](https://arxiv.org/abs/2510.20535).
|
| 14 |
+
Code: [ARC-Encoder repository](https://github.com/kyutai-labs/ARC-Encoder)
|
| 15 |
|
| 16 |
+
## Models Details
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
+
All the encoders released here are trained on web crawl filtered using [Dactory](https://github.com/kyutai-labs/dactory) based on a [Llama3.2-3B](https://github.com/meta-llama/llama-cookbook) base backbone. It consists in two ARC-Encoder specifically trained for one decoder and one for two decoders in the same time:
|
|
|
|
|
|
|
| 19 |
- `ARC8-Encoder_Llama`, trained on 2.6B tokens on [Llama3.1-8B](https://github.com/meta-llama/llama-cookbook) base specifically with a pooling factor of 8.
|
| 20 |
+
- `ARC8-Encoder_Mistral`, trained on 2.6B tokens on [Mistral-7B](https://www.mistralai.com/news/announcing-mistral-7b/) base specifically with a pooling factor of 8.
|
| 21 |
- `ARC8-Encoder_multi`, trained by sampling among the two decoders with a pooling factor of 8.
|
| 22 |
|
| 23 |
+
### Uses
|
| 24 |
|
| 25 |
+
As described in the [paper](https://arxiv.org/abs/2510.20535), the pretrained ARC-Encoders can be fine-tuned to perform various downstream tasks.
|
| 26 |
You can also adapt an ARC-Encoder to a new pooling factor (PF) by fine-tuning it on the desired PF.
|
| 27 |
For optimal results, we recommend fine-tuning toward a lower PF than the one used during pretraining.
|
| 28 |
To reproduce the results presented in the paper, you can use our released fine-tuning dataset, [ARC_finetuning](https://huggingface.co/datasets/kyutai/ARC_finetuning).
|
| 29 |
|
| 30 |
+
### Licensing
|
| 31 |
|
| 32 |
+
ARC-Encoders are licensed under the CC-BY 4.0 license.
|
| 33 |
|
| 34 |
Terms of use: As the released models are pretrained from Llama3.2 3B backbone, ARC-Encoders are subject to the Llama Terms of Use found at [Llama license](https://www.llama.com/license/).
|
| 35 |
|
| 36 |
+
## Usage
|
| 37 |
+
|
| 38 |
+
To load the pre-trained ARC-Encoders, use the following code snippet from the [ARC-Encoder repository](https://github.com/kyutai-labs/ARC-Encoder):
|
| 39 |
+
|
| 40 |
+
```python
|
| 41 |
+
from embed_llm.models.augmented_model import load_and_save_released_models
|
| 42 |
+
|
| 43 |
+
# ARC8_Encoder_multi, ARC8_Encoder_Llama or ARC8_Encoder_Mistral
|
| 44 |
+
load_and_save_released_models(ARC8_Encoder_Mistral, hf_token=<HF_TOKEN>)
|
| 45 |
+
```
|
| 46 |
+
|
| 47 |
+
***Remark:*** This code snippet loads the model from Hugging Face and then creates appropriate folders at `<TMP_PATH>` containing the checkpoint and additional necessary files for fine-tuning or evaluation with the `ARC-Encoder` codebase. To reduce occupied memory space, you can then delete the model from your Hugging Face cache.
|
| 48 |
+
|
| 49 |
+
## Citations
|
| 50 |
|
| 51 |
+
If you use one of these models, please cite:
|
| 52 |
|
| 53 |
+
```bibtex
|
| 54 |
@misc{pilchen2025arcencoderlearningcompressedtext,
|
| 55 |
title={ARC-Encoder: learning compressed text representations for large language models},
|
| 56 |
author={Hippolyte Pilchen and Edouard Grave and Patrick Pérez},
|