| --- |
| dataset_info: |
| features: |
| - name: filename |
| dtype: string |
| - name: chunks |
| dtype: string |
| - name: repo_name |
| dtype: string |
| splits: |
| - name: train |
| num_bytes: 111941 |
| num_examples: 251 |
| download_size: 37308 |
| dataset_size: 111941 |
| configs: |
| - config_name: default |
| data_files: |
| - split: train |
| path: data/train-* |
| --- |
| |
|
|
| ## Dataset info |
|
|
| This dataset contains documentation chunks from repositories (ADD REPOS). |
|
|
| ### Postprocessing |
|
|
| After some inspection, some chunks contain text too short to be meaningful, so we decided to remove those by removing chunks whose number of tokens (computed |
| with the same tokenizer of the model to be used for the embeddings) is lower or equal to the 5%: |
|
|
| ```python |
| from transformers import AutoTokenizer |
| tokenizer = AutoTokenizer.from_pretrained("BAAI/bge-base-en-v1.5") |
| |
| df = ds.to_pandas() |
| df["token_length"] = df["chunks"].apply(lambda x: len(tokenizer.encode(x))) |
| |
| df_short = df[df["token_length"] >= df["token_length"].quantile(0.05)] |
| ds = Dataset.from_pandas(df_short[["filename", "chunks", "repo_name"]], preserve_index=False) |
| ``` |