Datasets:

Add link to paper

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +13 -13
README.md CHANGED
@@ -1,25 +1,25 @@
1
  ---
2
  language:
3
- - en
4
  license: cc-by-4.0
5
  multilinguality: monolingual
6
  task_categories:
7
- - sentence-similarity
8
- - feature-extraction
9
- tags:
10
- - sonar
11
- - speech-embeddings
12
- - text-embeddings
13
- - common-voice
14
- - interpretability
15
  pretty_name: FLiP-data
 
 
 
 
 
 
16
  ---
17
 
18
  # FLiP-data
19
 
20
- Preprocessed data for the [FLiP](https://github.com/BUTSpeechFIT/FLiP) project — **Factorized Linear Projection for Interpreting Multimodal Multilingual Sentence Embeddings**.
21
 
22
- FLiP trains a factorized log-linear model to recover lexical content (keywords) from pretrained sentence embeddings via a single linear projection, with no fine-tuning of the encoder.
23
 
24
  ## Contents
25
 
@@ -57,10 +57,10 @@ train_text = np.load("cv_15/en/sonar_embeddings/train_text_embs.npy")
57
  ```bibtex
58
  @misc{kesiraju2026flip,
59
  title = {{FLiP}: Towards understanding and interpreting multimodal multilingual sentence embeddings},
60
- author = {Kesiraju, Santosh and Yusuf, Bolaji and Sedl{\'a}{\v{c}}ek, Simon and Plchot, Old{\v{r}}ich and Schwarz, Petr},
61
  year = {2026},
62
  eprint = {2026.XXXXX},
63
  archivePrefix = {arXiv},
64
  primaryClass = {cs.CL}
65
  }
66
- ```
 
1
  ---
2
  language:
3
+ - en
4
  license: cc-by-4.0
5
  multilinguality: monolingual
6
  task_categories:
7
+ - sentence-similarity
8
+ - feature-extraction
 
 
 
 
 
 
9
  pretty_name: FLiP-data
10
+ tags:
11
+ - sonar
12
+ - speech-embeddings
13
+ - text-embeddings
14
+ - common-voice
15
+ - interpretability
16
  ---
17
 
18
  # FLiP-data
19
 
20
+ Preprocessed data for the paper [FLiP: Towards understanding and interpreting multimodal multilingual sentence embeddings](https://huggingface.co/papers/2604.18109).
21
 
22
+ FLiP trains a factorized log-linear model to recover lexical content (keywords) from pretrained sentence embeddings via a single linear projection, with no fine-tuning of the encoder. The project code is available on [GitHub](https://github.com/BUTSpeechFIT/FLiP).
23
 
24
  ## Contents
25
 
 
57
  ```bibtex
58
  @misc{kesiraju2026flip,
59
  title = {{FLiP}: Towards understanding and interpreting multimodal multilingual sentence embeddings},
60
+ author = {Kesiraju, Santosh and Yusuf, Bolaji and Sedl{\'{a}}{\v{c}}ek, Simon and Plchot, Old{\v{r}}ich and Schwarz, Petr},
61
  year = {2026},
62
  eprint = {2026.XXXXX},
63
  archivePrefix = {arXiv},
64
  primaryClass = {cs.CL}
65
  }
66
+ ```