Update README.md
Browse files
README.md
CHANGED
|
@@ -81,8 +81,8 @@ These inputs are short enough to fit within the model’s 512-token limit — no
|
|
| 81 |
|
| 82 |
The cited-paper encoder was trained jointly with the query-talk encoder under a **dual-encoder contrastive framework** inspired by Dense Passage Retrieval (Karpukhin et al., 2020).
|
| 83 |
|
| 84 |
-
Each talk
|
| 85 |
-
Their dot-product similarity
|
| 86 |
|
| 87 |
$$
|
| 88 |
L = - \sum_i [y_i \log \sigma(s_i) + (1 - y_i)\log(1 - \sigma(s_i))]
|
|
|
|
| 81 |
|
| 82 |
The cited-paper encoder was trained jointly with the query-talk encoder under a **dual-encoder contrastive framework** inspired by Dense Passage Retrieval (Karpukhin et al., 2020).
|
| 83 |
|
| 84 |
+
Each talk $Ti$ and paper $Rj$ is encoded into embeddings $fT(Ti)$ and $fR(Rj)$.
|
| 85 |
+
Their dot-product similarity $sij = fT(Ti)·fR(Rj)$ is optimized using a sigmoid-based binary loss supporting multiple positives per query:
|
| 86 |
|
| 87 |
$$
|
| 88 |
L = - \sum_i [y_i \log \sigma(s_i) + (1 - y_i)\log(1 - \sigma(s_i))]
|