Upload marion.json with huggingface_hub
Browse files- marion.json +6 -0
marion.json
CHANGED
|
@@ -310,5 +310,11 @@
|
|
| 310 |
"query":"What is the conventional workflow for BERT ?",
|
| 311 |
"target_page":1,
|
| 312 |
"target_passage":"The conventional workflow for BERT consists of two stages: pre-training and fine-tuning. "
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 313 |
}
|
| 314 |
]
|
|
|
|
| 310 |
"query":"What is the conventional workflow for BERT ?",
|
| 311 |
"target_page":1,
|
| 312 |
"target_passage":"The conventional workflow for BERT consists of two stages: pre-training and fine-tuning. "
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"source_file":"arxiv2_taclccby4_license.pdf",
|
| 316 |
+
"query":"Is syntaxis encoded with Bert model ?",
|
| 317 |
+
"target_page":2,
|
| 318 |
+
"target_passage":" As far as how syntaxis represented, it seems that syntactic structure is not directly encoded in self-attention weights."
|
| 319 |
}
|
| 320 |
]
|