model update
#1
by TheWeeeed - opened
- .gitattributes +0 -1
- README.md +1 -5
- all_results.json +0 -4
- eval_nbest_predictions.json +0 -3
- eval_predictions.json +0 -0
- model.safetensors +1 -1
- tokenizer.json +1 -6
.gitattributes
CHANGED
|
@@ -33,4 +33,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
-
eval_nbest_predictions.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
README.md
CHANGED
|
@@ -18,8 +18,4 @@ pipeline_tag: question-answering
|
|
| 18 |
* **相關項目/GitHub**: https://github.com/YuTsyh/Chinese-Extractive-Question-Answering-QA-.git
|
| 19 |
* **相關模型**:
|
| 20 |
* TheWeeeed/chinese-paragraph-selector
|
| 21 |
-
* TheWeeeed/chinese-extractive-qa
|
| 22 |
-
|
| 23 |
-
## 更新紀錄
|
| 24 |
-
* **20/06/25**:更新模型
|
| 25 |
-
* **02/06/25**:更新模型
|
|
|
|
| 18 |
* **相關項目/GitHub**: https://github.com/YuTsyh/Chinese-Extractive-Question-Answering-QA-.git
|
| 19 |
* **相關模型**:
|
| 20 |
* TheWeeeed/chinese-paragraph-selector
|
| 21 |
+
* TheWeeeed/chinese-extractive-qa
|
|
|
|
|
|
|
|
|
|
|
|
all_results.json
DELETED
|
@@ -1,4 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"eval_exact_match": 78.36490528414755,
|
| 3 |
-
"eval_f1": 78.36490528414755
|
| 4 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
eval_nbest_predictions.json
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:9651879e5185fb7326df5dd59c4115d63a4c2e9ea5a41b31db4b33e80db51a73
|
| 3 |
-
size 13720443
|
|
|
|
|
|
|
|
|
|
|
|
eval_predictions.json
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 406737680
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:590a8b1bc317dc7b6806b71dbd44e423dd9ae1029ef54bc8cd7989ade0c4ad95
|
| 3 |
size 406737680
|
tokenizer.json
CHANGED
|
@@ -1,11 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
-
"truncation":
|
| 4 |
-
"direction": "Right",
|
| 5 |
-
"max_length": 512,
|
| 6 |
-
"strategy": "OnlySecond",
|
| 7 |
-
"stride": 128
|
| 8 |
-
},
|
| 9 |
"padding": null,
|
| 10 |
"added_tokens": [
|
| 11 |
{
|
|
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
+
"truncation": null,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
"padding": null,
|
| 5 |
"added_tokens": [
|
| 6 |
{
|