Update README.md
Browse files
README.md
CHANGED
|
@@ -183,7 +183,7 @@ bert-chunker-3 (prob_threshold=0.50543) | N/A | 0 | 90.4 ± 28.7 | 3.3 ± 3.1 |
|
|
| 183 |
★ bert-chunker-3.5 | <= 200 | 0 | 90.4 ± 26.2 | 7.7 ± 5.7 | 29.2 ± 17.9 | 7.6 ± 5.7 |**O(N)** | **Yes**
|
| 184 |
## Future
|
| 185 |
This model is undertrained due to lack of money and laziness. I observed it is still undertrained for two reasons:
|
| 186 |
-
- I
|
| 187 |
- The outputs from two non-overlapping windows show poor comparability in split point probabilities, which will undermines the performance when max_tokens_per_chunk is large. I think more data will amplify probability difference and enhance their comparability. This was corroborated by some of my experiments. So next version will probably just be more data.
|
| 188 |
## Citation
|
| 189 |
```bibtex
|
|
|
|
| 183 |
★ bert-chunker-3.5 | <= 200 | 0 | 90.4 ± 26.2 | 7.7 ± 5.7 | 29.2 ± 17.9 | 7.6 ± 5.7 |**O(N)** | **Yes**
|
| 184 |
## Future
|
| 185 |
This model is undertrained due to lack of money and laziness. I observed it is still undertrained for two reasons:
|
| 186 |
+
- I trained it on 110 K window data for 2~3 epoch and the eval loss kept decreasing.
|
| 187 |
- The outputs from two non-overlapping windows show poor comparability in split point probabilities, which will undermines the performance when max_tokens_per_chunk is large. I think more data will amplify probability difference and enhance their comparability. This was corroborated by some of my experiments. So next version will probably just be more data.
|
| 188 |
## Citation
|
| 189 |
```bibtex
|