File size: 876 Bytes
2f5fb67
 
 
 
 
 
 
 
 
303f1d0
 
 
d035106
d10b95c
374554f
c0f8df8
ad8e05c
449f9fb
ad8e05c
4d081f3
4c8f2b4
4d081f3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
---
title: README
emoji: 😻
colorFrom: purple
colorTo: gray
sdk: static
pinned: false
---

# Legend:

- **pruned** datasets are reduced in size to contain only examples in which the source paragraph and the target paragraph are not identical
- **hard** datasets have their training and test split created from separate pools of books with no overlap (so all paragraphs from a given book are contained in only a single split)
- **transduced** datasets have their training split processed by a rule-based normalizer

Models were accordingly created based on the 4 dataset variants. 

# Evaluation repositories:

[https://github.com/kedudzic/pl-normalizer-evaluation](https://github.com/kedudzic/pl-normalizer-evaluation) (private)

[https://github.com/kedudzic/pl-normalizer-evaluation-just-results](https://github.com/kedudzic/pl-normalizer-evaluation-just-results) (public)