| { |
| "splits": [ |
| "train", |
| "validation", |
| "test" |
| ], |
| "groups": [ |
| "base", |
| "bert_surprisal", |
| "character_arcs", |
| "emotional", |
| "gc_academic", |
| "gc_basic", |
| "gc_char_diversity", |
| "gc_concreteness", |
| "gc_dialogue", |
| "gc_discourse", |
| "gc_narrative", |
| "gc_polarity", |
| "gc_pos", |
| "gc_pronouns", |
| "gc_punctuation", |
| "gc_readability", |
| "gc_syntax", |
| "gc_temporal", |
| "ngram", |
| "ngram_surprisal", |
| "plot_shifts", |
| "rst", |
| "structure", |
| "surprisal" |
| ], |
| "stats": { |
| "train": { |
| "base": { |
| "n_samples": 117503, |
| "n_features": 21, |
| "size_mb": 7.204872131347656 |
| }, |
| "bert_surprisal": { |
| "n_samples": 117503, |
| "n_features": 29, |
| "size_mb": 15.838018417358398 |
| }, |
| "character_arcs": { |
| "n_samples": 117503, |
| "n_features": 4, |
| "size_mb": 0.20000839233398438 |
| }, |
| "emotional": { |
| "n_samples": 117503, |
| "n_features": 5, |
| "size_mb": 4.4886064529418945 |
| }, |
| "gc_academic": { |
| "n_samples": 117503, |
| "n_features": 8, |
| "size_mb": 0.6636857986450195 |
| }, |
| "gc_basic": { |
| "n_samples": 117503, |
| "n_features": 3, |
| "size_mb": 1.3420581817626953 |
| }, |
| "gc_char_diversity": { |
| "n_samples": 117503, |
| "n_features": 8, |
| "size_mb": 3.832453727722168 |
| }, |
| "gc_concreteness": { |
| "n_samples": 117503, |
| "n_features": 22, |
| "size_mb": 10.268058776855469 |
| }, |
| "gc_dialogue": { |
| "n_samples": 117503, |
| "n_features": 6, |
| "size_mb": 0.571375846862793 |
| }, |
| "gc_discourse": { |
| "n_samples": 117503, |
| "n_features": 7, |
| "size_mb": 1.0773344039916992 |
| }, |
| "gc_narrative": { |
| "n_samples": 117503, |
| "n_features": 10, |
| "size_mb": 0.8629426956176758 |
| }, |
| "gc_polarity": { |
| "n_samples": 117503, |
| "n_features": 23, |
| "size_mb": 19.75270366668701 |
| }, |
| "gc_pos": { |
| "n_samples": 117503, |
| "n_features": 8, |
| "size_mb": 1.6754398345947266 |
| }, |
| "gc_pronouns": { |
| "n_samples": 117503, |
| "n_features": 12, |
| "size_mb": 1.9187965393066406 |
| }, |
| "gc_punctuation": { |
| "n_samples": 117503, |
| "n_features": 8, |
| "size_mb": 0.7266168594360352 |
| }, |
| "gc_readability": { |
| "n_samples": 117503, |
| "n_features": 4, |
| "size_mb": 1.960618019104004 |
| }, |
| "gc_syntax": { |
| "n_samples": 117503, |
| "n_features": 575, |
| "size_mb": 30.994938850402832 |
| }, |
| "gc_temporal": { |
| "n_samples": 117503, |
| "n_features": 5, |
| "size_mb": 0.5964765548706055 |
| }, |
| "ngram": { |
| "n_samples": 117503, |
| "n_features": 9, |
| "size_mb": 1.9910192489624023 |
| }, |
| "ngram_surprisal": { |
| "n_samples": 117503, |
| "n_features": 86, |
| "size_mb": 1.2012948989868164 |
| }, |
| "plot_shifts": { |
| "n_samples": 117503, |
| "n_features": 3, |
| "size_mb": 1.5663232803344727 |
| }, |
| "rst": { |
| "n_samples": 117503, |
| "n_features": 46, |
| "size_mb": 3.908018112182617 |
| }, |
| "structure": { |
| "n_samples": 117503, |
| "n_features": 5, |
| "size_mb": 1.2650909423828125 |
| }, |
| "surprisal": { |
| "n_samples": 117503, |
| "n_features": 6, |
| "size_mb": 5.623771667480469 |
| } |
| }, |
| "validation": { |
| "base": { |
| "n_samples": 8052, |
| "n_features": 21, |
| "size_mb": 0.670506477355957 |
| }, |
| "bert_surprisal": { |
| "n_samples": 8052, |
| "n_features": 29, |
| "size_mb": 1.1183900833129883 |
| }, |
| "character_arcs": { |
| "n_samples": 8052, |
| "n_features": 4, |
| "size_mb": 0.0215606689453125 |
| }, |
| "emotional": { |
| "n_samples": 8052, |
| "n_features": 5, |
| "size_mb": 0.3020200729370117 |
| }, |
| "gc_academic": { |
| "n_samples": 8052, |
| "n_features": 8, |
| "size_mb": 0.07343864440917969 |
| }, |
| "gc_basic": { |
| "n_samples": 8052, |
| "n_features": 3, |
| "size_mb": 0.12804508209228516 |
| }, |
| "gc_char_diversity": { |
| "n_samples": 8052, |
| "n_features": 8, |
| "size_mb": 0.3675346374511719 |
| }, |
| "gc_concreteness": { |
| "n_samples": 8052, |
| "n_features": 22, |
| "size_mb": 0.806300163269043 |
| }, |
| "gc_dialogue": { |
| "n_samples": 8052, |
| "n_features": 6, |
| "size_mb": 0.056652069091796875 |
| }, |
| "gc_discourse": { |
| "n_samples": 8052, |
| "n_features": 7, |
| "size_mb": 0.10312652587890625 |
| }, |
| "gc_narrative": { |
| "n_samples": 8052, |
| "n_features": 10, |
| "size_mb": 0.08792686462402344 |
| }, |
| "gc_polarity": { |
| "n_samples": 8052, |
| "n_features": 23, |
| "size_mb": 1.337418556213379 |
| }, |
| "gc_pos": { |
| "n_samples": 8052, |
| "n_features": 8, |
| "size_mb": 0.1473827362060547 |
| }, |
| "gc_pronouns": { |
| "n_samples": 8052, |
| "n_features": 12, |
| "size_mb": 0.1820669174194336 |
| }, |
| "gc_punctuation": { |
| "n_samples": 8052, |
| "n_features": 8, |
| "size_mb": 0.07291126251220703 |
| }, |
| "gc_readability": { |
| "n_samples": 8052, |
| "n_features": 4, |
| "size_mb": 0.18763160705566406 |
| }, |
| "gc_syntax": { |
| "n_samples": 8052, |
| "n_features": 488, |
| "size_mb": 3.2519569396972656 |
| }, |
| "gc_temporal": { |
| "n_samples": 8052, |
| "n_features": 5, |
| "size_mb": 0.05423545837402344 |
| }, |
| "ngram": { |
| "n_samples": 8052, |
| "n_features": 9, |
| "size_mb": 0.18585777282714844 |
| }, |
| "ngram_surprisal": { |
| "n_samples": 8052, |
| "n_features": 86, |
| "size_mb": 0.14410018920898438 |
| }, |
| "plot_shifts": { |
| "n_samples": 8052, |
| "n_features": 3, |
| "size_mb": 0.13669204711914062 |
| }, |
| "rst": { |
| "n_samples": 8052, |
| "n_features": 46, |
| "size_mb": 0.3528108596801758 |
| }, |
| "structure": { |
| "n_samples": 8052, |
| "n_features": 5, |
| "size_mb": 0.159698486328125 |
| }, |
| "surprisal": { |
| "n_samples": 8052, |
| "n_features": 6, |
| "size_mb": 0.3733844757080078 |
| } |
| }, |
| "test": { |
| "base": { |
| "n_samples": 8156, |
| "n_features": 21, |
| "size_mb": 0.6687898635864258 |
| }, |
| "bert_surprisal": { |
| "n_samples": 8156, |
| "n_features": 29, |
| "size_mb": 1.1588869094848633 |
| }, |
| "character_arcs": { |
| "n_samples": 8156, |
| "n_features": 4, |
| "size_mb": 0.021600723266601562 |
| }, |
| "emotional": { |
| "n_samples": 8156, |
| "n_features": 5, |
| "size_mb": 0.3055391311645508 |
| }, |
| "gc_academic": { |
| "n_samples": 8156, |
| "n_features": 8, |
| "size_mb": 0.0720987319946289 |
| }, |
| "gc_basic": { |
| "n_samples": 8156, |
| "n_features": 3, |
| "size_mb": 0.12851428985595703 |
| }, |
| "gc_char_diversity": { |
| "n_samples": 8156, |
| "n_features": 8, |
| "size_mb": 0.36421966552734375 |
| }, |
| "gc_concreteness": { |
| "n_samples": 8156, |
| "n_features": 22, |
| "size_mb": 0.7973165512084961 |
| }, |
| "gc_dialogue": { |
| "n_samples": 8156, |
| "n_features": 6, |
| "size_mb": 0.05322837829589844 |
| }, |
| "gc_discourse": { |
| "n_samples": 8156, |
| "n_features": 7, |
| "size_mb": 0.101531982421875 |
| }, |
| "gc_narrative": { |
| "n_samples": 8156, |
| "n_features": 10, |
| "size_mb": 0.08516597747802734 |
| }, |
| "gc_polarity": { |
| "n_samples": 8156, |
| "n_features": 23, |
| "size_mb": 1.3533782958984375 |
| }, |
| "gc_pos": { |
| "n_samples": 8156, |
| "n_features": 8, |
| "size_mb": 0.14973068237304688 |
| }, |
| "gc_pronouns": { |
| "n_samples": 8156, |
| "n_features": 12, |
| "size_mb": 0.17980289459228516 |
| }, |
| "gc_punctuation": { |
| "n_samples": 8156, |
| "n_features": 8, |
| "size_mb": 0.07043266296386719 |
| }, |
| "gc_readability": { |
| "n_samples": 8156, |
| "n_features": 4, |
| "size_mb": 0.18674278259277344 |
| }, |
| "gc_syntax": { |
| "n_samples": 8156, |
| "n_features": 480, |
| "size_mb": 3.1574296951293945 |
| }, |
| "gc_temporal": { |
| "n_samples": 8156, |
| "n_features": 5, |
| "size_mb": 0.05355262756347656 |
| }, |
| "ngram": { |
| "n_samples": 8156, |
| "n_features": 9, |
| "size_mb": 0.18451881408691406 |
| }, |
| "ngram_surprisal": { |
| "n_samples": 8156, |
| "n_features": 86, |
| "size_mb": 0.1438894271850586 |
| }, |
| "plot_shifts": { |
| "n_samples": 8156, |
| "n_features": 3, |
| "size_mb": 0.13460922241210938 |
| }, |
| "rst": { |
| "n_samples": 8156, |
| "n_features": 46, |
| "size_mb": 0.3458251953125 |
| }, |
| "structure": { |
| "n_samples": 8156, |
| "n_features": 5, |
| "size_mb": 0.1639995574951172 |
| }, |
| "surprisal": { |
| "n_samples": 8156, |
| "n_features": 6, |
| "size_mb": 0.378387451171875 |
| } |
| } |
| }, |
| "total_size_mb": 140.1113634109497 |
| } |