gsaltintas commited on
Commit
22cb4f0
·
verified ·
1 Parent(s): b6a4ffc

Uploading tokenizer_robustness_completion_italian_dialects subset

Browse files
README.md CHANGED
@@ -764,6 +764,132 @@ dataset_info:
764
  num_examples: 6
765
  download_size: 29387
766
  dataset_size: 3452
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
767
  configs:
768
  - config_name: tokenizer_robustness_completion_italian_abbreviations
769
  data_files:
@@ -789,6 +915,10 @@ configs:
789
  data_files:
790
  - split: test
791
  path: tokenizer_robustness_completion_italian_date_formats/test-*
 
 
 
 
792
  ---
793
 
794
  # Dataset Card for Tokenization Robustness
 
764
  num_examples: 6
765
  download_size: 29387
766
  dataset_size: 3452
767
+ - config_name: tokenizer_robustness_completion_italian_dialects
768
+ features:
769
+ - name: question
770
+ dtype: string
771
+ - name: choices
772
+ list: string
773
+ - name: answer
774
+ dtype: int64
775
+ - name: answer_label
776
+ dtype: string
777
+ - name: split
778
+ dtype: string
779
+ - name: subcategories
780
+ dtype: string
781
+ - name: category
782
+ dtype: string
783
+ - name: lang
784
+ dtype: string
785
+ - name: second_lang
786
+ dtype: string
787
+ - name: notes
788
+ dtype: string
789
+ - name: id
790
+ dtype: string
791
+ - name: set_id
792
+ dtype: string
793
+ - name: variation_id
794
+ dtype: string
795
+ - name: perturbed_word
796
+ dtype: string
797
+ - name: vanilla_cos_sim_to_canonical
798
+ struct:
799
+ - name: CohereLabs/aya-expanse-8b
800
+ dtype: float64
801
+ - name: Qwen/Qwen3-8B
802
+ dtype: float64
803
+ - name: bigscience/bloom
804
+ dtype: float64
805
+ - name: common-pile/comma-v0.1-1t
806
+ dtype: float64
807
+ - name: facebook/xglm-564M
808
+ dtype: float64
809
+ - name: google-bert/bert-base-multilingual-cased
810
+ dtype: float64
811
+ - name: google/byt5-small
812
+ dtype: float64
813
+ - name: google/gemma-2-2b
814
+ dtype: float64
815
+ - name: gpt2
816
+ dtype: float64
817
+ - name: meta-llama/Llama-3.2-1B
818
+ dtype: float64
819
+ - name: microsoft/Phi-3-mini-4k-instruct
820
+ dtype: float64
821
+ - name: mistralai/tekken
822
+ dtype: float64
823
+ - name: tiktoken/gpt-4o
824
+ dtype: float64
825
+ - name: tokenmonster/englishcode-32000-consistent-v1
826
+ dtype: float64
827
+ - name: trimmed_cos_sim_to_canonical
828
+ struct:
829
+ - name: CohereLabs/aya-expanse-8b
830
+ dtype: float64
831
+ - name: Qwen/Qwen3-8B
832
+ dtype: float64
833
+ - name: bigscience/bloom
834
+ dtype: float64
835
+ - name: common-pile/comma-v0.1-1t
836
+ dtype: float64
837
+ - name: facebook/xglm-564M
838
+ dtype: float64
839
+ - name: google-bert/bert-base-multilingual-cased
840
+ dtype: float64
841
+ - name: google/byt5-small
842
+ dtype: float64
843
+ - name: google/gemma-2-2b
844
+ dtype: float64
845
+ - name: gpt2
846
+ dtype: float64
847
+ - name: meta-llama/Llama-3.2-1B
848
+ dtype: float64
849
+ - name: microsoft/Phi-3-mini-4k-instruct
850
+ dtype: float64
851
+ - name: mistralai/tekken
852
+ dtype: float64
853
+ - name: tiktoken/gpt-4o
854
+ dtype: float64
855
+ - name: tokenmonster/englishcode-32000-consistent-v1
856
+ dtype: float64
857
+ - name: token_counts
858
+ struct:
859
+ - name: CohereLabs/aya-expanse-8b
860
+ dtype: int64
861
+ - name: Qwen/Qwen3-8B
862
+ dtype: int64
863
+ - name: bigscience/bloom
864
+ dtype: int64
865
+ - name: common-pile/comma-v0.1-1t
866
+ dtype: int64
867
+ - name: facebook/xglm-564M
868
+ dtype: int64
869
+ - name: google-bert/bert-base-multilingual-cased
870
+ dtype: int64
871
+ - name: google/byt5-small
872
+ dtype: int64
873
+ - name: google/gemma-2-2b
874
+ dtype: int64
875
+ - name: gpt2
876
+ dtype: int64
877
+ - name: meta-llama/Llama-3.2-1B
878
+ dtype: int64
879
+ - name: microsoft/Phi-3-mini-4k-instruct
880
+ dtype: int64
881
+ - name: mistralai/tekken
882
+ dtype: int64
883
+ - name: tiktoken/gpt-4o
884
+ dtype: int64
885
+ - name: tokenmonster/englishcode-32000-consistent-v1
886
+ dtype: int64
887
+ splits:
888
+ - name: test
889
+ num_bytes: 16359
890
+ num_examples: 31
891
+ download_size: 35120
892
+ dataset_size: 16359
893
  configs:
894
  - config_name: tokenizer_robustness_completion_italian_abbreviations
895
  data_files:
 
915
  data_files:
916
  - split: test
917
  path: tokenizer_robustness_completion_italian_date_formats/test-*
918
+ - config_name: tokenizer_robustness_completion_italian_dialects
919
+ data_files:
920
+ - split: test
921
+ path: tokenizer_robustness_completion_italian_dialects/test-*
922
  ---
923
 
924
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_italian_dialects/test-00000-of-00001.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf3591693393237cda0cb45b69d28caed012d6660ba94f58a2f9c2fef815d346
3
- size 35296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:152c80bcb3eef6f590e4f5000f4c7403e1f4d0363d3e29be3bd1d99f53305fc7
3
+ size 35120