gsaltintas commited on
Commit
0d3c8bd
·
verified ·
1 Parent(s): 5a62272

Uploading tokenizer_robustness_completion_italian_grammatical_errors subset

Browse files
README.md CHANGED
@@ -1016,6 +1016,132 @@ dataset_info:
1016
  num_examples: 68
1017
  download_size: 41589
1018
  dataset_size: 36901
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1019
  configs:
1020
  - config_name: tokenizer_robustness_completion_italian_abbreviations
1021
  data_files:
@@ -1049,6 +1175,10 @@ configs:
1049
  data_files:
1050
  - split: test
1051
  path: tokenizer_robustness_completion_italian_english_keyboard/test-*
 
 
 
 
1052
  ---
1053
 
1054
  # Dataset Card for Tokenization Robustness
 
1016
  num_examples: 68
1017
  download_size: 41589
1018
  dataset_size: 36901
1019
+ - config_name: tokenizer_robustness_completion_italian_grammatical_errors
1020
+ features:
1021
+ - name: question
1022
+ dtype: string
1023
+ - name: choices
1024
+ list: string
1025
+ - name: answer
1026
+ dtype: int64
1027
+ - name: answer_label
1028
+ dtype: string
1029
+ - name: split
1030
+ dtype: string
1031
+ - name: subcategories
1032
+ dtype: string
1033
+ - name: category
1034
+ dtype: string
1035
+ - name: lang
1036
+ dtype: string
1037
+ - name: second_lang
1038
+ dtype: string
1039
+ - name: notes
1040
+ dtype: string
1041
+ - name: id
1042
+ dtype: string
1043
+ - name: set_id
1044
+ dtype: string
1045
+ - name: variation_id
1046
+ dtype: string
1047
+ - name: perturbed_word
1048
+ dtype: string
1049
+ - name: vanilla_cos_sim_to_canonical
1050
+ struct:
1051
+ - name: CohereLabs/aya-expanse-8b
1052
+ dtype: float64
1053
+ - name: Qwen/Qwen3-8B
1054
+ dtype: float64
1055
+ - name: bigscience/bloom
1056
+ dtype: float64
1057
+ - name: common-pile/comma-v0.1-1t
1058
+ dtype: float64
1059
+ - name: facebook/xglm-564M
1060
+ dtype: float64
1061
+ - name: google-bert/bert-base-multilingual-cased
1062
+ dtype: float64
1063
+ - name: google/byt5-small
1064
+ dtype: float64
1065
+ - name: google/gemma-2-2b
1066
+ dtype: float64
1067
+ - name: gpt2
1068
+ dtype: float64
1069
+ - name: meta-llama/Llama-3.2-1B
1070
+ dtype: float64
1071
+ - name: microsoft/Phi-3-mini-4k-instruct
1072
+ dtype: float64
1073
+ - name: mistralai/tekken
1074
+ dtype: float64
1075
+ - name: tiktoken/gpt-4o
1076
+ dtype: float64
1077
+ - name: tokenmonster/englishcode-32000-consistent-v1
1078
+ dtype: float64
1079
+ - name: trimmed_cos_sim_to_canonical
1080
+ struct:
1081
+ - name: CohereLabs/aya-expanse-8b
1082
+ dtype: float64
1083
+ - name: Qwen/Qwen3-8B
1084
+ dtype: float64
1085
+ - name: bigscience/bloom
1086
+ dtype: float64
1087
+ - name: common-pile/comma-v0.1-1t
1088
+ dtype: float64
1089
+ - name: facebook/xglm-564M
1090
+ dtype: float64
1091
+ - name: google-bert/bert-base-multilingual-cased
1092
+ dtype: float64
1093
+ - name: google/byt5-small
1094
+ dtype: float64
1095
+ - name: google/gemma-2-2b
1096
+ dtype: float64
1097
+ - name: gpt2
1098
+ dtype: float64
1099
+ - name: meta-llama/Llama-3.2-1B
1100
+ dtype: float64
1101
+ - name: microsoft/Phi-3-mini-4k-instruct
1102
+ dtype: float64
1103
+ - name: mistralai/tekken
1104
+ dtype: float64
1105
+ - name: tiktoken/gpt-4o
1106
+ dtype: float64
1107
+ - name: tokenmonster/englishcode-32000-consistent-v1
1108
+ dtype: float64
1109
+ - name: token_counts
1110
+ struct:
1111
+ - name: CohereLabs/aya-expanse-8b
1112
+ dtype: int64
1113
+ - name: Qwen/Qwen3-8B
1114
+ dtype: int64
1115
+ - name: bigscience/bloom
1116
+ dtype: int64
1117
+ - name: common-pile/comma-v0.1-1t
1118
+ dtype: int64
1119
+ - name: facebook/xglm-564M
1120
+ dtype: int64
1121
+ - name: google-bert/bert-base-multilingual-cased
1122
+ dtype: int64
1123
+ - name: google/byt5-small
1124
+ dtype: int64
1125
+ - name: google/gemma-2-2b
1126
+ dtype: int64
1127
+ - name: gpt2
1128
+ dtype: int64
1129
+ - name: meta-llama/Llama-3.2-1B
1130
+ dtype: int64
1131
+ - name: microsoft/Phi-3-mini-4k-instruct
1132
+ dtype: int64
1133
+ - name: mistralai/tekken
1134
+ dtype: int64
1135
+ - name: tiktoken/gpt-4o
1136
+ dtype: int64
1137
+ - name: tokenmonster/englishcode-32000-consistent-v1
1138
+ dtype: int64
1139
+ splits:
1140
+ - name: test
1141
+ num_bytes: 49702
1142
+ num_examples: 86
1143
+ download_size: 49167
1144
+ dataset_size: 49702
1145
  configs:
1146
  - config_name: tokenizer_robustness_completion_italian_abbreviations
1147
  data_files:
 
1175
  data_files:
1176
  - split: test
1177
  path: tokenizer_robustness_completion_italian_english_keyboard/test-*
1178
+ - config_name: tokenizer_robustness_completion_italian_grammatical_errors
1179
+ data_files:
1180
+ - split: test
1181
+ path: tokenizer_robustness_completion_italian_grammatical_errors/test-*
1182
  ---
1183
 
1184
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_italian_grammatical_errors/test-00000-of-00001.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8faf9a4c689c4c890c36eee86713f3082ffe56487397691100b53e0f19a83aa
3
  size 49167
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd4cf5e072ce40ae48b3deec6f0a47382b548abdde7d3ab99666d930abec3c5b
3
  size 49167