gsaltintas commited on
Commit
516dbb1
·
verified ·
1 Parent(s): 43772d5

Uploading tokenizer_robustness_completion_italian_equivalent_expressions subset

Browse files
README.md CHANGED
@@ -40,6 +40,10 @@ configs:
40
  data_files:
41
  - split: test
42
  path: tokenizer_robustness_completion_italian_english_keyboard/test-*
 
 
 
 
43
  dataset_info:
44
  - config_name: tokenizer_robustness_completion_italian_abbreviations
45
  features:
@@ -1049,6 +1053,132 @@ dataset_info:
1049
  num_examples: 68
1050
  download_size: 41603
1051
  dataset_size: 36901
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1052
  ---
1053
 
1054
  # Dataset Card for Tokenization Robustness
 
40
  data_files:
41
  - split: test
42
  path: tokenizer_robustness_completion_italian_english_keyboard/test-*
43
+ - config_name: tokenizer_robustness_completion_italian_equivalent_expressions
44
+ data_files:
45
+ - split: test
46
+ path: tokenizer_robustness_completion_italian_equivalent_expressions/test-*
47
  dataset_info:
48
  - config_name: tokenizer_robustness_completion_italian_abbreviations
49
  features:
 
1053
  num_examples: 68
1054
  download_size: 41603
1055
  dataset_size: 36901
1056
+ - config_name: tokenizer_robustness_completion_italian_equivalent_expressions
1057
+ features:
1058
+ - name: question
1059
+ dtype: string
1060
+ - name: choices
1061
+ list: string
1062
+ - name: answer
1063
+ dtype: int64
1064
+ - name: answer_label
1065
+ dtype: string
1066
+ - name: split
1067
+ dtype: string
1068
+ - name: subcategories
1069
+ dtype: string
1070
+ - name: category
1071
+ dtype: string
1072
+ - name: lang
1073
+ dtype: string
1074
+ - name: second_lang
1075
+ dtype: string
1076
+ - name: notes
1077
+ dtype: string
1078
+ - name: id
1079
+ dtype: string
1080
+ - name: set_id
1081
+ dtype: string
1082
+ - name: variation_id
1083
+ dtype: string
1084
+ - name: perturbed_word
1085
+ dtype: string
1086
+ - name: vanilla_cos_sim_to_canonical
1087
+ struct:
1088
+ - name: CohereLabs/aya-expanse-8b
1089
+ dtype: float64
1090
+ - name: Qwen/Qwen3-8B
1091
+ dtype: float64
1092
+ - name: bigscience/bloom
1093
+ dtype: float64
1094
+ - name: common-pile/comma-v0.1-1t
1095
+ dtype: float64
1096
+ - name: facebook/xglm-564M
1097
+ dtype: float64
1098
+ - name: google-bert/bert-base-multilingual-cased
1099
+ dtype: float64
1100
+ - name: google/byt5-small
1101
+ dtype: float64
1102
+ - name: google/gemma-2-2b
1103
+ dtype: float64
1104
+ - name: gpt2
1105
+ dtype: float64
1106
+ - name: meta-llama/Llama-3.2-1B
1107
+ dtype: float64
1108
+ - name: microsoft/Phi-3-mini-4k-instruct
1109
+ dtype: float64
1110
+ - name: mistralai/tekken
1111
+ dtype: float64
1112
+ - name: tiktoken/gpt-4o
1113
+ dtype: float64
1114
+ - name: tokenmonster/englishcode-32000-consistent-v1
1115
+ dtype: float64
1116
+ - name: trimmed_cos_sim_to_canonical
1117
+ struct:
1118
+ - name: CohereLabs/aya-expanse-8b
1119
+ dtype: float64
1120
+ - name: Qwen/Qwen3-8B
1121
+ dtype: float64
1122
+ - name: bigscience/bloom
1123
+ dtype: float64
1124
+ - name: common-pile/comma-v0.1-1t
1125
+ dtype: float64
1126
+ - name: facebook/xglm-564M
1127
+ dtype: float64
1128
+ - name: google-bert/bert-base-multilingual-cased
1129
+ dtype: float64
1130
+ - name: google/byt5-small
1131
+ dtype: float64
1132
+ - name: google/gemma-2-2b
1133
+ dtype: float64
1134
+ - name: gpt2
1135
+ dtype: float64
1136
+ - name: meta-llama/Llama-3.2-1B
1137
+ dtype: float64
1138
+ - name: microsoft/Phi-3-mini-4k-instruct
1139
+ dtype: float64
1140
+ - name: mistralai/tekken
1141
+ dtype: float64
1142
+ - name: tiktoken/gpt-4o
1143
+ dtype: float64
1144
+ - name: tokenmonster/englishcode-32000-consistent-v1
1145
+ dtype: float64
1146
+ - name: token_counts
1147
+ struct:
1148
+ - name: CohereLabs/aya-expanse-8b
1149
+ dtype: int64
1150
+ - name: Qwen/Qwen3-8B
1151
+ dtype: int64
1152
+ - name: bigscience/bloom
1153
+ dtype: int64
1154
+ - name: common-pile/comma-v0.1-1t
1155
+ dtype: int64
1156
+ - name: facebook/xglm-564M
1157
+ dtype: int64
1158
+ - name: google-bert/bert-base-multilingual-cased
1159
+ dtype: int64
1160
+ - name: google/byt5-small
1161
+ dtype: int64
1162
+ - name: google/gemma-2-2b
1163
+ dtype: int64
1164
+ - name: gpt2
1165
+ dtype: int64
1166
+ - name: meta-llama/Llama-3.2-1B
1167
+ dtype: int64
1168
+ - name: microsoft/Phi-3-mini-4k-instruct
1169
+ dtype: int64
1170
+ - name: mistralai/tekken
1171
+ dtype: int64
1172
+ - name: tiktoken/gpt-4o
1173
+ dtype: int64
1174
+ - name: tokenmonster/englishcode-32000-consistent-v1
1175
+ dtype: int64
1176
+ splits:
1177
+ - name: test
1178
+ num_bytes: 1692
1179
+ num_examples: 3
1180
+ download_size: 28979
1181
+ dataset_size: 1692
1182
  ---
1183
 
1184
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_italian_equivalent_expressions/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a663c7a84569d953fe88f3288b2328f053658a948ca04c7f053c71c0190e4997
3
+ size 28979