gsaltintas commited on
Commit
55c95d8
·
verified ·
1 Parent(s): 9afdb5e

Uploading tokenizer_robustness_completion_italian_typographical_errors subset

Browse files
README.md CHANGED
@@ -2024,6 +2024,132 @@ dataset_info:
2024
  num_examples: 37
2025
  download_size: 35213
2026
  dataset_size: 20098
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2027
  configs:
2028
  - config_name: tokenizer_robustness_completion_italian_abbreviations
2029
  data_files:
@@ -2089,6 +2215,10 @@ configs:
2089
  data_files:
2090
  - split: test
2091
  path: tokenizer_robustness_completion_italian_spelled_out/test-*
 
 
 
 
2092
  ---
2093
 
2094
  # Dataset Card for Tokenization Robustness
 
2024
  num_examples: 37
2025
  download_size: 35213
2026
  dataset_size: 20098
2027
+ - config_name: tokenizer_robustness_completion_italian_typographical_errors
2028
+ features:
2029
+ - name: question
2030
+ dtype: string
2031
+ - name: choices
2032
+ list: string
2033
+ - name: answer
2034
+ dtype: int64
2035
+ - name: answer_label
2036
+ dtype: string
2037
+ - name: split
2038
+ dtype: string
2039
+ - name: subcategories
2040
+ dtype: string
2041
+ - name: category
2042
+ dtype: string
2043
+ - name: lang
2044
+ dtype: string
2045
+ - name: second_lang
2046
+ dtype: string
2047
+ - name: notes
2048
+ dtype: string
2049
+ - name: id
2050
+ dtype: string
2051
+ - name: set_id
2052
+ dtype: string
2053
+ - name: variation_id
2054
+ dtype: string
2055
+ - name: perturbed_word
2056
+ dtype: string
2057
+ - name: vanilla_cos_sim_to_canonical
2058
+ struct:
2059
+ - name: CohereLabs/aya-expanse-8b
2060
+ dtype: float64
2061
+ - name: Qwen/Qwen3-8B
2062
+ dtype: float64
2063
+ - name: bigscience/bloom
2064
+ dtype: float64
2065
+ - name: common-pile/comma-v0.1-1t
2066
+ dtype: float64
2067
+ - name: facebook/xglm-564M
2068
+ dtype: float64
2069
+ - name: google-bert/bert-base-multilingual-cased
2070
+ dtype: float64
2071
+ - name: google/byt5-small
2072
+ dtype: float64
2073
+ - name: google/gemma-2-2b
2074
+ dtype: float64
2075
+ - name: gpt2
2076
+ dtype: float64
2077
+ - name: meta-llama/Llama-3.2-1B
2078
+ dtype: float64
2079
+ - name: microsoft/Phi-3-mini-4k-instruct
2080
+ dtype: float64
2081
+ - name: mistralai/tekken
2082
+ dtype: float64
2083
+ - name: tiktoken/gpt-4o
2084
+ dtype: float64
2085
+ - name: tokenmonster/englishcode-32000-consistent-v1
2086
+ dtype: float64
2087
+ - name: trimmed_cos_sim_to_canonical
2088
+ struct:
2089
+ - name: CohereLabs/aya-expanse-8b
2090
+ dtype: float64
2091
+ - name: Qwen/Qwen3-8B
2092
+ dtype: float64
2093
+ - name: bigscience/bloom
2094
+ dtype: float64
2095
+ - name: common-pile/comma-v0.1-1t
2096
+ dtype: float64
2097
+ - name: facebook/xglm-564M
2098
+ dtype: float64
2099
+ - name: google-bert/bert-base-multilingual-cased
2100
+ dtype: float64
2101
+ - name: google/byt5-small
2102
+ dtype: float64
2103
+ - name: google/gemma-2-2b
2104
+ dtype: float64
2105
+ - name: gpt2
2106
+ dtype: float64
2107
+ - name: meta-llama/Llama-3.2-1B
2108
+ dtype: float64
2109
+ - name: microsoft/Phi-3-mini-4k-instruct
2110
+ dtype: float64
2111
+ - name: mistralai/tekken
2112
+ dtype: float64
2113
+ - name: tiktoken/gpt-4o
2114
+ dtype: float64
2115
+ - name: tokenmonster/englishcode-32000-consistent-v1
2116
+ dtype: float64
2117
+ - name: token_counts
2118
+ struct:
2119
+ - name: CohereLabs/aya-expanse-8b
2120
+ dtype: int64
2121
+ - name: Qwen/Qwen3-8B
2122
+ dtype: int64
2123
+ - name: bigscience/bloom
2124
+ dtype: int64
2125
+ - name: common-pile/comma-v0.1-1t
2126
+ dtype: int64
2127
+ - name: facebook/xglm-564M
2128
+ dtype: int64
2129
+ - name: google-bert/bert-base-multilingual-cased
2130
+ dtype: int64
2131
+ - name: google/byt5-small
2132
+ dtype: int64
2133
+ - name: google/gemma-2-2b
2134
+ dtype: int64
2135
+ - name: gpt2
2136
+ dtype: int64
2137
+ - name: meta-llama/Llama-3.2-1B
2138
+ dtype: int64
2139
+ - name: microsoft/Phi-3-mini-4k-instruct
2140
+ dtype: int64
2141
+ - name: mistralai/tekken
2142
+ dtype: int64
2143
+ - name: tiktoken/gpt-4o
2144
+ dtype: int64
2145
+ - name: tokenmonster/englishcode-32000-consistent-v1
2146
+ dtype: int64
2147
+ splits:
2148
+ - name: test
2149
+ num_bytes: 158477
2150
+ num_examples: 281
2151
+ download_size: 89075
2152
+ dataset_size: 158477
2153
  configs:
2154
  - config_name: tokenizer_robustness_completion_italian_abbreviations
2155
  data_files:
 
2215
  data_files:
2216
  - split: test
2217
  path: tokenizer_robustness_completion_italian_spelled_out/test-*
2218
+ - config_name: tokenizer_robustness_completion_italian_typographical_errors
2219
+ data_files:
2220
+ - split: test
2221
+ path: tokenizer_robustness_completion_italian_typographical_errors/test-*
2222
  ---
2223
 
2224
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_italian_typographical_errors/test-00000-of-00001.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:088eba2d17a83cfcdc93cbec43d0a3001276711c7e41879b6adb5e5684f82923
3
- size 88299
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c5f2ade7f77127c9371ff981184f93fe20f584f1fcb44678a94915717956eab
3
+ size 89075