Commit
Β·
d0b8175
1
Parent(s):
02a5ceb
Add irony data files
Browse files- README.md +13 -5
- dataset_infos.json +11 -44
- irony/test-00000-of-00001.parquet +3 -0
- irony/train-00000-of-00001.parquet +3 -0
- irony/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
|
@@ -135,16 +135,16 @@ dataset_info:
|
|
| 135 |
'1': irony
|
| 136 |
splits:
|
| 137 |
- name: train
|
| 138 |
-
num_bytes:
|
| 139 |
num_examples: 2862
|
| 140 |
- name: test
|
| 141 |
-
num_bytes:
|
| 142 |
num_examples: 784
|
| 143 |
- name: validation
|
| 144 |
-
num_bytes:
|
| 145 |
num_examples: 955
|
| 146 |
-
download_size:
|
| 147 |
-
dataset_size:
|
| 148 |
- config_name: offensive
|
| 149 |
features:
|
| 150 |
- name: text
|
|
@@ -330,6 +330,14 @@ configs:
|
|
| 330 |
path: hate/test-*
|
| 331 |
- split: validation
|
| 332 |
path: hate/validation-*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 333 |
train-eval-index:
|
| 334 |
- config: emotion
|
| 335 |
task: text-classification
|
|
|
|
| 135 |
'1': irony
|
| 136 |
splits:
|
| 137 |
- name: train
|
| 138 |
+
num_bytes: 259187
|
| 139 |
num_examples: 2862
|
| 140 |
- name: test
|
| 141 |
+
num_bytes: 75897
|
| 142 |
num_examples: 784
|
| 143 |
- name: validation
|
| 144 |
+
num_bytes: 86017
|
| 145 |
num_examples: 955
|
| 146 |
+
download_size: 297647
|
| 147 |
+
dataset_size: 421101
|
| 148 |
- config_name: offensive
|
| 149 |
features:
|
| 150 |
- name: text
|
|
|
|
| 330 |
path: hate/test-*
|
| 331 |
- split: validation
|
| 332 |
path: hate/validation-*
|
| 333 |
+
- config_name: irony
|
| 334 |
+
data_files:
|
| 335 |
+
- split: train
|
| 336 |
+
path: irony/train-*
|
| 337 |
+
- split: test
|
| 338 |
+
path: irony/test-*
|
| 339 |
+
- split: validation
|
| 340 |
+
path: irony/validation-*
|
| 341 |
train-eval-index:
|
| 342 |
- config: emotion
|
| 343 |
task: text-classification
|
dataset_infos.json
CHANGED
|
@@ -180,27 +180,21 @@
|
|
| 180 |
"features": {
|
| 181 |
"text": {
|
| 182 |
"dtype": "string",
|
| 183 |
-
"id": null,
|
| 184 |
"_type": "Value"
|
| 185 |
},
|
| 186 |
"label": {
|
| 187 |
-
"num_classes": 2,
|
| 188 |
"names": [
|
| 189 |
"non_irony",
|
| 190 |
"irony"
|
| 191 |
],
|
| 192 |
-
"names_file": null,
|
| 193 |
-
"id": null,
|
| 194 |
"_type": "ClassLabel"
|
| 195 |
}
|
| 196 |
},
|
| 197 |
-
"
|
| 198 |
-
"
|
| 199 |
-
"builder_name": " tweet_eval",
|
| 200 |
"config_name": "irony",
|
| 201 |
"version": {
|
| 202 |
"version_str": "1.1.0",
|
| 203 |
-
"description": null,
|
| 204 |
"major": 1,
|
| 205 |
"minor": 1,
|
| 206 |
"patch": 0
|
|
@@ -208,53 +202,26 @@
|
|
| 208 |
"splits": {
|
| 209 |
"train": {
|
| 210 |
"name": "train",
|
| 211 |
-
"num_bytes":
|
| 212 |
"num_examples": 2862,
|
| 213 |
-
"dataset_name":
|
| 214 |
},
|
| 215 |
"test": {
|
| 216 |
"name": "test",
|
| 217 |
-
"num_bytes":
|
| 218 |
"num_examples": 784,
|
| 219 |
-
"dataset_name":
|
| 220 |
},
|
| 221 |
"validation": {
|
| 222 |
"name": "validation",
|
| 223 |
-
"num_bytes":
|
| 224 |
"num_examples": 955,
|
| 225 |
-
"dataset_name":
|
| 226 |
-
}
|
| 227 |
-
},
|
| 228 |
-
"download_checksums": {
|
| 229 |
-
"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/train_text.txt": {
|
| 230 |
-
"num_bytes": 231594,
|
| 231 |
-
"checksum": "a888125a44f7dfaa25b026318748d0e62cc9a300d20f66eafd62011a19eaea23"
|
| 232 |
-
},
|
| 233 |
-
"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/train_labels.txt": {
|
| 234 |
-
"num_bytes": 5724,
|
| 235 |
-
"checksum": "fc69e6106c0f1f433a91536e08f83c71a391d7b219f7684d42f243a8089af77d"
|
| 236 |
-
},
|
| 237 |
-
"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/test_text.txt": {
|
| 238 |
-
"num_bytes": 68057,
|
| 239 |
-
"checksum": "53103da934a7308eee82f05f2a9781a8ea3e88604fdc1e02d3101108505c64be"
|
| 240 |
-
},
|
| 241 |
-
"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/test_labels.txt": {
|
| 242 |
-
"num_bytes": 1568,
|
| 243 |
-
"checksum": "08e2095e1725e74907a380614c220204e356bb46e3e8c93deb74e83e5b15ab38"
|
| 244 |
-
},
|
| 245 |
-
"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/val_text.txt": {
|
| 246 |
-
"num_bytes": 76760,
|
| 247 |
-
"checksum": "8806cf3793e300a485cfae34892fc3a0a2f9a183deb06c750c6531515c83051e"
|
| 248 |
-
},
|
| 249 |
-
"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/val_labels.txt": {
|
| 250 |
-
"num_bytes": 1910,
|
| 251 |
-
"checksum": "ccf429f63b4e8d0e7f425ca09445f7c31f7cea8a1b7c283b015b117c4002fd07"
|
| 252 |
}
|
| 253 |
},
|
| 254 |
-
"download_size":
|
| 255 |
-
"
|
| 256 |
-
"
|
| 257 |
-
"size_in_bytes": 806726
|
| 258 |
},
|
| 259 |
"offensive": {
|
| 260 |
"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n",
|
|
|
|
| 180 |
"features": {
|
| 181 |
"text": {
|
| 182 |
"dtype": "string",
|
|
|
|
| 183 |
"_type": "Value"
|
| 184 |
},
|
| 185 |
"label": {
|
|
|
|
| 186 |
"names": [
|
| 187 |
"non_irony",
|
| 188 |
"irony"
|
| 189 |
],
|
|
|
|
|
|
|
| 190 |
"_type": "ClassLabel"
|
| 191 |
}
|
| 192 |
},
|
| 193 |
+
"builder_name": "tweet_eval",
|
| 194 |
+
"dataset_name": "tweet_eval",
|
|
|
|
| 195 |
"config_name": "irony",
|
| 196 |
"version": {
|
| 197 |
"version_str": "1.1.0",
|
|
|
|
| 198 |
"major": 1,
|
| 199 |
"minor": 1,
|
| 200 |
"patch": 0
|
|
|
|
| 202 |
"splits": {
|
| 203 |
"train": {
|
| 204 |
"name": "train",
|
| 205 |
+
"num_bytes": 259187,
|
| 206 |
"num_examples": 2862,
|
| 207 |
+
"dataset_name": null
|
| 208 |
},
|
| 209 |
"test": {
|
| 210 |
"name": "test",
|
| 211 |
+
"num_bytes": 75897,
|
| 212 |
"num_examples": 784,
|
| 213 |
+
"dataset_name": null
|
| 214 |
},
|
| 215 |
"validation": {
|
| 216 |
"name": "validation",
|
| 217 |
+
"num_bytes": 86017,
|
| 218 |
"num_examples": 955,
|
| 219 |
+
"dataset_name": null
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 220 |
}
|
| 221 |
},
|
| 222 |
+
"download_size": 297647,
|
| 223 |
+
"dataset_size": 421101,
|
| 224 |
+
"size_in_bytes": 718748
|
|
|
|
| 225 |
},
|
| 226 |
"offensive": {
|
| 227 |
"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n",
|
irony/test-00000-of-00001.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6bcf4eae70637cacc9239773684a37cbfa3ef98fde5f173355651afee477862
|
| 3 |
+
size 53984
|
irony/train-00000-of-00001.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:541de8c14f53024660fa5ae1534489b51106c5f4905da294e47ded7e09d73ad0
|
| 3 |
+
size 182571
|
irony/validation-00000-of-00001.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2ac02a5943b02808f1998dfacba7fce89033be6b3e4bbb465a26faafabf21191
|
| 3 |
+
size 61092
|