upload the next config
Browse files
README.md
CHANGED
@@ -341,6 +341,44 @@ dataset_info:
|
|
341 |
num_examples: 546
|
342 |
download_size: 571864348
|
343 |
dataset_size: 662086441.09
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
344 |
- config_name: matching_from_pixels
|
345 |
features:
|
346 |
- name: image
|
@@ -696,6 +734,14 @@ configs:
|
|
696 |
path: matching_2/validation-*
|
697 |
- split: test
|
698 |
path: matching_2/test-*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
699 |
- config_name: matching_from_pixels
|
700 |
data_files:
|
701 |
- split: train
|
|
|
341 |
num_examples: 546
|
342 |
download_size: 571864348
|
343 |
dataset_size: 662086441.09
|
344 |
+
- config_name: matching_3
|
345 |
+
features:
|
346 |
+
- name: image
|
347 |
+
dtype: image
|
348 |
+
- name: contest_number
|
349 |
+
dtype: int32
|
350 |
+
- name: image_location
|
351 |
+
dtype: string
|
352 |
+
- name: image_description
|
353 |
+
dtype: string
|
354 |
+
- name: image_uncanny_description
|
355 |
+
dtype: string
|
356 |
+
- name: entities
|
357 |
+
sequence: string
|
358 |
+
- name: questions
|
359 |
+
sequence: string
|
360 |
+
- name: caption_choices
|
361 |
+
sequence: string
|
362 |
+
- name: from_description
|
363 |
+
dtype: string
|
364 |
+
- name: label
|
365 |
+
dtype: string
|
366 |
+
- name: n_tokens_label
|
367 |
+
dtype: int32
|
368 |
+
- name: instance_id
|
369 |
+
dtype: string
|
370 |
+
splits:
|
371 |
+
- name: train
|
372 |
+
num_bytes: 615620189.53
|
373 |
+
num_examples: 9630
|
374 |
+
- name: validation
|
375 |
+
num_bytes: 34829502.0
|
376 |
+
num_examples: 546
|
377 |
+
- name: test
|
378 |
+
num_bytes: 33697178.0
|
379 |
+
num_examples: 540
|
380 |
+
download_size: 571744845
|
381 |
+
dataset_size: 684146869.53
|
382 |
- config_name: matching_from_pixels
|
383 |
features:
|
384 |
- name: image
|
|
|
734 |
path: matching_2/validation-*
|
735 |
- split: test
|
736 |
path: matching_2/test-*
|
737 |
+
- config_name: matching_3
|
738 |
+
data_files:
|
739 |
+
- split: train
|
740 |
+
path: matching_3/train-*
|
741 |
+
- split: validation
|
742 |
+
path: matching_3/validation-*
|
743 |
+
- split: test
|
744 |
+
path: matching_3/test-*
|
745 |
- config_name: matching_from_pixels
|
746 |
data_files:
|
747 |
- split: train
|
dataset_infos.json
CHANGED
@@ -1470,89 +1470,67 @@
|
|
1470 |
"license": "",
|
1471 |
"features": {
|
1472 |
"image": {
|
1473 |
-
"decode": true,
|
1474 |
-
"id": null,
|
1475 |
"_type": "Image"
|
1476 |
},
|
1477 |
"contest_number": {
|
1478 |
"dtype": "int32",
|
1479 |
-
"id": null,
|
1480 |
"_type": "Value"
|
1481 |
},
|
1482 |
"image_location": {
|
1483 |
"dtype": "string",
|
1484 |
-
"id": null,
|
1485 |
"_type": "Value"
|
1486 |
},
|
1487 |
"image_description": {
|
1488 |
"dtype": "string",
|
1489 |
-
"id": null,
|
1490 |
"_type": "Value"
|
1491 |
},
|
1492 |
"image_uncanny_description": {
|
1493 |
"dtype": "string",
|
1494 |
-
"id": null,
|
1495 |
"_type": "Value"
|
1496 |
},
|
1497 |
"entities": {
|
1498 |
"feature": {
|
1499 |
"dtype": "string",
|
1500 |
-
"id": null,
|
1501 |
"_type": "Value"
|
1502 |
},
|
1503 |
-
"length": -1,
|
1504 |
-
"id": null,
|
1505 |
"_type": "Sequence"
|
1506 |
},
|
1507 |
"questions": {
|
1508 |
"feature": {
|
1509 |
"dtype": "string",
|
1510 |
-
"id": null,
|
1511 |
"_type": "Value"
|
1512 |
},
|
1513 |
-
"length": -1,
|
1514 |
-
"id": null,
|
1515 |
"_type": "Sequence"
|
1516 |
},
|
1517 |
"caption_choices": {
|
1518 |
"feature": {
|
1519 |
"dtype": "string",
|
1520 |
-
"id": null,
|
1521 |
"_type": "Value"
|
1522 |
},
|
1523 |
-
"length": -1,
|
1524 |
-
"id": null,
|
1525 |
"_type": "Sequence"
|
1526 |
},
|
1527 |
"from_description": {
|
1528 |
"dtype": "string",
|
1529 |
-
"id": null,
|
1530 |
"_type": "Value"
|
1531 |
},
|
1532 |
"label": {
|
1533 |
"dtype": "string",
|
1534 |
-
"id": null,
|
1535 |
"_type": "Value"
|
1536 |
},
|
1537 |
"n_tokens_label": {
|
1538 |
"dtype": "int32",
|
1539 |
-
"id": null,
|
1540 |
"_type": "Value"
|
1541 |
},
|
1542 |
"instance_id": {
|
1543 |
"dtype": "string",
|
1544 |
-
"id": null,
|
1545 |
"_type": "Value"
|
1546 |
}
|
1547 |
},
|
1548 |
-
"post_processed": null,
|
1549 |
-
"supervised_keys": null,
|
1550 |
-
"task_templates": null,
|
1551 |
"builder_name": "newyorker_caption_contest",
|
|
|
1552 |
"config_name": "matching_3",
|
1553 |
"version": {
|
1554 |
"version_str": "1.0.0",
|
1555 |
-
"description": null,
|
1556 |
"major": 1,
|
1557 |
"minor": 0,
|
1558 |
"patch": 0
|
@@ -1560,37 +1538,26 @@
|
|
1560 |
"splits": {
|
1561 |
"train": {
|
1562 |
"name": "train",
|
1563 |
-
"num_bytes":
|
1564 |
"num_examples": 9630,
|
1565 |
-
"dataset_name":
|
1566 |
},
|
1567 |
"validation": {
|
1568 |
"name": "validation",
|
1569 |
-
"num_bytes":
|
1570 |
"num_examples": 546,
|
1571 |
-
"dataset_name":
|
1572 |
},
|
1573 |
"test": {
|
1574 |
"name": "test",
|
1575 |
-
"num_bytes":
|
1576 |
"num_examples": 540,
|
1577 |
-
"dataset_name":
|
1578 |
-
}
|
1579 |
-
},
|
1580 |
-
"download_checksums": {
|
1581 |
-
"https://storage.googleapis.com/ai2-jack-public/caption_contest_data_public/huggingface_hub/v1.0/matching_3.zip": {
|
1582 |
-
"num_bytes": 3307856,
|
1583 |
-
"checksum": "21f934f5b7db7864856878346c1a30112dbe479a594e42db2867c88e5cfe2a57"
|
1584 |
-
},
|
1585 |
-
"https://storage.googleapis.com/ai2-jack-public/caption_contest_data_public/all_contest_images.zip": {
|
1586 |
-
"num_bytes": 39523209,
|
1587 |
-
"checksum": "741527b4ef7198d16cee42ae74eacbe239bcc7377f8b86811c27d627fdc77748"
|
1588 |
}
|
1589 |
},
|
1590 |
-
"download_size":
|
1591 |
-
"
|
1592 |
-
"
|
1593 |
-
"size_in_bytes": 57513620
|
1594 |
},
|
1595 |
"matching_from_pixels_3": {
|
1596 |
"description": "There are 3 caption contest tasks, described in the paper. In the Matching multiple choice task, models must recognize a caption written about a cartoon (vs. options that were not). In the Quality Ranking task, models must evaluate the quality\nof that caption by scoring it more highly than a lower quality option from the same contest. In the Explanation Generation task, models must explain why the joke is funny.\nYou are given a cartoon and 5 captions. Only one of the captions was truly written about the cartoon. You must select it.\n",
|
|
|
1470 |
"license": "",
|
1471 |
"features": {
|
1472 |
"image": {
|
|
|
|
|
1473 |
"_type": "Image"
|
1474 |
},
|
1475 |
"contest_number": {
|
1476 |
"dtype": "int32",
|
|
|
1477 |
"_type": "Value"
|
1478 |
},
|
1479 |
"image_location": {
|
1480 |
"dtype": "string",
|
|
|
1481 |
"_type": "Value"
|
1482 |
},
|
1483 |
"image_description": {
|
1484 |
"dtype": "string",
|
|
|
1485 |
"_type": "Value"
|
1486 |
},
|
1487 |
"image_uncanny_description": {
|
1488 |
"dtype": "string",
|
|
|
1489 |
"_type": "Value"
|
1490 |
},
|
1491 |
"entities": {
|
1492 |
"feature": {
|
1493 |
"dtype": "string",
|
|
|
1494 |
"_type": "Value"
|
1495 |
},
|
|
|
|
|
1496 |
"_type": "Sequence"
|
1497 |
},
|
1498 |
"questions": {
|
1499 |
"feature": {
|
1500 |
"dtype": "string",
|
|
|
1501 |
"_type": "Value"
|
1502 |
},
|
|
|
|
|
1503 |
"_type": "Sequence"
|
1504 |
},
|
1505 |
"caption_choices": {
|
1506 |
"feature": {
|
1507 |
"dtype": "string",
|
|
|
1508 |
"_type": "Value"
|
1509 |
},
|
|
|
|
|
1510 |
"_type": "Sequence"
|
1511 |
},
|
1512 |
"from_description": {
|
1513 |
"dtype": "string",
|
|
|
1514 |
"_type": "Value"
|
1515 |
},
|
1516 |
"label": {
|
1517 |
"dtype": "string",
|
|
|
1518 |
"_type": "Value"
|
1519 |
},
|
1520 |
"n_tokens_label": {
|
1521 |
"dtype": "int32",
|
|
|
1522 |
"_type": "Value"
|
1523 |
},
|
1524 |
"instance_id": {
|
1525 |
"dtype": "string",
|
|
|
1526 |
"_type": "Value"
|
1527 |
}
|
1528 |
},
|
|
|
|
|
|
|
1529 |
"builder_name": "newyorker_caption_contest",
|
1530 |
+
"dataset_name": "newyorker_caption_contest",
|
1531 |
"config_name": "matching_3",
|
1532 |
"version": {
|
1533 |
"version_str": "1.0.0",
|
|
|
1534 |
"major": 1,
|
1535 |
"minor": 0,
|
1536 |
"patch": 0
|
|
|
1538 |
"splits": {
|
1539 |
"train": {
|
1540 |
"name": "train",
|
1541 |
+
"num_bytes": 615620189.53,
|
1542 |
"num_examples": 9630,
|
1543 |
+
"dataset_name": null
|
1544 |
},
|
1545 |
"validation": {
|
1546 |
"name": "validation",
|
1547 |
+
"num_bytes": 34829502.0,
|
1548 |
"num_examples": 546,
|
1549 |
+
"dataset_name": null
|
1550 |
},
|
1551 |
"test": {
|
1552 |
"name": "test",
|
1553 |
+
"num_bytes": 33697178.0,
|
1554 |
"num_examples": 540,
|
1555 |
+
"dataset_name": null
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1556 |
}
|
1557 |
},
|
1558 |
+
"download_size": 571744845,
|
1559 |
+
"dataset_size": 684146869.53,
|
1560 |
+
"size_in_bytes": 1255891714.53
|
|
|
1561 |
},
|
1562 |
"matching_from_pixels_3": {
|
1563 |
"description": "There are 3 caption contest tasks, described in the paper. In the Matching multiple choice task, models must recognize a caption written about a cartoon (vs. options that were not). In the Quality Ranking task, models must evaluate the quality\nof that caption by scoring it more highly than a lower quality option from the same contest. In the Explanation Generation task, models must explain why the joke is funny.\nYou are given a cartoon and 5 captions. Only one of the captions was truly written about the cartoon. You must select it.\n",
|
matching_3/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:419f15312bbbd483fd3b0c9e46be76aa1ba321d19f0d288d1073912b5949f21e
|
3 |
+
size 24456063
|
matching_3/train-00000-of-00002.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bbaa7d612af0cd46eb40263be8b4eff1ad89b35aee8af7cdc3c4cdd1cca983db
|
3 |
+
size 261520099
|
matching_3/train-00001-of-00002.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:03c9cb66aa2e3e2d0f0f32e7ed28b401898a94867349ccf6f7eb3af9da4cef3b
|
3 |
+
size 260208504
|
matching_3/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:67ef5906159541f0c50308f540cef66938c0bc1a783d1c4dc4c8db3a26322b50
|
3 |
+
size 25560179
|