url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.31B
node_id
stringlengths
18
32
number
int64
1
4.72k
title
stringlengths
1
276
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
null
comments
sequence
created_at
int64
1,587B
1,658B
updated_at
int64
1,587B
1,658B
closed_at
int64
1,587B
1,658B
author_association
stringclasses
3 values
active_lock_reason
null
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
3 values
draft
bool
2 classes
pull_request
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/4721
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4721/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4721/comments
https://api.github.com/repos/huggingface/datasets/issues/4721/events
https://github.com/huggingface/datasets/issues/4721
1,310,253,552
I_kwDODunzps5OGOHw
4,721
PyArrow Dataset error when calling `load_dataset`
{ "login": "piraka9011", "id": 16828657, "node_id": "MDQ6VXNlcjE2ODI4NjU3", "avatar_url": "https://avatars.githubusercontent.com/u/16828657?v=4", "gravatar_id": "", "url": "https://api.github.com/users/piraka9011", "html_url": "https://github.com/piraka9011", "followers_url": "https://api.github.com/users/piraka9011/followers", "following_url": "https://api.github.com/users/piraka9011/following{/other_user}", "gists_url": "https://api.github.com/users/piraka9011/gists{/gist_id}", "starred_url": "https://api.github.com/users/piraka9011/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/piraka9011/subscriptions", "organizations_url": "https://api.github.com/users/piraka9011/orgs", "repos_url": "https://api.github.com/users/piraka9011/repos", "events_url": "https://api.github.com/users/piraka9011/events{/privacy}", "received_events_url": "https://api.github.com/users/piraka9011/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[]
1,658,279,763,000
1,658,279,763,000
null
NONE
null
## Describe the bug I am fine tuning a wav2vec2 model following the script here using my own dataset: https://github.com/huggingface/transformers/blob/main/examples/pytorch/speech-recognition/run_speech_recognition_ctc.py Loading my Audio dataset from the hub which was originally generated from disk results in the following PyArrow error: ```sh File "/home/ubuntu/w2v2/run_speech_recognition_ctc.py", line 227, in main raw_datasets = load_dataset( File "/home/ubuntu/.virtualenvs/meval/lib/python3.8/site-packages/datasets/load.py", line 1679, in load_dataset builder_instance.download_and_prepare( File "/home/ubuntu/.virtualenvs/meval/lib/python3.8/site-packages/datasets/builder.py", line 704, in download_and_prepare self._download_and_prepare( File "/home/ubuntu/.virtualenvs/meval/lib/python3.8/site-packages/datasets/builder.py", line 793, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/ubuntu/.virtualenvs/meval/lib/python3.8/site-packages/datasets/builder.py", line 1268, in _prepare_split for key, table in logging.tqdm( File "/home/ubuntu/.virtualenvs/meval/lib/python3.8/site-packages/tqdm/std.py", line 1195, in __iter__ for obj in iterable: File "/home/ubuntu/.virtualenvs/meval/lib/python3.8/site-packages/datasets/packaged_modules/parquet/parquet.py", line 68, in _generate_tables for batch_idx, record_batch in enumerate( File "pyarrow/_parquet.pyx", line 1309, in iter_batches File "pyarrow/error.pxi", line 121, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Nested data conversions not implemented for chunked array outputs ``` ## Steps to reproduce the bug I created a dataset from a JSON lines manifest of `audio_filepath`, `text`, and `duration`. When creating the dataset, I do something like this: ```python import json from datasets import Dataset, Audio # manifest_lines is a list of dicts w/ "audio_filepath", "duration", and "text for line in manifest_lines: line = line.strip() if line: line_dict = json.loads(line) manifest_dict["audio"].append(f"{root_path}/{line_dict['audio_filepath']}") manifest_dict["duration"].append(line_dict["duration"]) manifest_dict["transcription"].append(line_dict["text"]) # Create a HF dataset dataset = Dataset.from_dict(manifest_dict).cast_column( "audio", Audio(sampling_rate=16_000), ) # From the docs for saving to disk # https://huggingface.co/docs/datasets/v2.3.2/en/package_reference/main_classes#datasets.Dataset.save_to_disk def read_audio_file(example): with open(example["audio"]["path"], "rb") as f: return {"audio": {"bytes": f.read()}} dataset = dataset.map(read_audio_file, num_proc=70) dataset.save_to_disk(f"/audio-data/hf/{artifact_name}") dataset.push_to_hub(f"{org-name}/{artifact_name}", max_shard_size="5GB", private=True) ``` Then when I call `load_dataset()` in my training script, with the same dataset I generated above, and download from the huggingface hub I get the above stack trace. I am able to load the dataset fine if I use `load_from_disk()`. ## Expected results `load_dataset()` should behave just like `load_from_disk()` and not cause any errors. ## Actual results See above ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> I am using the `huggingface/transformers-pytorch-gpu:latest` image - `datasets` version: 2.3.0 - Platform: Docker/Ubuntu 20.04 - Python version: 3.8 - PyArrow version: 8.0.0
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4721/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4721/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4720
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4720/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4720/comments
https://api.github.com/repos/huggingface/datasets/issues/4720/events
https://github.com/huggingface/datasets/issues/4720
1,309,980,195
I_kwDODunzps5OFLYj
4,720
Dataset Viewer issue for shamikbose89/lancaster_newsbooks
{ "login": "shamikbose", "id": 50837285, "node_id": "MDQ6VXNlcjUwODM3Mjg1", "avatar_url": "https://avatars.githubusercontent.com/u/50837285?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shamikbose", "html_url": "https://github.com/shamikbose", "followers_url": "https://api.github.com/users/shamikbose/followers", "following_url": "https://api.github.com/users/shamikbose/following{/other_user}", "gists_url": "https://api.github.com/users/shamikbose/gists{/gist_id}", "starred_url": "https://api.github.com/users/shamikbose/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shamikbose/subscriptions", "organizations_url": "https://api.github.com/users/shamikbose/orgs", "repos_url": "https://api.github.com/users/shamikbose/repos", "events_url": "https://api.github.com/users/shamikbose/events{/privacy}", "received_events_url": "https://api.github.com/users/shamikbose/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "It seems like the list of splits could not be obtained:\r\n\r\n```python\r\n>>> from datasets import get_dataset_split_names\r\n>>> get_dataset_split_names(\"shamikbose89/lancaster_newsbooks\", \"default\")\r\nUsing custom data configuration default\r\nTraceback (most recent call last):\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 354, in get_dataset_config_info\r\n for split_generator in builder._split_generators(\r\n File \"/home/slesage/.cache/huggingface/modules/datasets_modules/datasets/shamikbose89--lancaster_newsbooks/2d1c63d269bf7b9342accce0a95960b1710ab4bc774248878bd80eb96c1afaf7/lancaster_newsbooks.py\", line 73, in _split_generators\r\n data_dir = dl_manager.download_and_extract(_URL)\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py\", line 916, in download_and_extract\r\n return self.extract(self.download(url_or_urls))\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py\", line 879, in extract\r\n urlpaths = map_nested(self._extract, path_or_paths, map_tuple=True)\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/py_utils.py\", line 348, in map_nested\r\n return function(data_struct)\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py\", line 884, in _extract\r\n protocol = _get_extraction_protocol(urlpath, use_auth_token=self.download_config.use_auth_token)\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py\", line 388, in _get_extraction_protocol\r\n return _get_extraction_protocol_with_magic_number(f)\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py\", line 354, in _get_extraction_protocol_with_magic_number\r\n f.seek(0)\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/fsspec/implementations/http.py\", line 684, in seek\r\n raise ValueError(\"Cannot seek streaming HTTP file\")\r\nValueError: Cannot seek streaming HTTP file\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 404, in get_dataset_split_names\r\n info = get_dataset_config_info(\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 359, in get_dataset_config_info\r\n raise SplitsNotFoundError(\"The split names could not be parsed from the dataset config.\") from err\r\ndatasets.inspect.SplitsNotFoundError: The split names could not be parsed from the dataset config.\r\n```\r\n\r\nping @huggingface/datasets ", "Oh, I removed the 'split' key from `kwargs`. I put it back in, but there's still the same error" ]
1,658,260,807,000
1,658,262,112,000
null
NONE
null
### Link https://huggingface.co/datasets/shamikbose89/lancaster_newsbooks ### Description Status code: 400 Exception: ValueError Message: Cannot seek streaming HTTP file I am able to use the dataset loading script locally and it also runs when I'm using the one from the hub, but the viewer still doesn't load ### Owner Yes
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4720/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4720/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4719
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4719/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4719/comments
https://api.github.com/repos/huggingface/datasets/issues/4719/events
https://github.com/huggingface/datasets/issues/4719
1,309,854,492
I_kwDODunzps5OEssc
4,719
Issue loading TheNoob3131/mosquito-data dataset
{ "login": "thenerd31", "id": 53668030, "node_id": "MDQ6VXNlcjUzNjY4MDMw", "avatar_url": "https://avatars.githubusercontent.com/u/53668030?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thenerd31", "html_url": "https://github.com/thenerd31", "followers_url": "https://api.github.com/users/thenerd31/followers", "following_url": "https://api.github.com/users/thenerd31/following{/other_user}", "gists_url": "https://api.github.com/users/thenerd31/gists{/gist_id}", "starred_url": "https://api.github.com/users/thenerd31/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thenerd31/subscriptions", "organizations_url": "https://api.github.com/users/thenerd31/orgs", "repos_url": "https://api.github.com/users/thenerd31/repos", "events_url": "https://api.github.com/users/thenerd31/events{/privacy}", "received_events_url": "https://api.github.com/users/thenerd31/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I am also getting a ValueError: 'Couldn't cast' at the bottom. Is this because of some delimiter issue? My dataset is on the Huggingface Hub. If you could look at it, that would be greatly appreciated.", "Hi @thenerd31, thanks for reporting.\r\n\r\nPlease note that your issue is not caused by the Hugging Face Datasets library, but it has to do with the specific implementation of your dataset on the Hub.\r\n\r\nTherefore, I'm transferring this discussion to your own dataset Community tab: https://huggingface.co/datasets/TheNoob3131/mosquito-data/discussions/1" ]
1,658,252,857,000
1,658,299,617,000
1,658,299,562,000
NONE
null
![image](https://user-images.githubusercontent.com/53668030/179815591-d75fa7d3-3122-485f-a852-b06a68909066.png) So my dataset is public in the Huggingface Hub, but when I try to load it using the load_dataset command, it shows that it is downloading the files, but throws a ValueError. When I went to my directory to see if the files were downloaded, the folder was blank. Here is the error below: ValueError Traceback (most recent call last) Input In [8], in <cell line: 3>() 1 from datasets import load_dataset ----> 3 dataset = load_dataset("TheNoob3131/mosquito-data", split="train") File ~\Anaconda3\lib\site-packages\datasets\load.py:1679, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) 1676 try_from_hf_gcs = path not in _PACKAGED_DATASETS_MODULES 1678 # Download and prepare data -> 1679 builder_instance.download_and_prepare( 1680 download_config=download_config, 1681 download_mode=download_mode, 1682 ignore_verifications=ignore_verifications, 1683 try_from_hf_gcs=try_from_hf_gcs, 1684 use_auth_token=use_auth_token, 1685 ) 1687 # Build dataset for splits 1688 keep_in_memory = ( 1689 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size) 1690 ) Is the dataset in the wrong format or is there some security permission that I should enable?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4719/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4719/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4718
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4718/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4718/comments
https://api.github.com/repos/huggingface/datasets/issues/4718/events
https://github.com/huggingface/datasets/pull/4718
1,309,520,453
PR_kwDODunzps47prWR
4,718
Make Extractor accept Path as input
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4718). All of your documentation changes will be reflected on that endpoint." ]
1,658,237,106,000
1,658,237,565,000
null
MEMBER
null
This PR: - Makes `Extractor` accept instance of `Path` as input - Removes unnecessary castings of `Path` to `str`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4718/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4718/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4718", "html_url": "https://github.com/huggingface/datasets/pull/4718", "diff_url": "https://github.com/huggingface/datasets/pull/4718.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4718.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4717
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4717/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4717/comments
https://api.github.com/repos/huggingface/datasets/issues/4717/events
https://github.com/huggingface/datasets/issues/4717
1,309,512,483
I_kwDODunzps5ODZMj
4,717
Dataset Viewer issue for LawalAfeez/englishreview-ds-mini
{ "login": "lawalAfeez820", "id": 69974956, "node_id": "MDQ6VXNlcjY5OTc0OTU2", "avatar_url": "https://avatars.githubusercontent.com/u/69974956?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lawalAfeez820", "html_url": "https://github.com/lawalAfeez820", "followers_url": "https://api.github.com/users/lawalAfeez820/followers", "following_url": "https://api.github.com/users/lawalAfeez820/following{/other_user}", "gists_url": "https://api.github.com/users/lawalAfeez820/gists{/gist_id}", "starred_url": "https://api.github.com/users/lawalAfeez820/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lawalAfeez820/subscriptions", "organizations_url": "https://api.github.com/users/lawalAfeez820/orgs", "repos_url": "https://api.github.com/users/lawalAfeez820/repos", "events_url": "https://api.github.com/users/lawalAfeez820/events{/privacy}", "received_events_url": "https://api.github.com/users/lawalAfeez820/received_events", "type": "User", "site_admin": false }
[ { "id": 3470211881, "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer", "name": "dataset-viewer", "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co" } ]
open
false
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "organizations_url": "https://api.github.com/users/severo/orgs", "repos_url": "https://api.github.com/users/severo/repos", "events_url": "https://api.github.com/users/severo/events{/privacy}", "received_events_url": "https://api.github.com/users/severo/received_events", "type": "User", "site_admin": false }
[ { "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "organizations_url": "https://api.github.com/users/severo/orgs", "repos_url": "https://api.github.com/users/severo/repos", "events_url": "https://api.github.com/users/severo/events{/privacy}", "received_events_url": "https://api.github.com/users/severo/received_events", "type": "User", "site_admin": false } ]
null
[ "It's currently working, as far as I understand\r\n\r\nhttps://huggingface.co/datasets/LawalAfeez/englishreview-ds-mini/viewer/LawalAfeez--englishreview-ds-mini/train\r\n\r\n<img width=\"1556\" alt=\"Capture d’écran 2022-07-19 à 09 24 01\" src=\"https://user-images.githubusercontent.com/1676121/179761130-2d7980b9-c0f6-4093-8b1d-f0a3872fef3f.png\">\r\n\r\n---\r\n\r\nWhat was your issue?" ]
1,658,236,779,000
1,658,237,090,000
null
NONE
null
### Link _No response_ ### Description Unable to view the split data ### Owner _No response_
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4717/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4717/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4716
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4716/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4716/comments
https://api.github.com/repos/huggingface/datasets/issues/4716/events
https://github.com/huggingface/datasets/pull/4716
1,309,455,838
PR_kwDODunzps47pdbh
4,716
Support "tags" yaml tag
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4716). All of your documentation changes will be reflected on that endpoint." ]
1,658,234,071,000
1,658,234,463,000
null
MEMBER
null
Added the "tags" YAML tag, so that users can specify data domain/topics keywords for dataset search
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4716/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4716/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4716", "html_url": "https://github.com/huggingface/datasets/pull/4716", "diff_url": "https://github.com/huggingface/datasets/pull/4716.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4716.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4715
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4715/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4715/comments
https://api.github.com/repos/huggingface/datasets/issues/4715/events
https://github.com/huggingface/datasets/pull/4715
1,309,405,980
PR_kwDODunzps47pSui
4,715
Fix POS tags
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "CI failures are about missing content in the dataset cards or bad tags, and this is unrelated to this PR. Merging :)" ]
1,658,231,574,000
1,658,235,274,000
1,658,234,476,000
MEMBER
null
We're now using `part-of-speech` and not `part-of-speech-tagging`, see discussion here: https://github.com/huggingface/datasets/commit/114c09aff2fa1519597b46fbcd5a8e0c0d3ae020#r78794777
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4715/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4715/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4715", "html_url": "https://github.com/huggingface/datasets/pull/4715", "diff_url": "https://github.com/huggingface/datasets/pull/4715.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4715.patch", "merged_at": 1658234475000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4714
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4714/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4714/comments
https://api.github.com/repos/huggingface/datasets/issues/4714/events
https://github.com/huggingface/datasets/pull/4714
1,309,265,682
PR_kwDODunzps47o0YG
4,714
Fix named split casting and sorting
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4714). All of your documentation changes will be reflected on that endpoint." ]
1,658,224,108,000
1,658,234,026,000
null
MEMBER
null
This PR: - makes `NamedSplit` sortable: so that `sorted()` can be called on them - removes unnecessary `sorted()` on `dict.keys()`: `dict_keys` view is already like a `set` - removes unnecessary casting of `NamedSplit` to `str`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4714/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4714/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4714", "html_url": "https://github.com/huggingface/datasets/pull/4714", "diff_url": "https://github.com/huggingface/datasets/pull/4714.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4714.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4713
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4713/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4713/comments
https://api.github.com/repos/huggingface/datasets/issues/4713/events
https://github.com/huggingface/datasets/pull/4713
1,309,184,756
PR_kwDODunzps47ojC1
4,713
Document installation of sox OS dependency for audio
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4713). All of your documentation changes will be reflected on that endpoint." ]
1,658,220,155,000
1,658,220,558,000
null
MEMBER
null
The `sox` OS package needs being installed manually using the distribution package manager. This PR adds this explanation to the docs.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4713/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4713/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4713", "html_url": "https://github.com/huggingface/datasets/pull/4713", "diff_url": "https://github.com/huggingface/datasets/pull/4713.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4713.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4712
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4712/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4712/comments
https://api.github.com/repos/huggingface/datasets/issues/4712/events
https://github.com/huggingface/datasets/pull/4712
1,309,177,302
PR_kwDODunzps47ohdr
4,712
Update README.md
{ "login": "sbroadhurst-hf", "id": 108879611, "node_id": "U_kgDOBn1e-w", "avatar_url": "https://avatars.githubusercontent.com/u/108879611?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sbroadhurst-hf", "html_url": "https://github.com/sbroadhurst-hf", "followers_url": "https://api.github.com/users/sbroadhurst-hf/followers", "following_url": "https://api.github.com/users/sbroadhurst-hf/following{/other_user}", "gists_url": "https://api.github.com/users/sbroadhurst-hf/gists{/gist_id}", "starred_url": "https://api.github.com/users/sbroadhurst-hf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sbroadhurst-hf/subscriptions", "organizations_url": "https://api.github.com/users/sbroadhurst-hf/orgs", "repos_url": "https://api.github.com/users/sbroadhurst-hf/repos", "events_url": "https://api.github.com/users/sbroadhurst-hf/events{/privacy}", "received_events_url": "https://api.github.com/users/sbroadhurst-hf/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[]
1,658,219,780,000
1,658,219,780,000
null
NONE
null
Highlight that the licence granted by Amazon only covers non-commercial research use.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4712/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4712/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4712", "html_url": "https://github.com/huggingface/datasets/pull/4712", "diff_url": "https://github.com/huggingface/datasets/pull/4712.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4712.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4711
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4711/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4711/comments
https://api.github.com/repos/huggingface/datasets/issues/4711/events
https://github.com/huggingface/datasets/issues/4711
1,309,138,570
I_kwDODunzps5OB96K
4,711
Document how to create a dataset loading script for audio/vision
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
open
false
null
[]
null
[]
1,658,217,820,000
1,658,217,820,000
null
MEMBER
null
Currently, in our docs for Audio/Vision/Text, we explain how to: - Load data - Process data However we only explain how to *Create a dataset loading script* for text data. I think it would be useful that we add the same for Audio/Vision as these have some specificities different from Text. See, for example: - #4697 CC: @stevhliu
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4711/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4711/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4710
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4710/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4710/comments
https://api.github.com/repos/huggingface/datasets/issues/4710/events
https://github.com/huggingface/datasets/pull/4710
1,308,958,525
PR_kwDODunzps47ny0L
4,710
Add object detection processing tutorial
{ "login": "nateraw", "id": 32437151, "node_id": "MDQ6VXNlcjMyNDM3MTUx", "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nateraw", "html_url": "https://github.com/nateraw", "followers_url": "https://api.github.com/users/nateraw/followers", "following_url": "https://api.github.com/users/nateraw/following{/other_user}", "gists_url": "https://api.github.com/users/nateraw/gists{/gist_id}", "starred_url": "https://api.github.com/users/nateraw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nateraw/subscriptions", "organizations_url": "https://api.github.com/users/nateraw/orgs", "repos_url": "https://api.github.com/users/nateraw/repos", "events_url": "https://api.github.com/users/nateraw/events{/privacy}", "received_events_url": "https://api.github.com/users/nateraw/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4710). All of your documentation changes will be reflected on that endpoint.", "Great idea! Now that we have more than one task, it makes sense to separate image classification and object detection so it'll be easier for users to follow.", "@lhoestq do we want to do that in this PR, or should we merge it and let @stevhliu reorganize separately? " ]
1,658,204,626,000
1,658,255,506,000
null
CONTRIBUTOR
null
The following adds a quick guide on how to process object detection datasets with `albumentations`.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4710/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4710/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4710", "html_url": "https://github.com/huggingface/datasets/pull/4710", "diff_url": "https://github.com/huggingface/datasets/pull/4710.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4710.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4709
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4709/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4709/comments
https://api.github.com/repos/huggingface/datasets/issues/4709/events
https://github.com/huggingface/datasets/issues/4709
1,308,633,093
I_kwDODunzps5OACgF
4,709
WMT21 & WMT22
{ "login": "Muennighoff", "id": 62820084, "node_id": "MDQ6VXNlcjYyODIwMDg0", "avatar_url": "https://avatars.githubusercontent.com/u/62820084?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Muennighoff", "html_url": "https://github.com/Muennighoff", "followers_url": "https://api.github.com/users/Muennighoff/followers", "following_url": "https://api.github.com/users/Muennighoff/following{/other_user}", "gists_url": "https://api.github.com/users/Muennighoff/gists{/gist_id}", "starred_url": "https://api.github.com/users/Muennighoff/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Muennighoff/subscriptions", "organizations_url": "https://api.github.com/users/Muennighoff/orgs", "repos_url": "https://api.github.com/users/Muennighoff/repos", "events_url": "https://api.github.com/users/Muennighoff/events{/privacy}", "received_events_url": "https://api.github.com/users/Muennighoff/received_events", "type": "User", "site_admin": false }
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "Hi ! That would be awesome to have them indeed, thanks for opening this issue\r\n\r\nI just added you to the WMT org on the HF Hub if you're interested in adding those datasets.\r\n\r\nFeel free to create a dataset repository for each dataset and upload the data files there :) preferably in ZIP archives instead of TAR archives (the current WMT scripts don't support streaming TAR archives, so it would break the dataset preview). We've also had issues with the `statmt.org` host (data unavailable, slow download speed), that's why I think it's better if we re-host the files on the Hub.\r\n\r\n`wmt21` (and wmt22) can be added in this GitHub repository I think, for consistency with the previous ones.\r\nTo add it, you can copy paste the code of the previous one (e.g. wmt19), and add the new data:\r\n- in wmt_utils.py, add the new data subsets. You need to provide the download URLs, as well as the target and source languages\r\n- in wmt21.py (renamed from wmt19.py), you can specify the subsets that WMT21 uses (i.e. the one you just added)\r\n- in wmt_utils.py, define the python function that must be used to parse the subsets you added. To do so, you must go in `_generate_examples` and chose the proper `sub_generator` based on the subset name. For example, the `paracrawl_v3` subset uses the `_parse_tmx` function:\r\n\r\nhttps://github.com/huggingface/datasets/blob/ede72d3f9796339701ec59899c7c31d2427046fb/datasets/wmt19/wmt_utils.py#L834-L835\r\n\r\nHopefully the data is in a format that is already supported and there's no need to write a new `_parse_*` function for the new subsets. Let me know if you have questions or if I can help :)" ]
1,658,178,333,000
1,658,225,264,000
null
NONE
null
## Adding a Dataset - **Name:** WMT21 & WMT22 - **Description:** We are going to have three tracks: two small tasks and a large task. The small tracks evaluate translation between fairly related languages and English (all pairs). The large track uses 101 languages. - **Paper:** / - **Data:** https://statmt.org/wmt21/large-scale-multilingual-translation-task.html https://statmt.org/wmt22/large-scale-multilingual-translation-task.html - **Motivation:** Many more languages than previous WMT versions - Could be very high impact Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/main/ADD_NEW_DATASET.md). I could also tackle this. I saw the existing logic for WMT models is a bit complex (datasets are stored on the wmt account & retrieved in separate wmt datasets afaict). How long do you think it would take me? @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4709/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4709/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4708
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4708/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4708/comments
https://api.github.com/repos/huggingface/datasets/issues/4708/events
https://github.com/huggingface/datasets/pull/4708
1,308,279,700
PR_kwDODunzps47lewm
4,708
Fix require torchaudio and refactor test requirements
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4708). All of your documentation changes will be reflected on that endpoint." ]
1,658,165,068,000
1,658,165,527,000
null
MEMBER
null
Currently there is a bug in `require_torchaudio` (indeed it is requiring `sox` instead): ```python def require_torchaudio(test_case): if find_spec("sox") is None: ... ``` The bug was introduced by: - #3685 - Commit: https://github.com/huggingface/datasets/pull/3685/commits/b5a3e7122d49c4dcc9333b1d8d18a833fc04b940 which moved ```python require_sndfile = pytest.mark.skipif( # In Windows and OS X, soundfile installs sndfile (sys.platform != "linux" and find_spec("soundfile") is None) # In Linux, soundfile throws RuntimeError if sndfile not installed with distribution package manager or (sys.platform == "linux" and find_library("sndfile") is None), reason="Test requires 'sndfile': `pip install soundfile`; " "Linux requires sndfile installed with distribution package manager, e.g.: `sudo apt-get install libsndfile1`", ) require_sox = pytest.mark.skipif( find_library("sox") is None, reason="Test requires 'sox'; only available in non-Windows, e.g.: `sudo apt-get install sox`", ) require_torchaudio = pytest.mark.skipif(find_spec("torchaudio") is None, reason="Test requires 'torchaudio'") ``` to ```python def require_sndfile(test_case): """ Decorator marking a test that requires soundfile. These tests are skipped when soundfile isn't installed. """ if (sys.platform != "linux" and find_spec("soundfile") is None) or ( sys.platform == "linux" and find_library("sndfile") is None ): test_case = unittest.skip( "test requires 'sndfile': `pip install soundfile`; " "Linux requires sndfile installed with distribution package manager, e.g.: `sudo apt-get install libsndfile1`", )(test_case) return test_case def require_sox(test_case): """ Decorator marking a test that requires sox. These tests are skipped when sox isn't installed. """ if find_library("sox") is None: return unittest.skip("test requires 'sox'; only available in non-Windows, e.g.: `sudo apt-get install sox`")( test_case ) return test_case def require_torchaudio(test_case): """ Decorator marking a test that requires torchaudio. These tests are skipped when torchaudio isn't installed. """ if find_spec("sox") is None: return unittest.skip("test requires 'torchaudio'")(test_case) return test_case ``` This PR; - fixes the bug in `require_torchaudio` - refactors the test requirements back to using `pytest` instead of `unittest` - the text in `pytest.skipif` `reason` can be used if needed in a test body: `require_torchaudio.kwargs["reason"]`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4708/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4708/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4708", "html_url": "https://github.com/huggingface/datasets/pull/4708", "diff_url": "https://github.com/huggingface/datasets/pull/4708.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4708.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4707
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4707/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4707/comments
https://api.github.com/repos/huggingface/datasets/issues/4707/events
https://github.com/huggingface/datasets/issues/4707
1,308,251,405
I_kwDODunzps5N-lUN
4,707
Dataset Viewer issue for TheNoob3131/mosquito-data
{ "login": "thenerd31", "id": 53668030, "node_id": "MDQ6VXNlcjUzNjY4MDMw", "avatar_url": "https://avatars.githubusercontent.com/u/53668030?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thenerd31", "html_url": "https://github.com/thenerd31", "followers_url": "https://api.github.com/users/thenerd31/followers", "following_url": "https://api.github.com/users/thenerd31/following{/other_user}", "gists_url": "https://api.github.com/users/thenerd31/gists{/gist_id}", "starred_url": "https://api.github.com/users/thenerd31/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thenerd31/subscriptions", "organizations_url": "https://api.github.com/users/thenerd31/orgs", "repos_url": "https://api.github.com/users/thenerd31/repos", "events_url": "https://api.github.com/users/thenerd31/events{/privacy}", "received_events_url": "https://api.github.com/users/thenerd31/received_events", "type": "User", "site_admin": false }
[ { "id": 3470211881, "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer", "name": "dataset-viewer", "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co" } ]
closed
false
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "organizations_url": "https://api.github.com/users/severo/orgs", "repos_url": "https://api.github.com/users/severo/repos", "events_url": "https://api.github.com/users/severo/events{/privacy}", "received_events_url": "https://api.github.com/users/severo/received_events", "type": "User", "site_admin": false }
[ { "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "organizations_url": "https://api.github.com/users/severo/orgs", "repos_url": "https://api.github.com/users/severo/repos", "events_url": "https://api.github.com/users/severo/events{/privacy}", "received_events_url": "https://api.github.com/users/severo/received_events", "type": "User", "site_admin": false } ]
null
[ "Thanks for reporting. I refreshed the dataset viewer and it now works as expected.\r\n\r\nhttps://huggingface.co/datasets/TheNoob3131/mosquito-data\r\n\r\n<img width=\"1135\" alt=\"Capture d’écran 2022-07-18 à 13 15 22\" src=\"https://user-images.githubusercontent.com/1676121/179566497-e47f1a27-fd84-4a8d-9d7f-2e0f2da803df.png\">\r\n\r\nWe will investigate why it occurred in the first place\r\n", "By chance, could you provide some details about the operations done on the dataset: was it private? gated?", "Yes, it was a private dataset, and when I made it public, the Dataset Preview did not work. \r\n\r\nHowever, now when I make the dataset private, it says that the Dataset Preview has been disabled. Why is this?", "Thanks for the details. For now, the dataset viewer is always disabled on private datasets (see https://huggingface.co/docs/hub/datasets-viewer for more details)", "Hi, it was working fine for a few hours, but then I can't see the dataset viewer again (public dataset). Why is this still happening?\r\nIt's the same error too:\r\n![image](https://user-images.githubusercontent.com/53668030/179602465-f220f971-d3aa-49ba-a31b-60510f4c2a89.png)\r\n", "OK? This is a bug, thanks for help spotting and reproducing it (it occurs when a dataset is switched to private, then to public). We will be working on it, meanwhile, I've restored the dataset viewer manually again." ]
1,658,164,039,000
1,658,173,486,000
1,658,164,550,000
NONE
null
### Link _No response_ ### Description Getting this error when trying to view dataset preview: Message: 401, message='Unauthorized', url=URL('https://huggingface.co/datasets/TheNoob3131/mosquito-data/resolve/8aceebd6c4a359d216d10ef020868bd9e8c986dd/0_Africa_train.csv') ### Owner _No response_
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4707/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4707/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4706
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4706/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4706/comments
https://api.github.com/repos/huggingface/datasets/issues/4706/events
https://github.com/huggingface/datasets/pull/4706
1,308,198,454
PR_kwDODunzps47lNBg
4,706
Fix empty examples in xtreme dataset for bucc18 config
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "I guess the report link is this instead: https://huggingface.co/datasets/xtreme/discussions/1" ]
1,658,161,366,000
1,658,212,874,000
1,658,212,157,000
MEMBER
null
As reported in https://huggingface.co/muibk, there are empty examples in xtreme/bucc18.de I applied your fix @mustaszewski I also used a dict to make the dataset generation much faster
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4706/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4706/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4706", "html_url": "https://github.com/huggingface/datasets/pull/4706", "diff_url": "https://github.com/huggingface/datasets/pull/4706.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4706.patch", "merged_at": 1658212157000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4705
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4705/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4705/comments
https://api.github.com/repos/huggingface/datasets/issues/4705/events
https://github.com/huggingface/datasets/pull/4705
1,308,161,794
PR_kwDODunzps47lFDo
4,705
Fix crd3
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4705). All of your documentation changes will be reflected on that endpoint." ]
1,658,159,624,000
1,658,162,434,000
null
MEMBER
null
As reported in https://huggingface.co/datasets/crd3/discussions/1#62cc377073b2512b81662794, each split of the dataset was containing the same data. This issues comes from a bug in the dataset script I fixed it and also uploaded the data to hf.co to make the dataset work in streaming mode
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4705/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4705/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4705", "html_url": "https://github.com/huggingface/datasets/pull/4705", "diff_url": "https://github.com/huggingface/datasets/pull/4705.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4705.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4704
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4704/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4704/comments
https://api.github.com/repos/huggingface/datasets/issues/4704/events
https://github.com/huggingface/datasets/pull/4704
1,308,147,876
PR_kwDODunzps47lCFt
4,704
Skip tests only for lz4/zstd params if not installed
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
1,658,158,900,000
1,658,235,751,000
1,658,234,958,000
MEMBER
null
Currently, if `zstandard` or `lz4` are not installed, `test_compression_filesystems` and `test_streaming_dl_manager_extract_all_supported_single_file_compression_types` are skipped for all compression format parameters. This PR fixes these tests, so that if `zstandard` or `lz4` are not installed, the tests are skipped only for the corresponding compression parameters (`zstd` or `lz4`), whereas the tests are not skipped for all the other compression parameters (`gzip`, `xz` and `bz2`). Related to: - #4688
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4704/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4704/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4704", "html_url": "https://github.com/huggingface/datasets/pull/4704", "diff_url": "https://github.com/huggingface/datasets/pull/4704.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4704.patch", "merged_at": 1658234958000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4703
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4703/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4703/comments
https://api.github.com/repos/huggingface/datasets/issues/4703/events
https://github.com/huggingface/datasets/pull/4703
1,307,844,097
PR_kwDODunzps47kABf
4,703
Make cast in `from_pandas` more robust
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4703). All of your documentation changes will be reflected on that endpoint." ]
1,658,145,349,000
1,658,145,736,000
null
CONTRIBUTOR
null
Make the cast in `from_pandas` more robust (as it was done for the packaged modules in https://github.com/huggingface/datasets/pull/4364) This should be useful in situations like [this one](https://discuss.huggingface.co/t/loading-custom-audio-dataset-and-fine-tuning-model/8836/4).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4703/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4703/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4703", "html_url": "https://github.com/huggingface/datasets/pull/4703", "diff_url": "https://github.com/huggingface/datasets/pull/4703.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4703.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4702
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4702/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4702/comments
https://api.github.com/repos/huggingface/datasets/issues/4702/events
https://github.com/huggingface/datasets/issues/4702
1,307,793,811
I_kwDODunzps5N81mT
4,702
Domain specific dataset discovery on the Hugging Face hub
{ "login": "davanstrien", "id": 8995957, "node_id": "MDQ6VXNlcjg5OTU5NTc=", "avatar_url": "https://avatars.githubusercontent.com/u/8995957?v=4", "gravatar_id": "", "url": "https://api.github.com/users/davanstrien", "html_url": "https://github.com/davanstrien", "followers_url": "https://api.github.com/users/davanstrien/followers", "following_url": "https://api.github.com/users/davanstrien/following{/other_user}", "gists_url": "https://api.github.com/users/davanstrien/gists{/gist_id}", "starred_url": "https://api.github.com/users/davanstrien/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davanstrien/subscriptions", "organizations_url": "https://api.github.com/users/davanstrien/orgs", "repos_url": "https://api.github.com/users/davanstrien/repos", "events_url": "https://api.github.com/users/davanstrien/events{/privacy}", "received_events_url": "https://api.github.com/users/davanstrien/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Hi! I added a link to this issue in our internal request for adding keywords/topics to the Hub, which is identical to the `topic tags` solution. The `collections` solution seems too complex (as you point out). Regarding the `domain tags` solution, we primarily focus on machine learning, so I'm not sure if it's a good idea to make our current taxonomy more complex.", "> Hi! I added a link to this issue in our internal request for adding keywords/topics to the Hub, which is identical to the `topic tags` solution. The `collections` solution seems too complex (as you point out). Regarding the `domain tags` solution, we primarily focus on machine learning, so I'm not sure if it's a good idea to make our current taxonomy more complex.\r\n\r\nThanks, for letting me know. Will you allow the topic tags to be user-generated or only chosen from a list?", "Thanks for opening this issue @davanstrien.\r\n\r\nAs we discussed last week, the tag approach would be in principle the simpler to be implemented, either the domain tag (with closed vocabulary: more reliable but also more rigid), or the topic tag (with open vocabulary: more flexible for user needs)", "Hi @davanstrien If i remember correctly this was also discussed inside a hf.co Discussion, would you be able to link it here too?\r\n\r\n(where i suggested using `tags: - foo - bar` IIRC.\r\n\r\nThanks a ton!", "> Hi @davanstrien If i remember correctly this was also discussed inside a hf.co Discussion, would you be able to link it here too?\r\n> \r\n> (where i suggested using `tags: - foo - bar` IIRC.\r\n> \r\n> Thanks a ton!\r\n\r\nThis doesn't ring a bell - I did a quick search of https://discuss.huggingface.co but didn't find anything. \r\n\r\nThe `tags: ` approach sounds like a good option for this. It would be especially nice if these could suggest existing tags, but this probably won't be easily possible through the current interface. \r\n", "I opened a PR to add \"tags\" to the YAML validator:\r\nhttps://github.com/huggingface/datasets/pull/4716\r\n\r\nI also added \"tags\" to the [tagging app](https://huggingface.co/spaces/huggingface/datasets-tagging), with suggestions like \"bio\" or \"newspapers\"", "Thanks @lhoestq for the initiative.\r\n \r\nJust one question: are \"tags\" already supported on the Hub? \r\n\r\nI think they aren't. Thus, the Hub should support them so that they are properly displayed.", "I think they're not displayed, but at least it should enable users to filter by tag in using `huggingface_hub` or using the appropriate query params on the website (not sure if it's possible yet though)", "> I think they're not displayed, but at least it should enable users to filter by tag in using `huggingface_hub` or using the appropriate query params on the website (not sure if it's possible yet though)\r\n\r\nI think this would already be a helpful start. I'm happy to try this out with the datasets added to https://huggingface.co/organizations/biglam and use the `huggingface_hub` to filter those datasets using the tags. " ]
1,658,142,843,000
1,658,243,891,000
null
CONTRIBUTOR
null
**Is your feature request related to a problem? Please describe.** ## The problem The datasets hub currently has `8,239` datasets. These datasets span a wide range of different modalities and tasks (currently with a bias towards textual data). There are various ways of identifying datasets that may be relevant for a particular use case: - searching - various filters Currently, however, there isn't an easy way to identify datasets belonging to a specific domain. For example, I want to browse machine learning datasets related to 'social science' or 'climate change research'. The ability to identify datasets relating to a specific domain has come up in discussions around the [BigLA](https://github.com/bigscience-workshop/lam/) datasets hackathon https://github.com/bigscience-workshop/lam/discussions/31#discussioncomment-3123610. As part of the hackathon, we're currently collecting datasets related to Libraries, Archives and Museums and making them available via the hub. We currently do this under a Hugging Face organization (https://huggingface.co/biglam). However, going forward, I can see some of these datasets being migrated to sit under an organization that is the custodian of the dataset (for example, a national library the data was originally from). At this point, it becomes more difficult to quickly identify datasets from this domain without relying on search. This is also related to some existing issues on Github related to metadata on the hub: - https://github.com/huggingface/datasets/issues/3625 - https://github.com/huggingface/datasets/issues/3877 **Describe the solution you'd like** ### Some possible solutions that may help with this: #### Enable domain tags (from a controlled vocabulary) - This would add metadata field to the YAML for the domain a dataset relates to - Advantages: - the list is controlled, allowing it to be more easily integrated into the datasets tag app (https://huggingface.co/space/huggingface/datasets-tagging) - the controlled vocabulary could align with an existing controlled vocabulary - this additional metadata can be used to perform filtering by domain - disadvantages - choosing the best controlled vocab may be difficult - there are many datasets that are likely to fit into the 'machine learning' domain (i.e. there is a long tail of datasets that aren't in more 'generic' machine learning domain #### Enable topic tags (user-generated) Enable 'free form' topic tags for datasets and models. This would be closer to GitHub's repository topics which can be chosen from a controlled list (https://github.com/topics/) but can also be more user/org specific. This could potentially be useful for organizations to also manage their own models and datasets as the number they hold in their org grows. For example, they may create 'topic tags' for a specific project, so it's clearer which datasets /models are related to that project. #### Collections This solution would likely be the biggest shift and may require significant changes in the hub fronted. Collections could work in several different ways but would include: Users can curate particular datasets, models, spaces, etc., into a collection. For example, they may create a collection of 'historic newspapers suitable for training language models'. These collections would not be mutually exclusive, i.e. a dataset can belong to zero, one or many collections. Collections can also potentially be nested under other collections. This is fairly common on other data reposotiores for example the following collections: <img width="293" alt="Screenshot 2022-07-18 at 11 50 44" src="https://user-images.githubusercontent.com/8995957/179496445-963ed122-5e26-4574-96e8-41081bce3e2b.png"> all belong under a higher level collection (https://bl.iro.bl.uk/collections/353c908d-b495-4413-b047-87236d2573e3?locale=en). There are different models one could use for how these collections could be created: - only within an org - for any dataset/model - the owner or a dataset/model has to agree to be added to a collection - a collection owner can have people suggest additions to their collection - other models.... These collections could be thematic, related to particular training approaches, curate models with particular inference properties etc. Whilst some of these features may duplicate current/or future tag filters on the hub, they offer the advantage of being flexible and not having to predict what users will want to do upfront. There is also potential for automating the creation of these collections based on existing metadata. For example, one could collect models trained on a collection of datasets so for example, if we had a collection of 'historic newspapers suitable for training language models' that contained 30 datasets, we could create another collection 'historic newspaper language models' that takes any model on the hub whose metadata says it used one or more of those 30 datasets. There is also the option of exploring ML approaches to suggest models/datasets may be relevant to a particular collection. This approach is likely to be quite difficult to implement well and would require significant thought. There is also likely to be a benefit in doing quite a bit of upfront work in curating useful collections to demonstrate the benefits of collections. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. It is possible to collate this information externally, i.e. one could link back to the relevant models/datasets from an external platform. **Additional context** Add any other context about the feature request here. I'm cc'ing others involved in the BigLAM hackathon who may also have thoughts @cakiki @clancyoftheoverflow @albertvillanova
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4702/reactions", "total_count": 2, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4702/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4701
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4701/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4701/comments
https://api.github.com/repos/huggingface/datasets/issues/4701/events
https://github.com/huggingface/datasets/pull/4701
1,307,689,625
PR_kwDODunzps47jeE9
4,701
Added more information in the README about contributors of the Arabic Speech Corpus
{ "login": "nawarhalabi", "id": 2845798, "node_id": "MDQ6VXNlcjI4NDU3OTg=", "avatar_url": "https://avatars.githubusercontent.com/u/2845798?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nawarhalabi", "html_url": "https://github.com/nawarhalabi", "followers_url": "https://api.github.com/users/nawarhalabi/followers", "following_url": "https://api.github.com/users/nawarhalabi/following{/other_user}", "gists_url": "https://api.github.com/users/nawarhalabi/gists{/gist_id}", "starred_url": "https://api.github.com/users/nawarhalabi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nawarhalabi/subscriptions", "organizations_url": "https://api.github.com/users/nawarhalabi/orgs", "repos_url": "https://api.github.com/users/nawarhalabi/repos", "events_url": "https://api.github.com/users/nawarhalabi/events{/privacy}", "received_events_url": "https://api.github.com/users/nawarhalabi/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[]
1,658,137,683,000
1,658,212,646,000
null
NONE
null
Added more information in the README about contributors and encouraged reading the thesis for more infos
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4701/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4701/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4701", "html_url": "https://github.com/huggingface/datasets/pull/4701", "diff_url": "https://github.com/huggingface/datasets/pull/4701.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4701.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4700
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4700/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4700/comments
https://api.github.com/repos/huggingface/datasets/issues/4700/events
https://github.com/huggingface/datasets/pull/4700
1,307,599,161
PR_kwDODunzps47jKNx
4,700
Support extract lz4 compressed data files
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
1,658,133,691,000
1,658,155,439,000
1,658,154,707,000
MEMBER
null
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4700/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4700/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4700", "html_url": "https://github.com/huggingface/datasets/pull/4700", "diff_url": "https://github.com/huggingface/datasets/pull/4700.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4700.patch", "merged_at": 1658154707000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4699
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4699/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4699/comments
https://api.github.com/repos/huggingface/datasets/issues/4699/events
https://github.com/huggingface/datasets/pull/4699
1,307,555,592
PR_kwDODunzps47jA6Z
4,699
Fix Authentification Error while streaming
{ "login": "hkjeon13", "id": 37480967, "node_id": "MDQ6VXNlcjM3NDgwOTY3", "avatar_url": "https://avatars.githubusercontent.com/u/37480967?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hkjeon13", "html_url": "https://github.com/hkjeon13", "followers_url": "https://api.github.com/users/hkjeon13/followers", "following_url": "https://api.github.com/users/hkjeon13/following{/other_user}", "gists_url": "https://api.github.com/users/hkjeon13/gists{/gist_id}", "starred_url": "https://api.github.com/users/hkjeon13/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hkjeon13/subscriptions", "organizations_url": "https://api.github.com/users/hkjeon13/orgs", "repos_url": "https://api.github.com/users/hkjeon13/repos", "events_url": "https://api.github.com/users/hkjeon13/events{/privacy}", "received_events_url": "https://api.github.com/users/hkjeon13/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[]
1,658,131,421,000
1,658,229,560,000
null
NONE
null
I fixed a few errors when it occurs while streaming the private dataset on the Huggingface Hub. ``` from datasets import load_dataset dataset = load_dataset(<repo_id>, use_auth_token=<private_token>, streaming=True) for d in dataset['train']: print(d) break # this is for checking ``` This code is an example for streaming private datasets. when the version of the datasets is 2.2.2, it works well but datasets>2.2.2 occurs error like this, ``` /usr/local/lib/python3.7/dist-packages/aiohttp/client_reqrep.py in raise_for_status(self) 1007 status=self.status, 1008 message=self.reason, → 1009 headers=self.headers, 1010 ) 1011 ClientResponseError: 401, message='Unauthorized', url=URL('https://huggingface.co/datasets/.../train-00000-of-00001-168b451062c67c34.parquet') ``` (this is an example on the dataset has `parquet` extenstion) It seems that the `xisfile `module in `download/streaming_download_manager.py` couldn't recognize the file on "https://huggingface.co/~". so I add three lines. With this change, there is no error anymore(but this code is ad-hoc).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4699/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4699/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4699", "html_url": "https://github.com/huggingface/datasets/pull/4699", "diff_url": "https://github.com/huggingface/datasets/pull/4699.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4699.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4698
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4698/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4698/comments
https://api.github.com/repos/huggingface/datasets/issues/4698/events
https://github.com/huggingface/datasets/pull/4698
1,307,539,585
PR_kwDODunzps47i9gN
4,698
Enable streaming dataset to use the "all" split
{ "login": "cakiki", "id": 3664563, "node_id": "MDQ6VXNlcjM2NjQ1NjM=", "avatar_url": "https://avatars.githubusercontent.com/u/3664563?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cakiki", "html_url": "https://github.com/cakiki", "followers_url": "https://api.github.com/users/cakiki/followers", "following_url": "https://api.github.com/users/cakiki/following{/other_user}", "gists_url": "https://api.github.com/users/cakiki/gists{/gist_id}", "starred_url": "https://api.github.com/users/cakiki/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cakiki/subscriptions", "organizations_url": "https://api.github.com/users/cakiki/orgs", "repos_url": "https://api.github.com/users/cakiki/repos", "events_url": "https://api.github.com/users/cakiki/events{/privacy}", "received_events_url": "https://api.github.com/users/cakiki/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4698). All of your documentation changes will be reflected on that endpoint.", "@albertvillanova \r\nAdding the validation split causes these two `assert_called_once` assertions to fail with `AssertionError: Expected 'ArrowWriter' to have been called once. Called 2 times`:\r\n\r\nhttps://github.com/huggingface/datasets/blob/main/tests/test_builder.py#L548-L562\r\n\r\nIt might be better to create a new dummy generator for the streaming tests, WDYT? Alternatively we could test for `self.call_count` equalling 2." ]
1,658,130,459,000
1,658,132,260,000
null
CONTRIBUTOR
null
Fixes #4637
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4698/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4698/timeline
null
null
true
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4698", "html_url": "https://github.com/huggingface/datasets/pull/4698", "diff_url": "https://github.com/huggingface/datasets/pull/4698.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4698.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4697
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4697/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4697/comments
https://api.github.com/repos/huggingface/datasets/issues/4697/events
https://github.com/huggingface/datasets/issues/4697
1,307,332,253
I_kwDODunzps5N7E6d
4,697
Trouble with streaming of vision dataset with TAR archive
{ "login": "frgfm", "id": 26927750, "node_id": "MDQ6VXNlcjI2OTI3NzUw", "avatar_url": "https://avatars.githubusercontent.com/u/26927750?v=4", "gravatar_id": "", "url": "https://api.github.com/users/frgfm", "html_url": "https://github.com/frgfm", "followers_url": "https://api.github.com/users/frgfm/followers", "following_url": "https://api.github.com/users/frgfm/following{/other_user}", "gists_url": "https://api.github.com/users/frgfm/gists{/gist_id}", "starred_url": "https://api.github.com/users/frgfm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/frgfm/subscriptions", "organizations_url": "https://api.github.com/users/frgfm/orgs", "repos_url": "https://api.github.com/users/frgfm/repos", "events_url": "https://api.github.com/users/frgfm/events{/privacy}", "received_events_url": "https://api.github.com/users/frgfm/received_events", "type": "User", "site_admin": false }
[ { "id": 3287858981, "node_id": "MDU6TGFiZWwzMjg3ODU4OTgx", "url": "https://api.github.com/repos/huggingface/datasets/labels/streaming", "name": "streaming", "color": "fef2c0", "default": false, "description": "" } ]
open
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
[ "Hi @frgfm, thanks for reporting.\r\n\r\nAs the error message says, streaming mode is not supported out of the box when the dataset contains TAR archive files.\r\n\r\nTo make the dataset streamable, you have to use `dl_manager.iter_archive`.\r\n\r\nThere are several examples in other datasets, e.g. food101: https://huggingface.co/datasets/food101/blob/main/food101.py\r\n\r\nAnd yes, as the link you pointed out, for the streaming to be possible, the metadata file must be loaded before all of the images:\r\n- either this is the case when iterating the archive (and you get the metadata file before the images)\r\n- or you have to extract the metadata file by hand and upload it separately to the Hub", "Hi @albertvillanova :wave:\r\n\r\nThanks! Yeah I saw that but since I didn't have any metadata, I wasn't sure whether I should create them myself.\r\n\r\nSo one last question:\r\nWhat is the metadata supposed to be for archives? The relative path of all files in it?\r\n_(Sorry I'm a bit confused since it's quite hard to debug using the single error message from the data preview :sweat_smile: )_" ]
1,658,112,669,000
1,658,219,232,000
null
NONE
null
### Link https://huggingface.co/datasets/frgfm/imagenette ### Description Hello there :wave: Thanks for the amazing work you've done with HF Datasets! I've just started playing with it, and managed to upload my first dataset. But for the second one, I'm having trouble with the preview since there is some archive extraction involved :sweat_smile: Basically, I get a: ``` Status code: 400 Exception: NotImplementedError Message: Extraction protocol for TAR archives like 'https://s3.amazonaws.com/fast-ai-imageclas/imagenette2.tgz' is not implemented in streaming mode. Please use `dl_manager.iter_archive` instead. ``` I've tried several things and checked this issue https://github.com/huggingface/datasets/issues/4181 as well, but no luck so far! Could you point me in the right direction please? :pray: ### Owner Yes
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4697/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4697/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4696
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4696/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4696/comments
https://api.github.com/repos/huggingface/datasets/issues/4696/events
https://github.com/huggingface/datasets/issues/4696
1,307,183,099
I_kwDODunzps5N6gf7
4,696
Cannot load LinCE dataset
{ "login": "finiteautomata", "id": 167943, "node_id": "MDQ6VXNlcjE2Nzk0Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/167943?v=4", "gravatar_id": "", "url": "https://api.github.com/users/finiteautomata", "html_url": "https://github.com/finiteautomata", "followers_url": "https://api.github.com/users/finiteautomata/followers", "following_url": "https://api.github.com/users/finiteautomata/following{/other_user}", "gists_url": "https://api.github.com/users/finiteautomata/gists{/gist_id}", "starred_url": "https://api.github.com/users/finiteautomata/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/finiteautomata/subscriptions", "organizations_url": "https://api.github.com/users/finiteautomata/orgs", "repos_url": "https://api.github.com/users/finiteautomata/repos", "events_url": "https://api.github.com/users/finiteautomata/events{/privacy}", "received_events_url": "https://api.github.com/users/finiteautomata/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
[ "Hi @finiteautomata, thanks for reporting.\r\n\r\nUnfortunately, I'm not able to reproduce your issue:\r\n```python\r\nIn [1]: from datasets import load_dataset\r\n ...: dataset = load_dataset(\"lince\", \"ner_spaeng\")\r\nDownloading builder script: 20.8kB [00:00, 9.09MB/s] \r\nDownloading metadata: 31.2kB [00:00, 13.5MB/s] \r\nDownloading and preparing dataset lince/ner_spaeng (download: 2.93 MiB, generated: 18.45 MiB, post-processed: Unknown size, total: 21.38 MiB) to .../.cache/huggingface/datasets/lince/ner_spaeng/1.0.0/10d41747f55f0849fa84ac579ea1acfa7df49aa2015b60426bc459c111b3d589...\r\nDownloading data: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3.08M/3.08M [00:01<00:00, 2.73MB/s]\r\nDataset lince downloaded and prepared to .../.cache/huggingface/datasets/lince/ner_spaeng/1.0.0/10d41747f55f0849fa84ac579ea1acfa7df49aa2015b60426bc459c111b3d589. Subsequent calls will reuse this data.\r\n100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3/3 [00:00<00:00, 630.66it/s]\r\n\r\nIn [2]: dataset\r\nOut[2]: \r\nDatasetDict({\r\n train: Dataset({\r\n features: ['idx', 'words', 'lid', 'ner'],\r\n num_rows: 33611\r\n })\r\n validation: Dataset({\r\n features: ['idx', 'words', 'lid', 'ner'],\r\n num_rows: 10085\r\n })\r\n test: Dataset({\r\n features: ['idx', 'words', 'lid', 'ner'],\r\n num_rows: 23527\r\n })\r\n})\r\n``` \r\n\r\nPlease note that for this dataset, the original data files are not hosted on the Hugging Face Hub, but on https://ritual.uh.edu\r\nAnd sometimes, the server might be temporarily unavailable, as your error message said (trying to connect to the server timed out):\r\n```\r\nConnectionError: Couldn't reach https://ritual.uh.edu/lince/libaccess/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9/ner_spaeng.zip (ConnectTimeout(MaxRetryError(\"HTTPSConnectionPool(host='ritual.uh.edu', port=443): Max retries exceeded with url: /lince/libaccess/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9/ner_spaeng.zip (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7feb1c45a690>, 'Connection to ritual.uh.edu timed out. (connect timeout=100)'))\")))\r\n```\r\nIn these cases you could:\r\n- either contact the owners of the data server where the data is hosted to inform them about the issue in their server\r\n- or re-try after waiting some time: usually these issues are just temporary", "Great, thanks for checking out!" ]
1,658,084,514,000
1,658,136,040,000
1,658,129,062,000
NONE
null
## Describe the bug Cannot load LinCE dataset due to a connection error ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("lince", "ner_spaeng") ``` A notebook with this code and corresponding error can be found at https://colab.research.google.com/drive/1pgX3bNB9amuUwAVfPFm-XuMV5fEg-cD2 ## Expected results It should load the dataset ## Actual results ```python --------------------------------------------------------------------------- ConnectionError Traceback (most recent call last) <ipython-input-2-fc551ddcebef> in <module>() 1 from datasets import load_dataset 2 ----> 3 dataset = load_dataset("lince", "ner_spaeng") 10 frames /usr/local/lib/python3.7/dist-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) 1682 ignore_verifications=ignore_verifications, 1683 try_from_hf_gcs=try_from_hf_gcs, -> 1684 use_auth_token=use_auth_token, 1685 ) 1686 /usr/local/lib/python3.7/dist-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs) 703 if not downloaded_from_gcs: 704 self._download_and_prepare( --> 705 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 706 ) 707 # Sync info /usr/local/lib/python3.7/dist-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos) 1219 1220 def _download_and_prepare(self, dl_manager, verify_infos): -> 1221 super()._download_and_prepare(dl_manager, verify_infos, check_duplicate_keys=verify_infos) 1222 1223 def _get_examples_iterable_for_split(self, split_generator: SplitGenerator) -> ExamplesIterable: /usr/local/lib/python3.7/dist-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 769 split_dict = SplitDict(dataset_name=self.name) 770 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) --> 771 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) 772 773 # Checksums verification /root/.cache/huggingface/modules/datasets_modules/datasets/lince/10d41747f55f0849fa84ac579ea1acfa7df49aa2015b60426bc459c111b3d589/lince.py in _split_generators(self, dl_manager) 481 def _split_generators(self, dl_manager): 482 """Returns SplitGenerators.""" --> 483 lince_dir = dl_manager.download_and_extract(f"{_LINCE_URL}/{self.config.name}.zip") 484 data_dir = os.path.join(lince_dir, self.config.data_dir) 485 return [ /usr/local/lib/python3.7/dist-packages/datasets/download/download_manager.py in download_and_extract(self, url_or_urls) 429 extracted_path(s): `str`, extracted paths of given URL(s). 430 """ --> 431 return self.extract(self.download(url_or_urls)) 432 433 def get_recorded_sizes_checksums(self): /usr/local/lib/python3.7/dist-packages/datasets/download/download_manager.py in download(self, url_or_urls) 313 num_proc=download_config.num_proc, 314 disable_tqdm=not is_progress_bar_enabled(), --> 315 desc="Downloading data files", 316 ) 317 duration = datetime.now() - start_time /usr/local/lib/python3.7/dist-packages/datasets/utils/py_utils.py in map_nested(function, data_struct, dict_only, map_list, map_tuple, map_numpy, num_proc, types, disable_tqdm, desc) 346 # Singleton 347 if not isinstance(data_struct, dict) and not isinstance(data_struct, types): --> 348 return function(data_struct) 349 350 disable_tqdm = disable_tqdm or not logging.is_progress_bar_enabled() /usr/local/lib/python3.7/dist-packages/datasets/download/download_manager.py in _download(self, url_or_filename, download_config) 333 # append the relative path to the base_path 334 url_or_filename = url_or_path_join(self._base_path, url_or_filename) --> 335 return cached_path(url_or_filename, download_config=download_config) 336 337 def iter_archive(self, path_or_buf: Union[str, io.BufferedReader]): /usr/local/lib/python3.7/dist-packages/datasets/utils/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs) 195 use_auth_token=download_config.use_auth_token, 196 ignore_url_params=download_config.ignore_url_params, --> 197 download_desc=download_config.download_desc, 198 ) 199 elif os.path.exists(url_or_filename): /usr/local/lib/python3.7/dist-packages/datasets/utils/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token, ignore_url_params, download_desc) 531 _raise_if_offline_mode_is_enabled(f"Tried to reach {url}") 532 if head_error is not None: --> 533 raise ConnectionError(f"Couldn't reach {url} ({repr(head_error)})") 534 elif response is not None: 535 raise ConnectionError(f"Couldn't reach {url} (error {response.status_code})") ConnectionError: Couldn't reach https://ritual.uh.edu/lince/libaccess/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9/ner_spaeng.zip (ConnectTimeout(MaxRetryError("HTTPSConnectionPool(host='ritual.uh.edu', port=443): Max retries exceeded with url: /lince/libaccess/eyJ1c2VybmFtZSI6ICJodWdnaW5nZmFjZSBubHAiLCAidXNlcl9pZCI6IDExMSwgImVtYWlsIjogImR1bW15QGVtYWlsLmNvbSJ9/ner_spaeng.zip (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x7feb1c45a690>, 'Connection to ritual.uh.edu timed out. (connect timeout=100)'))"))) ``` ## Environment info - `datasets` version: 2.3.2 - Platform: Linux-5.4.188+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - PyArrow version: 6.0.1 - Pandas version: 1.3.5
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4696/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4696/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4695
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4695/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4695/comments
https://api.github.com/repos/huggingface/datasets/issues/4695/events
https://github.com/huggingface/datasets/pull/4695
1,307,134,701
PR_kwDODunzps47hobQ
4,695
Add MANtIS dataset
{ "login": "bhavitvyamalik", "id": 19718818, "node_id": "MDQ6VXNlcjE5NzE4ODE4", "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bhavitvyamalik", "html_url": "https://github.com/bhavitvyamalik", "followers_url": "https://api.github.com/users/bhavitvyamalik/followers", "following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}", "gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}", "starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions", "organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs", "repos_url": "https://api.github.com/users/bhavitvyamalik/repos", "events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}", "received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4695). All of your documentation changes will be reflected on that endpoint." ]
1,658,073,185,000
1,658,073,615,000
null
CONTRIBUTOR
null
This PR adds MANtIS dataset. Arxiv: [https://arxiv.org/abs/1912.04639](https://arxiv.org/abs/1912.04639) Github: [https://github.com/Guzpenha/MANtIS](https://github.com/Guzpenha/MANtIS) README and dataset tags are WIP.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4695/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4695/timeline
null
null
true
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4695", "html_url": "https://github.com/huggingface/datasets/pull/4695", "diff_url": "https://github.com/huggingface/datasets/pull/4695.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4695.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4694
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4694/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4694/comments
https://api.github.com/repos/huggingface/datasets/issues/4694/events
https://github.com/huggingface/datasets/issues/4694
1,306,958,380
I_kwDODunzps5N5pos
4,694
Distributed data parallel training for streaming datasets
{ "login": "cyk1337", "id": 13767887, "node_id": "MDQ6VXNlcjEzNzY3ODg3", "avatar_url": "https://avatars.githubusercontent.com/u/13767887?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cyk1337", "html_url": "https://github.com/cyk1337", "followers_url": "https://api.github.com/users/cyk1337/followers", "following_url": "https://api.github.com/users/cyk1337/following{/other_user}", "gists_url": "https://api.github.com/users/cyk1337/gists{/gist_id}", "starred_url": "https://api.github.com/users/cyk1337/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cyk1337/subscriptions", "organizations_url": "https://api.github.com/users/cyk1337/orgs", "repos_url": "https://api.github.com/users/cyk1337/repos", "events_url": "https://api.github.com/users/cyk1337/events{/privacy}", "received_events_url": "https://api.github.com/users/cyk1337/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[]
1,658,021,383,000
1,658,021,383,000
null
NONE
null
### Feature request Any documentations for the the `load_dataset(streaming=True)` for (multi-node multi-GPU) DDP training? ### Motivation Given a bunch of data files, it is expected to split them onto different GPUs. Is there a guide or documentation? ### Your contribution Does it requires manually split on data files for each worker in `DatasetBuilder._split_generator()`? What is`IterableDatasetShard` expected to do?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4694/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4694/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4693
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4693/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4693/comments
https://api.github.com/repos/huggingface/datasets/issues/4693/events
https://github.com/huggingface/datasets/pull/4693
1,306,788,322
PR_kwDODunzps47go-F
4,693
update `samsum` script
{ "login": "bhavitvyamalik", "id": 19718818, "node_id": "MDQ6VXNlcjE5NzE4ODE4", "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bhavitvyamalik", "html_url": "https://github.com/bhavitvyamalik", "followers_url": "https://api.github.com/users/bhavitvyamalik/followers", "following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}", "gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}", "starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions", "organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs", "repos_url": "https://api.github.com/users/bhavitvyamalik/repos", "events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}", "received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4693). All of your documentation changes will be reflected on that endpoint." ]
1,657,972,385,000
1,658,231,143,000
null
CONTRIBUTOR
null
update `samsum` script after #4672 was merged (citation is also updated)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4693/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4693/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4693", "html_url": "https://github.com/huggingface/datasets/pull/4693", "diff_url": "https://github.com/huggingface/datasets/pull/4693.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4693.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4692
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4692/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4692/comments
https://api.github.com/repos/huggingface/datasets/issues/4692/events
https://github.com/huggingface/datasets/issues/4692
1,306,609,680
I_kwDODunzps5N4UgQ
4,692
Unable to cast a column with `Image()` by using the `cast_column()` feature
{ "login": "skrishnan99", "id": 28833916, "node_id": "MDQ6VXNlcjI4ODMzOTE2", "avatar_url": "https://avatars.githubusercontent.com/u/28833916?v=4", "gravatar_id": "", "url": "https://api.github.com/users/skrishnan99", "html_url": "https://github.com/skrishnan99", "followers_url": "https://api.github.com/users/skrishnan99/followers", "following_url": "https://api.github.com/users/skrishnan99/following{/other_user}", "gists_url": "https://api.github.com/users/skrishnan99/gists{/gist_id}", "starred_url": "https://api.github.com/users/skrishnan99/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/skrishnan99/subscriptions", "organizations_url": "https://api.github.com/users/skrishnan99/orgs", "repos_url": "https://api.github.com/users/skrishnan99/repos", "events_url": "https://api.github.com/users/skrishnan99/events{/privacy}", "received_events_url": "https://api.github.com/users/skrishnan99/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi, thanks for reporting! A PR (https://github.com/huggingface/datasets/pull/4614) has already been opened to address this issue." ]
1,657,925,763,000
1,658,237,784,000
1,658,237,784,000
NONE
null
## Describe the bug A clear and concise description of what the bug is. When I create a dataset, then add a column to the created dataset through the `dataset.add_column` feature and then try to cast a column of the dataset (this column contains image paths) with `Image()` by using the `cast_column()` feature, I get the following error - ``` TypeError: Couldn't cast array of type string to {'bytes': Value(dtype='binary', id=None), 'path': Value(dtype='string', id=None)} ``` When I try and cast the same column, but without doing the `add_column` in the previous step, it works as expected. ## Steps to reproduce the bug ```python from datasets import Dataset, Image data_dict = { "img_path": ["https://picsum.photos/200/300"] } dataset = Dataset.from_dict(data_dict) #NOTE Comment out this line and use cast_column and it works properly dataset = dataset.add_column("yeet", [1]) #NOTE This line fails to execute properly if `add_column` is called before dataset = dataset.cast_column("img_path", Image()) # #NOTE This is my current workaround. This seems to work fine with/without `add_column`. While # # running this, make sure to comment out the `cast_column` line # new_features = dataset.features.copy() # new_features["img_path"] = Image() # dataset = dataset.cast(new_features) print(dataset) print(dataset.features) print(dataset[0]) ``` ## Expected results A clear and concise description of the expected results. Able to successfully use `cast_column` to cast a column containing img_paths to now be Image() features after modifying the dataset using `add_column` in a previous step ## Actual results Specify the actual results or traceback. ``` Traceback (most recent call last): File "/home/surya/Desktop/hf_bug_test.py", line 14, in <module> dataset = dataset.cast_column("img_path", Image()) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/fingerprint.py", line 458, in wrapper out = func(self, *args, **kwargs) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 1580, in cast_column dataset._data = dataset._data.cast(dataset.features.arrow_schema) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1487, in cast new_tables.append(subtable.cast(subschema, *args, **kwargs)) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 834, in cast return InMemoryTable(table_cast(self.table, *args, **kwargs)) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1897, in table_cast return cast_table_to_schema(table, schema) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1880, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1880, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1673, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1673, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1846, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") TypeError: Couldn't cast array of type string to {'bytes': Value(dtype='binary', id=None), 'path': Value(dtype='string', id=None)} ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.3.2 - Platform: Ubuntu 20.04.3 LTS - Python version: 3.9.7 - PyArrow version: 7.0.0
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4692/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4692/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4691
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4691/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4691/comments
https://api.github.com/repos/huggingface/datasets/issues/4691/events
https://github.com/huggingface/datasets/issues/4691
1,306,389,656
I_kwDODunzps5N3eyY
4,691
Dataset Viewer issue for rajistics/indian_food_images
{ "login": "rajshah4", "id": 6808012, "node_id": "MDQ6VXNlcjY4MDgwMTI=", "avatar_url": "https://avatars.githubusercontent.com/u/6808012?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rajshah4", "html_url": "https://github.com/rajshah4", "followers_url": "https://api.github.com/users/rajshah4/followers", "following_url": "https://api.github.com/users/rajshah4/following{/other_user}", "gists_url": "https://api.github.com/users/rajshah4/gists{/gist_id}", "starred_url": "https://api.github.com/users/rajshah4/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rajshah4/subscriptions", "organizations_url": "https://api.github.com/users/rajshah4/orgs", "repos_url": "https://api.github.com/users/rajshah4/repos", "events_url": "https://api.github.com/users/rajshah4/events{/privacy}", "received_events_url": "https://api.github.com/users/rajshah4/received_events", "type": "User", "site_admin": false }
[ { "id": 3470211881, "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer", "name": "dataset-viewer", "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co" } ]
closed
false
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "organizations_url": "https://api.github.com/users/severo/orgs", "repos_url": "https://api.github.com/users/severo/repos", "events_url": "https://api.github.com/users/severo/events{/privacy}", "received_events_url": "https://api.github.com/users/severo/received_events", "type": "User", "site_admin": false }
[ { "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "organizations_url": "https://api.github.com/users/severo/orgs", "repos_url": "https://api.github.com/users/severo/repos", "events_url": "https://api.github.com/users/severo/events{/privacy}", "received_events_url": "https://api.github.com/users/severo/received_events", "type": "User", "site_admin": false } ]
null
[ "Hi, thanks for reporting. I triggered a refresh of the preview for this dataset, and it works now. I'm not sure what occurred.\r\n<img width=\"1019\" alt=\"Capture d’écran 2022-07-18 à 11 01 52\" src=\"https://user-images.githubusercontent.com/1676121/179541327-f62ecd5e-a18a-4d91-b316-9e2ebde77a28.png\">\r\n\r\n" ]
1,657,911,795,000
1,658,156,523,000
1,658,156,523,000
NONE
null
### Link https://huggingface.co/datasets/rajistics/indian_food_images/viewer/rajistics--indian_food_images/train ### Description I have a train/test split in my dataset <img width="410" alt="Screen Shot 2022-07-15 at 11 44 42 AM" src="https://user-images.githubusercontent.com/6808012/179293215-7b419ec3-3527-46f2-8dad-adbc5568cfa0.png"> t The dataset viewer works for the test split (images of indian food), but does not show my train split. My guess is maybe there is some corrupt image file that is guessing this. But I have no idea. The original dataset was pulled from here: https://www.kaggle.com/datasets/l33tc0d3r/indian-food-classification?resource=download-directory ### Owner Yes
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4691/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4691/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4690
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4690/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4690/comments
https://api.github.com/repos/huggingface/datasets/issues/4690/events
https://github.com/huggingface/datasets/pull/4690
1,306,321,975
PR_kwDODunzps47fG6w
4,690
Refactor base extractors
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
1,657,907,268,000
1,658,134,016,000
1,658,133,289,000
MEMBER
null
This PR: - Refactors base extractors as subclasses of `BaseExtractor`: - this is an abstract class defining the interface with: - `is_extractable`: abstract class method - `extract`: abstract static method - Implements abstract `MagicNumberBaseExtractor` (as subclass of `BaseExtractor`): - this has a default implementation of `is_extractable` - this improves performance (reducing the number of file reads) by allowing passing already read `magic_number` - Refactors `Extractor`: - reads magic number from file only once This PR deprecates: ```python is_extractable, extractor = self.extractor.is_extractable(input_path, return_extractor=True) self.extractor.extract(input_path, output_path, extractor=extractor) ``` and uses more Pythonic instead: ```python extractor_format = self.extractor.infer_extractor_format(input_path) self.extractor.extract(input_path, output_path, extractor_format) ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4690/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4690/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4690", "html_url": "https://github.com/huggingface/datasets/pull/4690", "diff_url": "https://github.com/huggingface/datasets/pull/4690.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4690.patch", "merged_at": 1658133289000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4689
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4689/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4689/comments
https://api.github.com/repos/huggingface/datasets/issues/4689/events
https://github.com/huggingface/datasets/pull/4689
1,306,230,203
PR_kwDODunzps47eyw5
4,689
Test extractors for all compression formats
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
1,657,902,595,000
1,657,907,222,000
1,657,906,524,000
MEMBER
null
This PR: - Adds all compression formats to `test_extractor` - Tests each base extractor for all compression formats Note that all compression formats are tested except "rar".
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4689/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4689/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4689", "html_url": "https://github.com/huggingface/datasets/pull/4689", "diff_url": "https://github.com/huggingface/datasets/pull/4689.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4689.patch", "merged_at": 1657906524000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4688
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4688/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4688/comments
https://api.github.com/repos/huggingface/datasets/issues/4688/events
https://github.com/huggingface/datasets/pull/4688
1,306,100,488
PR_kwDODunzps47eW6C
4,688
Skip test_extractor only for zstd param if zstandard not installed
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
1,657,895,027,000
1,657,898,873,000
1,657,898,124,000
MEMBER
null
Currently, if `zstandard` is not installed, `test_extractor` is skipped for all compression format parameters. This PR fixes `test_extractor` so that if `zstandard` is not installed, `test_extractor` is skipped only for the `zstd` compression parameter, that is, it is not skipped for all the other compression parameters (`gzip`, `xz`,...).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4688/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4688/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4688", "html_url": "https://github.com/huggingface/datasets/pull/4688", "diff_url": "https://github.com/huggingface/datasets/pull/4688.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4688.patch", "merged_at": 1657898124000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4687
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4687/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4687/comments
https://api.github.com/repos/huggingface/datasets/issues/4687/events
https://github.com/huggingface/datasets/pull/4687
1,306,021,415
PR_kwDODunzps47eF_E
4,687
Trigger CI also on push to main
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
1,657,890,689,000
1,657,892,841,000
1,657,892,123,000
MEMBER
null
Currently, new CI (on GitHub Actions) is only triggered on pull requests branches when the base branch is main. This PR also triggers the CI when a PR is merged to main branch.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4687/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4687/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4687", "html_url": "https://github.com/huggingface/datasets/pull/4687", "diff_url": "https://github.com/huggingface/datasets/pull/4687.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4687.patch", "merged_at": 1657892123000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4686
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4686/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4686/comments
https://api.github.com/repos/huggingface/datasets/issues/4686/events
https://github.com/huggingface/datasets/pull/4686
1,305,974,924
PR_kwDODunzps47d8Jf
4,686
Align logging with Transformers (again)
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4686). All of your documentation changes will be reflected on that endpoint.", "I wasn't aware of https://github.com/huggingface/datasets/pull/1845 before opening this PR. This issue seems much more complex now ..." ]
1,657,887,869,000
1,657,898,863,000
null
CONTRIBUTOR
null
Fix #2832
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4686/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4686/timeline
null
null
true
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4686", "html_url": "https://github.com/huggingface/datasets/pull/4686", "diff_url": "https://github.com/huggingface/datasets/pull/4686.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4686.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4685
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4685/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4685/comments
https://api.github.com/repos/huggingface/datasets/issues/4685/events
https://github.com/huggingface/datasets/pull/4685
1,305,861,708
PR_kwDODunzps47dju8
4,685
Fix mock fsspec
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
1,657,880,592,000
1,657,890,303,000
1,657,889,560,000
MEMBER
null
This PR: - Removes an unused method from `DummyTestFS` - Refactors `mock_fsspec` to make it simpler
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4685/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4685/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4685", "html_url": "https://github.com/huggingface/datasets/pull/4685", "diff_url": "https://github.com/huggingface/datasets/pull/4685.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4685.patch", "merged_at": 1657889560000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4684
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4684/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4684/comments
https://api.github.com/repos/huggingface/datasets/issues/4684/events
https://github.com/huggingface/datasets/issues/4684
1,305,554,654
I_kwDODunzps5N0S7e
4,684
How to assign new values to Dataset?
{ "login": "beyondguo", "id": 37113676, "node_id": "MDQ6VXNlcjM3MTEzNjc2", "avatar_url": "https://avatars.githubusercontent.com/u/37113676?v=4", "gravatar_id": "", "url": "https://api.github.com/users/beyondguo", "html_url": "https://github.com/beyondguo", "followers_url": "https://api.github.com/users/beyondguo/followers", "following_url": "https://api.github.com/users/beyondguo/following{/other_user}", "gists_url": "https://api.github.com/users/beyondguo/gists{/gist_id}", "starred_url": "https://api.github.com/users/beyondguo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/beyondguo/subscriptions", "organizations_url": "https://api.github.com/users/beyondguo/orgs", "repos_url": "https://api.github.com/users/beyondguo/repos", "events_url": "https://api.github.com/users/beyondguo/events{/privacy}", "received_events_url": "https://api.github.com/users/beyondguo/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Hi! One option is use `map` with a function that overwrites the labels (`dset = dset.map(lamba _: {\"label\": 0}, features=dset.features`)). Or you can use the `remove_column` + `add_column` combination (`dset = dset.remove_columns(\"label\").add_column(\"label\", [0]*len(data)).cast(dset.features)`, but note that this approach creates an in-memory table for the added column instead of writing to disk, which could be problematic for large datasets." ]
1,657,858,677,000
1,657,901,901,000
null
NONE
null
![image](https://user-images.githubusercontent.com/37113676/179149159-bbbda0c8-a661-403c-87ed-dc2b4219cd68.png) Hi, if I want to change some values of the dataset, or add new columns to it, how can I do it? For example, I want to change all the labels of the SST2 dataset to `0`: ```python from datasets import load_dataset data = load_dataset('glue','sst2') data['train']['label'] = [0]*len(data) ``` I will get the error: ``` TypeError: 'Dataset' object does not support item assignment ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4684/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4684/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4683
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4683/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4683/comments
https://api.github.com/repos/huggingface/datasets/issues/4683/events
https://github.com/huggingface/datasets/pull/4683
1,305,443,253
PR_kwDODunzps47cLkm
4,683
Update create dataset card docs
{ "login": "stevhliu", "id": 59462357, "node_id": "MDQ6VXNlcjU5NDYyMzU3", "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stevhliu", "html_url": "https://github.com/stevhliu", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "repos_url": "https://api.github.com/users/stevhliu/repos", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
1,657,845,689,000
1,658,165,160,000
1,658,150,650,000
MEMBER
null
This PR proposes removing the [online dataset card creator](https://huggingface.co/datasets/card-creator/) in favor of simply copy/pasting a template and using the [Datasets Tagger app](https://huggingface.co/spaces/huggingface/datasets-tagging) to generate the tags. The Tagger app provides more guidance by showing all possible values a user can select in the dropdown menus, whereas the online dataset card creator doesn't, which can make it difficult to know what tag values to input. Let me know what you think! :)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4683/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4683/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4683", "html_url": "https://github.com/huggingface/datasets/pull/4683", "diff_url": "https://github.com/huggingface/datasets/pull/4683.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4683.patch", "merged_at": 1658150650000 }
true
https://api.github.com/repos/huggingface/datasets/issues/4682
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4682/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4682/comments
https://api.github.com/repos/huggingface/datasets/issues/4682/events
https://github.com/huggingface/datasets/issues/4682
1,304,788,215
I_kwDODunzps5NxXz3
4,682
weird issue/bug with columns (dataset iterable/stream mode)
{ "login": "eunseojo", "id": 12104720, "node_id": "MDQ6VXNlcjEyMTA0NzIw", "avatar_url": "https://avatars.githubusercontent.com/u/12104720?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eunseojo", "html_url": "https://github.com/eunseojo", "followers_url": "https://api.github.com/users/eunseojo/followers", "following_url": "https://api.github.com/users/eunseojo/following{/other_user}", "gists_url": "https://api.github.com/users/eunseojo/gists{/gist_id}", "starred_url": "https://api.github.com/users/eunseojo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eunseojo/subscriptions", "organizations_url": "https://api.github.com/users/eunseojo/orgs", "repos_url": "https://api.github.com/users/eunseojo/repos", "events_url": "https://api.github.com/users/eunseojo/events{/privacy}", "received_events_url": "https://api.github.com/users/eunseojo/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[]
1,657,805,207,000
1,657,805,207,000
null
NONE
null
I have a dataset online (CloverSearch/cc-news-mutlilingual) that has a bunch of columns, two of which are "score_title_maintext" and "score_title_description". the original files are jsonl formatted. I was trying to iterate through via streaming mode and grab all "score_title_description" values, but I kept getting key not found after a certain point of iteration. I found that some json objects in the file don't have "score_title_description". And in SOME cases, this returns a NONE and in others it just gets a key error. Why is there an inconsistency here and how can I fix it?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4682/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4682/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4681
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4681/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4681/comments
https://api.github.com/repos/huggingface/datasets/issues/4681/events
https://github.com/huggingface/datasets/issues/4681
1,304,617,484
I_kwDODunzps5NwuIM
4,681
IndexError when loading ImageFolder
{ "login": "johko", "id": 2843485, "node_id": "MDQ6VXNlcjI4NDM0ODU=", "avatar_url": "https://avatars.githubusercontent.com/u/2843485?v=4", "gravatar_id": "", "url": "https://api.github.com/users/johko", "html_url": "https://github.com/johko", "followers_url": "https://api.github.com/users/johko/followers", "following_url": "https://api.github.com/users/johko/following{/other_user}", "gists_url": "https://api.github.com/users/johko/gists{/gist_id}", "starred_url": "https://api.github.com/users/johko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/johko/subscriptions", "organizations_url": "https://api.github.com/users/johko/orgs", "repos_url": "https://api.github.com/users/johko/repos", "events_url": "https://api.github.com/users/johko/events{/privacy}", "received_events_url": "https://api.github.com/users/johko/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
[ { "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false } ]
null
[]
1,657,796,275,000
1,657,814,846,000
null
NONE
null
## Describe the bug Loading an image dataset with `imagefolder` throws `IndexError: list index out of range` when the given folder contains a non-image file (like a csv). ## Steps to reproduce the bug Put a csv file in a folder with images and load it: ```python import datasets datasets.load_dataset("imagefolder", data_dir=path/to/folder) ``` ## Expected results I would expect a better error message, like `Unsupported file` or even the dataset loader just ignoring every file that is not an image in that case. ## Actual results Here is the whole traceback: ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.3.2 - Platform: Linux-5.11.0-051100-generic-x86_64-with-glibc2.27 - Python version: 3.9.9 - PyArrow version: 8.0.0 - Pandas version: 1.4.3
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4681/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4681/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4680
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4680/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4680/comments
https://api.github.com/repos/huggingface/datasets/issues/4680/events
https://github.com/huggingface/datasets/issues/4680
1,304,534,770
I_kwDODunzps5NwZ7y
4,680
Dataset Viewer issue for codeparrot/xlcost-text-to-code
{ "login": "loubnabnl", "id": 44069155, "node_id": "MDQ6VXNlcjQ0MDY5MTU1", "avatar_url": "https://avatars.githubusercontent.com/u/44069155?v=4", "gravatar_id": "", "url": "https://api.github.com/users/loubnabnl", "html_url": "https://github.com/loubnabnl", "followers_url": "https://api.github.com/users/loubnabnl/followers", "following_url": "https://api.github.com/users/loubnabnl/following{/other_user}", "gists_url": "https://api.github.com/users/loubnabnl/gists{/gist_id}", "starred_url": "https://api.github.com/users/loubnabnl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/loubnabnl/subscriptions", "organizations_url": "https://api.github.com/users/loubnabnl/orgs", "repos_url": "https://api.github.com/users/loubnabnl/repos", "events_url": "https://api.github.com/users/loubnabnl/events{/privacy}", "received_events_url": "https://api.github.com/users/loubnabnl/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "There seems to be an issue with the `C++-snippet-level` config:\r\n\r\n```python\r\n>>> from datasets import get_dataset_split_names\r\n>>> get_dataset_split_names(\"codeparrot/xlcost-text-to-code\", \"C++-snippet-level\")\r\nTraceback (most recent call last):\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 352, in get_dataset_config_info\r\n info.splits = {\r\nTypeError: 'NoneType' object is not iterable\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 404, in get_dataset_split_names\r\n info = get_dataset_config_info(\r\n File \"/home/slesage/hf/datasets-server/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 359, in get_dataset_config_info\r\n raise SplitsNotFoundError(\"The split names could not be parsed from the dataset config.\") from err\r\ndatasets.inspect.SplitsNotFoundError: The split names could not be parsed from the dataset config.\r\n```\r\n\r\nI remove the dataset-viewer tag since it's not directly related.\r\n\r\nPinging @huggingface/datasets ", "Thanks I found that this subset wasn't properly defined the the config, I fixed it. Now I can see the subsets but I get this error for the viewer\r\n````\r\nStatus code: 400\r\nException: Status400Error\r\nMessage: The split cache is empty.\r\n```", "Yes, the cache is being refreshed, hopefully, it will work in some minutes for all the splits. Some are already here:\r\n\r\nhttps://huggingface.co/datasets/codeparrot/xlcost-text-to-code/viewer/Python-snippet-level/train\r\n\r\n<img width=\"1533\" alt=\"Capture d’écran 2022-07-18 à 12 04 06\" src=\"https://user-images.githubusercontent.com/1676121/179553933-64d874fa-ada9-4b82-900e-082619523c20.png\">\r\n", "I think all the splits are working as expected now", "Perfect, thank you!" ]
1,657,791,950,000
1,658,162,220,000
1,658,160,276,000
NONE
null
### Link https://huggingface.co/datasets/codeparrot/xlcost-text-to-code ### Description Error ``` Server Error Status code: 400 Exception: TypeError Message: 'NoneType' object is not iterable ``` Before I did a minor change in the dataset script (removing some comments), the viewer was working but not properely, it wasn't showing the dataset subsets. But the data can be loaded successfully. Thanks! ### Owner Yes
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4680/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4680/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4679
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4679/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4679/comments
https://api.github.com/repos/huggingface/datasets/issues/4679/events
https://github.com/huggingface/datasets/pull/4679
1,303,980,648
PR_kwDODunzps47XX67
4,679
Added method to remove excess nesting in a DatasetDict
{ "login": "CakeCrusher", "id": 37946988, "node_id": "MDQ6VXNlcjM3OTQ2OTg4", "avatar_url": "https://avatars.githubusercontent.com/u/37946988?v=4", "gravatar_id": "", "url": "https://api.github.com/users/CakeCrusher", "html_url": "https://github.com/CakeCrusher", "followers_url": "https://api.github.com/users/CakeCrusher/followers", "following_url": "https://api.github.com/users/CakeCrusher/following{/other_user}", "gists_url": "https://api.github.com/users/CakeCrusher/gists{/gist_id}", "starred_url": "https://api.github.com/users/CakeCrusher/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CakeCrusher/subscriptions", "organizations_url": "https://api.github.com/users/CakeCrusher/orgs", "repos_url": "https://api.github.com/users/CakeCrusher/repos", "events_url": "https://api.github.com/users/CakeCrusher/events{/privacy}", "received_events_url": "https://api.github.com/users/CakeCrusher/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Hi ! I think the issue you linked is closed and suggests to use `remove_columns`.\r\n\r\nMoreover if you end up with a dataset with an unnecessarily nested data, please modify your processing functions to not output nested data, or use `map(..., batched=True)` if you function take batches as input", "Hi @lhoestq , you are right about the issues this pull has steered beyond that issue. I created this [colab notebook](https://colab.research.google.com/drive/16aLu6QrDSV_aUYRdpufl5E4iS08qkUGj?usp=sharing) to present the error. I tried using batch and that won't resolve it either. I'm looking into that error right now.", "I think you just need to pass one example at a time to your tokenizer, this way you don't end up with nested data:\r\n```python\r\n\r\ndef preprocessFunction(row):\r\n collatedContext = tokenizer.eos_token.join([row[\"context\"+str(i+1)] for i in range(int(AMT_OF_CONTEXT))])\r\n response = row[\"response\"]\r\n tokenizedContext = tokenizer(\r\n collatedContext, max_length=max_context_length, truncation=True # don't pass as a list here\r\n )\r\n with tokenizer.as_target_tokenizer():\r\n tokenized_response = tokenizer(\r\n response, max_length=max_response_length, truncation=True # don't pass a a list here\r\n )\r\n tokenizedContext[\"labels\"] = tokenized_response[\"input_ids\"]\r\n return tokenizedContext\r\n```", "Yes that is correct, the purpose of this pull is to advise of a more general solution like with `def remove_excess_nesting(self)` or maybe automate the solution (stas00 advised not to automate it as it could \"not be backwards compatible\").", "I'm not sure I understand how having `remove_excess_nesting` would make more sense than just fixing the preprocessFunction to simply not return nested samples, can you elaborate ?", "Figuring out the issue can be a bit difficult to figure out. Only until I added batch does it make a little more sense with the error\r\n\r\n> sequence item 0: expected str instance, list found\r\n\r\nbut batch was never intended.\r\n\r\nWhen you run the colab you will notice that only until collating do you learn there is this error. So i figured it would be better to address it during at the `DatasetDict` level.\r\nI think it would be ideal if the user could be notified at the preprocess function.", "I'm not arguing that `remove_excess_nesting` is the right solution but what I aim to address is dealing with unnecessary nesting as early as possible.", "> When you run the colab you will notice that only until collating do you learn there is this error.\r\n\r\nI think users can just check the `dataset.features` and they would notice that the data are nested\r\n```python\r\n{\r\n 'input_ids': Sequence(Sequence(feature=Value(dtype='int32', id=None), length=-1, id=None), length=-1, id=None)\r\n ...\r\n}\r\n```\r\n\r\nSometime nested data are intentional, so you can't know in advance if it's a user's mistake or something planned." ]
1,657,748,977,000
1,658,236,877,000
null
NONE
null
Added the ability for a DatasetDict to remove additional nested layers within its features to avoid conflicts when collating. It is meant to accompany [this PR](https://github.com/huggingface/transformers/pull/18119) to resolve the same issue [#15505](https://github.com/huggingface/transformers/issues/15505). @stas00 @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4679/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4679/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4679", "html_url": "https://github.com/huggingface/datasets/pull/4679", "diff_url": "https://github.com/huggingface/datasets/pull/4679.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4679.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/4678
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4678/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4678/comments
https://api.github.com/repos/huggingface/datasets/issues/4678/events
https://github.com/huggingface/datasets/issues/4678
1,303,741,432
I_kwDODunzps5NtYP4
4,678
Cant pass streaming dataset to dataloader after take()
{ "login": "zankner", "id": 39166683, "node_id": "MDQ6VXNlcjM5MTY2Njgz", "avatar_url": "https://avatars.githubusercontent.com/u/39166683?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zankner", "html_url": "https://github.com/zankner", "followers_url": "https://api.github.com/users/zankner/followers", "following_url": "https://api.github.com/users/zankner/following{/other_user}", "gists_url": "https://api.github.com/users/zankner/gists{/gist_id}", "starred_url": "https://api.github.com/users/zankner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zankner/subscriptions", "organizations_url": "https://api.github.com/users/zankner/orgs", "repos_url": "https://api.github.com/users/zankner/repos", "events_url": "https://api.github.com/users/zankner/events{/privacy}", "received_events_url": "https://api.github.com/users/zankner/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[ "Hi! Calling `take` on an iterable/streamable dataset makes it not possible to shard the dataset, which in turn disables multi-process loading (attempts to split the workload over the shards), so to go past this limitation, you can either use single-process loading in `DataLoader` (`num_workers=None`) or fetch the first `50_000/batch_size` batches in the loop." ]
1,657,733,658,000
1,657,804,041,000
null
NONE
null
## Describe the bug I am trying to pass a streaming version of c4 to a dataloader, but it can't be passed after I call `dataset.take(n)`. Some functions such as `shuffle()` can be applied without breaking the dataloader but not take. ## Steps to reproduce the bug ```python import datasets import torch dset = datasets.load_dataset(path='c4', name='en', split="train", streaming=True) dset = dset.take(50_000) dset = dset.with_format("torch") num_workers = 8 batch_size = 512 loader = torch.utils.data.DataLoader(dataset=dset, batch_size=batch_size, num_workers=num_workers) for batch in loader: ... ``` ## Expected results No error thrown when iterating over the dataloader ## Actual results Original Traceback (most recent call last): File "/usr/local/lib/python3.9/dist-packages/torch/utils/data/_utils/worker.py", line 287, in _worker_loop data = fetcher.fetch(index) File "/usr/local/lib/python3.9/dist-packages/torch/utils/data/_utils/fetch.py", line 32, in fetch data.append(next(self.dataset_iter)) File "/root/.local/lib/python3.9/site-packages/datasets/formatting/dataset_wrappers/torch_iterable_dataset.py", line 48, in __iter__ for key, example in self._iter_shard(shard_idx): File "/root/.local/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 586, in _iter_shard yield from ex_iterable.shard_data_sources(shard_idx) File "/root/.local/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 60, in shard_data_sources raise NotImplementedError(f"{type(self)} doesn't implement shard_data_sources yet") NotImplementedError: <class 'datasets.iterable_dataset.TakeExamplesIterable'> doesn't implement shard_data_sources yet ## Environment info - `datasets` version: 2.3.2 - Platform: Linux-5.4.0-120-generic-x86_64-with-glibc2.31 - Python version: 3.9.13 - PyArrow version: 8.0.0 - Pandas version: 1.4.3
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/4678/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/4678/timeline
null
null
null
null
false
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/datasets-cards)

annotations_creators: [] language:

  • en language_creators: [] license: [] multilinguality: [] pretty_name: HuggingFace GitHub Issues size_categories: [] source_datasets: [] tags: [] task_categories:
  • text-classification
  • text-retrieval task_ids:
  • multi-class-classification
  • multi-label-classification
  • document-retrieval
Downloads last month
2
Edit dataset card