url
stringlengths
61
61
repository_url
stringclasses
1 value
labels_url
stringlengths
75
75
comments_url
stringlengths
70
70
events_url
stringlengths
68
68
html_url
stringlengths
51
51
id
int64
1.88B
2.51B
node_id
stringlengths
18
18
number
int64
6.22k
7.14k
title
stringlengths
2
150
user
dict
labels
listlengths
0
2
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
1
milestone
dict
comments
int64
0
17
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
active_lock_reason
null
draft
bool
0 classes
pull_request
dict
body
stringlengths
3
19.4k
closed_by
dict
reactions
dict
timeline_url
stringlengths
70
70
performed_via_github_app
null
state_reason
stringclasses
3 values
is_pull_request
bool
1 class
https://api.github.com/repos/huggingface/datasets/issues/6661
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6661/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6661/comments
https://api.github.com/repos/huggingface/datasets/issues/6661/events
https://github.com/huggingface/datasets/issues/6661
2,132,296,267
I_kwDODunzps5_GEJL
6,661
Import error on Google Colab
{ "login": "kithogue", "id": 16103566, "node_id": "MDQ6VXNlcjE2MTAzNTY2", "avatar_url": "https://avatars.githubusercontent.com/u/16103566?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kithogue", "html_url": "https://github.com/kithogue", "followers_url": "https://api.github.com/users/kithogue/followers", "following_url": "https://api.github.com/users/kithogue/following{/other_user}", "gists_url": "https://api.github.com/users/kithogue/gists{/gist_id}", "starred_url": "https://api.github.com/users/kithogue/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kithogue/subscriptions", "organizations_url": "https://api.github.com/users/kithogue/orgs", "repos_url": "https://api.github.com/users/kithogue/repos", "events_url": "https://api.github.com/users/kithogue/events{/privacy}", "received_events_url": "https://api.github.com/users/kithogue/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-02-13T13:12:40
2024-02-25T16:37:54
2024-02-14T08:04:47
NONE
null
null
null
### Describe the bug Cannot be imported on Google Colab, the import throws the following error: ValueError: pyarrow.lib.IpcWriteOptions size changed, may indicate binary incompatibility. Expected 88 from C header, got 72 from PyObject ### Steps to reproduce the bug 1. `! pip install -U datasets` 2. `import datasets` ### Expected behavior Should be possible to use the library ### Environment info - `datasets` version: 2.17.0 - Platform: Linux-6.1.58+-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.20.3 - PyArrow version: 15.0.0 - Pandas version: 1.5.3 - `fsspec` version: 2023.6.0
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6661/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6661/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6657
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6657/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6657/comments
https://api.github.com/repos/huggingface/datasets/issues/6657/events
https://github.com/huggingface/datasets/issues/6657
2,129,147,085
I_kwDODunzps5-6DTN
6,657
Release not pushed to conda channel
{ "login": "atulsaurav", "id": 7138162, "node_id": "MDQ6VXNlcjcxMzgxNjI=", "avatar_url": "https://avatars.githubusercontent.com/u/7138162?v=4", "gravatar_id": "", "url": "https://api.github.com/users/atulsaurav", "html_url": "https://github.com/atulsaurav", "followers_url": "https://api.github.com/users/atulsaurav/followers", "following_url": "https://api.github.com/users/atulsaurav/following{/other_user}", "gists_url": "https://api.github.com/users/atulsaurav/gists{/gist_id}", "starred_url": "https://api.github.com/users/atulsaurav/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/atulsaurav/subscriptions", "organizations_url": "https://api.github.com/users/atulsaurav/orgs", "repos_url": "https://api.github.com/users/atulsaurav/repos", "events_url": "https://api.github.com/users/atulsaurav/events{/privacy}", "received_events_url": "https://api.github.com/users/atulsaurav/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
5
2024-02-11T20:05:17
2024-03-06T15:06:22
2024-03-06T15:06:22
NONE
null
null
null
### Describe the bug The github actions step to publish the release 2.17.0 to conda channel has failed due to expired token. Can some one please update the anaconda token rerun the failed action? @albertvillanova ? ![image](https://github.com/huggingface/datasets/assets/7138162/1b56ad3d-7643-4778-9cce-4bf531717700) ### Steps to reproduce the bug Please see this actions [link](https://github.com/huggingface/datasets/actions/runs/7842473662) ### Expected behavior The action runs successfully and the latest release is pushed to HuggingFace conda channel ### Environment info Not applicable.
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6657/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6657/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6656
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6656/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6656/comments
https://api.github.com/repos/huggingface/datasets/issues/6656/events
https://github.com/huggingface/datasets/issues/6656
2,127,338,377
I_kwDODunzps5-zJuJ
6,656
Error when loading a big local json file
{ "login": "Riccorl", "id": 10062216, "node_id": "MDQ6VXNlcjEwMDYyMjE2", "avatar_url": "https://avatars.githubusercontent.com/u/10062216?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Riccorl", "html_url": "https://github.com/Riccorl", "followers_url": "https://api.github.com/users/Riccorl/followers", "following_url": "https://api.github.com/users/Riccorl/following{/other_user}", "gists_url": "https://api.github.com/users/Riccorl/gists{/gist_id}", "starred_url": "https://api.github.com/users/Riccorl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Riccorl/subscriptions", "organizations_url": "https://api.github.com/users/Riccorl/orgs", "repos_url": "https://api.github.com/users/Riccorl/repos", "events_url": "https://api.github.com/users/Riccorl/events{/privacy}", "received_events_url": "https://api.github.com/users/Riccorl/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
1
2024-02-09T15:14:21
2024-03-15T22:18:21
null
NONE
null
null
null
### Describe the bug When trying to load big json files from a local directory, `load_dataset` throws the following error ``` Traceback (most recent call last): File "/miniconda3/envs/conda-env/lib/python3.10/site-packages/datasets/builder.py", line 1989, in _prepare_split_single writer.write_table(table) File "miniconda3/envs/conda-env/lib/python3.10/site-packages/datasets/arrow_writer.py", line 573, in write_table pa_table = pa_table.combine_chunks() File "pyarrow/table.pxi", line 3638, in pyarrow.lib.Table.combine_chunks File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays ``` ### Steps to reproduce the bug 1. Download a big file, e.g. `https://dl.fbaipublicfiles.com/dpr/data/retriever/biencoder-nq-train.json.gz` 2. Load it like `data = load_dataset("json", data_files=["nq-train.json"], split="train")` ```python from datasets import load_dataset data = load_dataset("json", data_files=["nq-train.json"], split="train") ``` A similarly formatted but smaller file, e.g. e.g. `https://dl.fbaipublicfiles.com/dpr/data/retriever/biencoder-nq-dev.json.gz` is loaded without issues ```python from datasets import load_dataset data = load_dataset("json", data_files=["nq-dev.json"], split="train") ``` ### Expected behavior It should load normally ### Environment info - `datasets` version: 2.16.1 - Platform: Linux-5.18.10-76051810-generic-x86_64-with-glibc2.31 - Python version: 3.10.13 - `huggingface_hub` version: 0.20.3 - PyArrow version: 15.0.0 - Pandas version: 2.2.0 - `fsspec` version: 2023.10.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6656/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6656/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6655
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6655/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6655/comments
https://api.github.com/repos/huggingface/datasets/issues/6655/events
https://github.com/huggingface/datasets/issues/6655
2,127,020,042
I_kwDODunzps5-x8AK
6,655
Cannot load the dataset go_emotions
{ "login": "arame", "id": 688324, "node_id": "MDQ6VXNlcjY4ODMyNA==", "avatar_url": "https://avatars.githubusercontent.com/u/688324?v=4", "gravatar_id": "", "url": "https://api.github.com/users/arame", "html_url": "https://github.com/arame", "followers_url": "https://api.github.com/users/arame/followers", "following_url": "https://api.github.com/users/arame/following{/other_user}", "gists_url": "https://api.github.com/users/arame/gists{/gist_id}", "starred_url": "https://api.github.com/users/arame/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/arame/subscriptions", "organizations_url": "https://api.github.com/users/arame/orgs", "repos_url": "https://api.github.com/users/arame/repos", "events_url": "https://api.github.com/users/arame/events{/privacy}", "received_events_url": "https://api.github.com/users/arame/received_events", "type": "User", "site_admin": false }
[]
open
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
4
2024-02-09T12:15:39
2024-02-12T09:35:55
null
NONE
null
null
null
### Describe the bug When I run the following code I get an exception; `go_emotions = load_dataset("go_emotions")` > AttributeError Traceback (most recent call last) Cell In[6], [line 1](vscode-notebook-cell:?execution_count=6&line=1) ----> [1](vscode-notebook-cell:?execution_count=6&line=1) go_emotions = load_dataset("go_emotions") [2](vscode-notebook-cell:?execution_count=6&line=2) data = go_emotions.data File [c:\Users\hijik\anaconda3\Lib\site-packages\datasets\load.py:2523](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2523), in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, ignore_verifications, keep_in_memory, save_infos, revision, token, use_auth_token, task, streaming, num_proc, storage_options, trust_remote_code, **config_kwargs) [2518](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2518) verification_mode = VerificationMode( [2519](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2519) (verification_mode or VerificationMode.BASIC_CHECKS) if not save_infos else VerificationMode.ALL_CHECKS [2520](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2520) ) [2522](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2522) # Create a dataset builder -> [2523](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2523) builder_instance = load_dataset_builder( [2524](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2524) path=path, [2525](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2525) name=name, [2526](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2526) data_dir=data_dir, [2527](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2527) data_files=data_files, [2528](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2528) cache_dir=cache_dir, [2529](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2529) features=features, [2530](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2530) download_config=download_config, [2531](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2531) download_mode=download_mode, [2532](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2532) revision=revision, [2533](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2533) token=token, [2534](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2534) storage_options=storage_options, [2535](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2535) trust_remote_code=trust_remote_code, [2536](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2536) _require_default_config_name=name is None, ... ---> [63](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/utils/_dill.py:63) if issubclass(obj_type, transformers.PreTrainedTokenizerBase): [64](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/utils/_dill.py:64) pklregister(obj_type)(_save_transformersPreTrainedTokenizerBase) [66](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/utils/_dill.py:66) # Unwrap `torch.compile`-ed functions AttributeError: module 'transformers' has no attribute 'PreTrainedTokenizerBase' Output is truncated. View as a [scrollable element](command:cellOutput.enableScrolling?10bc0728-6947-456e-9a3e-f056872b04c6) or open in a [text editor](command:workbench.action.openLargeOutput?10bc0728-6947-456e-9a3e-f056872b04c6). Adjust cell output [settings](command:workbench.action.openSettings?%5B%22%40tag%3AnotebookOutputLayout%22%5D)... ### Steps to reproduce the bug ``` from datasets import load_dataset go_emotions = load_dataset("go_emotions") ``` ### Expected behavior Should simply load the variable with the data from the file ### Environment info Copy-and-paste the text below in your GitHub issue. - `datasets` version: 2.16.1 - Platform: Windows-10-10.0.22631-SP0 - Python version: 3.11.4 - `huggingface_hub` version: 0.20.3 - PyArrow version: 11.0.0 - Pandas version: 1.5.3 - `fsspec` version: 2023.10.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6655/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6655/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6654
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6654/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6654/comments
https://api.github.com/repos/huggingface/datasets/issues/6654/events
https://github.com/huggingface/datasets/issues/6654
2,126,939,358
I_kwDODunzps5-xoTe
6,654
Batched dataset map throws exception that cannot cast fixed length array to Sequence
{ "login": "keesjandevries", "id": 1029671, "node_id": "MDQ6VXNlcjEwMjk2NzE=", "avatar_url": "https://avatars.githubusercontent.com/u/1029671?v=4", "gravatar_id": "", "url": "https://api.github.com/users/keesjandevries", "html_url": "https://github.com/keesjandevries", "followers_url": "https://api.github.com/users/keesjandevries/followers", "following_url": "https://api.github.com/users/keesjandevries/following{/other_user}", "gists_url": "https://api.github.com/users/keesjandevries/gists{/gist_id}", "starred_url": "https://api.github.com/users/keesjandevries/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/keesjandevries/subscriptions", "organizations_url": "https://api.github.com/users/keesjandevries/orgs", "repos_url": "https://api.github.com/users/keesjandevries/repos", "events_url": "https://api.github.com/users/keesjandevries/events{/privacy}", "received_events_url": "https://api.github.com/users/keesjandevries/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-02-09T11:23:19
2024-02-12T08:26:53
2024-02-12T08:26:53
NONE
null
null
null
### Describe the bug I encountered a TypeError when batch processing a dataset with Sequence features in datasets package version 2.16.1. The error arises from a mismatch in handling fixed-size list arrays during the map function execution. Debugging pinpoints the issue to an if-statement in datasets/table.py, line 2093, failing to correctly process sequence lengths. ### Steps to reproduce the bug Create virtual environment and activate ``` virtualenv venv source venv/bin/activate ``` Then install the datasets package (I'm using the latest version) ``` pip install datasets==2.16.1 ``` Then run ```python # bug.py from datasets import Dataset from datasets.features import Features, Sequence, Value data = { "num": [[1, 2], [3, 4]], } features = Features({'num': Sequence(feature=Value(dtype='int32'), length=2)}) dataset = Dataset.from_dict(data, features=features) dataset.map(lambda x: x, batched=True, batch_size=1) ``` ### Expected behavior I get the following stack trace ``` Map: 50%|█████ | 1/2 [00:00<00:00, 423.92 examples/s] Traceback (most recent call last): File "/PATH/TO/BUG_PORT/bug.py", line 9, in <module> dataset.map(lambda x: x, batched=True, batch_size=1) File "/PATH/TO/BUG_PORT/venv/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 592, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/PATH/TO/BUG_PORT/venv/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/PATH/TO/BUG_PORT/venv/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3093, in map for rank, done, content in Dataset._map_single(**dataset_kwargs): File "/PATH/TO/BUG_PORT/venv/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3489, in _map_single writer.write_batch(batch) File "/PATH/TO/BUG_PORT/venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 551, in write_batch array = cast_array_to_feature(col_values, col_type) if col_type is not None else col_values File "/PATH/TO/BUG_PORT/venv/lib/python3.9/site-packages/datasets/table.py", line 1797, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/PATH/TO/BUG_PORT/venv/lib/python3.9/site-packages/datasets/table.py", line 1797, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/PATH/TO/BUG_PORT/venv/lib/python3.9/site-packages/datasets/table.py", line 2111, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") TypeError: Couldn't cast array of type fixed_size_list<item: int32>[2] to Sequence(feature=Value(dtype='int32', id=None), length=2, id=None) ``` After some debugging, I found that the if-statement that is actually failing is line 2093 in `datasets/table.py` ```python # datasets/table.py ... 2093 if feature.length * len(array) == len(array_values): 2094 return pa.FixedSizeListArray.from_arrays(_c(array_values, feature.feature), feature.length) ... ``` ### Environment info Platform: MacOS Datasets version: datasets==2.16.1 Python version: 3.9.6
{ "login": "keesjandevries", "id": 1029671, "node_id": "MDQ6VXNlcjEwMjk2NzE=", "avatar_url": "https://avatars.githubusercontent.com/u/1029671?v=4", "gravatar_id": "", "url": "https://api.github.com/users/keesjandevries", "html_url": "https://github.com/keesjandevries", "followers_url": "https://api.github.com/users/keesjandevries/followers", "following_url": "https://api.github.com/users/keesjandevries/following{/other_user}", "gists_url": "https://api.github.com/users/keesjandevries/gists{/gist_id}", "starred_url": "https://api.github.com/users/keesjandevries/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/keesjandevries/subscriptions", "organizations_url": "https://api.github.com/users/keesjandevries/orgs", "repos_url": "https://api.github.com/users/keesjandevries/repos", "events_url": "https://api.github.com/users/keesjandevries/events{/privacy}", "received_events_url": "https://api.github.com/users/keesjandevries/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6654/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6654/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6651
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6651/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6651/comments
https://api.github.com/repos/huggingface/datasets/issues/6651/events
https://github.com/huggingface/datasets/issues/6651
2,126,649,626
I_kwDODunzps5-whka
6,651
Slice splits support for datasets.load_from_disk
{ "login": "mhorlacher", "id": 37439882, "node_id": "MDQ6VXNlcjM3NDM5ODgy", "avatar_url": "https://avatars.githubusercontent.com/u/37439882?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mhorlacher", "html_url": "https://github.com/mhorlacher", "followers_url": "https://api.github.com/users/mhorlacher/followers", "following_url": "https://api.github.com/users/mhorlacher/following{/other_user}", "gists_url": "https://api.github.com/users/mhorlacher/gists{/gist_id}", "starred_url": "https://api.github.com/users/mhorlacher/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mhorlacher/subscriptions", "organizations_url": "https://api.github.com/users/mhorlacher/orgs", "repos_url": "https://api.github.com/users/mhorlacher/repos", "events_url": "https://api.github.com/users/mhorlacher/events{/privacy}", "received_events_url": "https://api.github.com/users/mhorlacher/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-02-09T08:00:21
2024-06-14T14:42:46
null
NONE
null
null
null
### Feature request Support for slice splits in `datasets.load_from_disk`, similar to how it's already supported for `datasets.load_dataset`. ### Motivation Slice splits are convienient in a numer of cases - adding support to `datasets.load_from_disk` would make working with local datasets easier and homogenize the APIs of load_from_disk and load_dataset. ### Your contribution Sure, if the devs think the feature request is sensible.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6651/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6651/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6650
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6650/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6650/comments
https://api.github.com/repos/huggingface/datasets/issues/6650/events
https://github.com/huggingface/datasets/issues/6650
2,125,680,991
I_kwDODunzps5-s1Ff
6,650
AttributeError: 'InMemoryTable' object has no attribute '_batches'
{ "login": "matsuobasho", "id": 13874772, "node_id": "MDQ6VXNlcjEzODc0Nzcy", "avatar_url": "https://avatars.githubusercontent.com/u/13874772?v=4", "gravatar_id": "", "url": "https://api.github.com/users/matsuobasho", "html_url": "https://github.com/matsuobasho", "followers_url": "https://api.github.com/users/matsuobasho/followers", "following_url": "https://api.github.com/users/matsuobasho/following{/other_user}", "gists_url": "https://api.github.com/users/matsuobasho/gists{/gist_id}", "starred_url": "https://api.github.com/users/matsuobasho/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/matsuobasho/subscriptions", "organizations_url": "https://api.github.com/users/matsuobasho/orgs", "repos_url": "https://api.github.com/users/matsuobasho/repos", "events_url": "https://api.github.com/users/matsuobasho/events{/privacy}", "received_events_url": "https://api.github.com/users/matsuobasho/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
3
2024-02-08T17:11:26
2024-02-21T00:34:41
null
NONE
null
null
null
### Describe the bug ``` Traceback (most recent call last): File "finetune.py", line 103, in <module> main(args) File "finetune.py", line 45, in main data_tokenized = data.map(partial(funcs.tokenize_function, tokenizer, File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/dataset_dict.py", line 868, in map { File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/dataset_dict.py", line 869, in <dictcomp> k: dataset.map( File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 592, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 3093, in map for rank, done, content in Dataset._map_single(**dataset_kwargs): File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 3432, in _map_single arrow_formatted_shard = shard.with_format("arrow") File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 2667, in with_format dataset = copy.deepcopy(self) File "/opt/conda/envs/ptca/lib/python3.8/copy.py", line 172, in deepcopy y = _reconstruct(x, memo, *rv) File "/opt/conda/envs/ptca/lib/python3.8/copy.py", line 270, in _reconstruct state = deepcopy(state, memo) File "/opt/conda/envs/ptca/lib/python3.8/copy.py", line 146, in deepcopy y = copier(x, memo) File "/opt/conda/envs/ptca/lib/python3.8/copy.py", line 230, in _deepcopy_dict y[deepcopy(key, memo)] = deepcopy(value, memo) File "/opt/conda/envs/ptca/lib/python3.8/copy.py", line 153, in deepcopy y = copier(memo) File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/table.py", line 176, in __deepcopy__ memo[id(self._batches)] = list(self._batches) AttributeError: 'InMemoryTable' object has no attribute '_batches' ``` ### Steps to reproduce the bug I'm running an MLOps flow using AzureML. The error appears when I run the following function in my training script: ```python data_tokenized = data.map(partial(funcs.tokenize_function, tokenizer, seq_length), batched=True, batch_size=batch_size, remove_columns=['col1', 'col2']) ``` ```python def tokenize_function(tok, seq_length, example) # Pad so that each batch has the same sequence length inp = tok(example['col1'], padding=True, truncation=True) outp = tok(example['col2'], padding="max_length", max_length=seq_length) res = { 'input_ids': inp['input_ids'], 'attention_mask': inp['attention_mask'], 'decoder_input_ids': outp['input_ids'], 'labels': outp['input_ids'], 'decoder_attention_mask': outp['attention_mask'] } return res ``` ### Expected behavior Processing proceeds without errors. I ran this same workflow 2 weeks ago without a problem. I recreated the environment since then but it doesn't appear that datasets versions have changed since Dec. '23. ### Environment info datasets 2.16.1 transformers 4.35.2 pyarrow 15.0.0 pyarrow-hotfix 0.6 torch 2.0.1 I'm not using the latest transformers version because there was an error due to a conflict with Azure mlflow when I tried the last time.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6650/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6650/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6645
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6645/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6645/comments
https://api.github.com/repos/huggingface/datasets/issues/6645/events
https://github.com/huggingface/datasets/issues/6645
2,122,956,818
I_kwDODunzps5-icAS
6,645
Support fsspec 2024.2
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-02-07T12:45:29
2024-02-29T15:12:19
2024-02-29T15:12:19
MEMBER
null
null
null
Support fsspec 2024.2. First, we should address: - #6644
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6645/reactions", "total_count": 8, "+1": 8, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6645/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6644
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6644/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6644/comments
https://api.github.com/repos/huggingface/datasets/issues/6644/events
https://github.com/huggingface/datasets/issues/6644
2,122,955,282
I_kwDODunzps5-iboS
6,644
Support fsspec 2023.12
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-02-07T12:44:39
2024-02-29T15:12:18
2024-02-29T15:12:18
MEMBER
null
null
null
Support fsspec 2023.12 by handling previous and new glob behavior.
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6644/reactions", "total_count": 6, "+1": 6, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6644/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6643
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6643/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6643/comments
https://api.github.com/repos/huggingface/datasets/issues/6643/events
https://github.com/huggingface/datasets/issues/6643
2,121,239,039
I_kwDODunzps5-b4n_
6,643
Faiss GPU index cannot be serialised when passed to trainer
{ "login": "rubenweitzman", "id": 56388976, "node_id": "MDQ6VXNlcjU2Mzg4OTc2", "avatar_url": "https://avatars.githubusercontent.com/u/56388976?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rubenweitzman", "html_url": "https://github.com/rubenweitzman", "followers_url": "https://api.github.com/users/rubenweitzman/followers", "following_url": "https://api.github.com/users/rubenweitzman/following{/other_user}", "gists_url": "https://api.github.com/users/rubenweitzman/gists{/gist_id}", "starred_url": "https://api.github.com/users/rubenweitzman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rubenweitzman/subscriptions", "organizations_url": "https://api.github.com/users/rubenweitzman/orgs", "repos_url": "https://api.github.com/users/rubenweitzman/repos", "events_url": "https://api.github.com/users/rubenweitzman/events{/privacy}", "received_events_url": "https://api.github.com/users/rubenweitzman/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
3
2024-02-06T16:41:00
2024-02-15T10:29:32
null
NONE
null
null
null
### Describe the bug I am working on a retrieval project and encountering I have encountered two issues in the hugging face faiss integration: 1. I am trying to pass in a dataset with a faiss index to the Huggingface trainer. The code works for a cpu faiss index, but doesn't for a gpu one, getting error: ``` File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/transformers/trainer.py", line 1543, in train return inner_training_loop( File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/transformers/trainer.py", line 1555, in _inner_training_loop train_dataloader = self.get_train_dataloader() File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/transformers/trainer.py", line 831, in get_train_dataloader train_dataset = self._remove_unused_columns(train_dataset, description="training") File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/transformers/trainer.py", line 725, in _remove_unused_columns return dataset.remove_columns(ignored_columns) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 592, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/datasets/fingerprint.py", line 481, in wrapper out = func(dataset, *args, **kwargs) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2146, in remove_columns dataset = copy.deepcopy(self) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 172, in deepcopy y = _reconstruct(x, memo, *rv) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 271, in _reconstruct state = deepcopy(state, memo) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 146, in deepcopy y = copier(x, memo) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 231, in _deepcopy_dict y[deepcopy(key, memo)] = deepcopy(value, memo) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 146, in deepcopy y = copier(x, memo) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 231, in _deepcopy_dict y[deepcopy(key, memo)] = deepcopy(value, memo) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 172, in deepcopy y = _reconstruct(x, memo, *rv) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 271, in _reconstruct state = deepcopy(state, memo) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 146, in deepcopy y = copier(x, memo) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 231, in _deepcopy_dict y[deepcopy(key, memo)] = deepcopy(value, memo) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/copy.py", line 161, in deepcopy rv = reductor(4) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/faiss/__init__.py", line 556, in index_getstate return {"this": serialize_index(self).tobytes()} File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/faiss/__init__.py", line 1607, in serialize_index write_index(index, writer) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/faiss/swigfaiss.py", line 9843, in write_index return _swigfaiss.write_index(*args) RuntimeError: Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at /project/faiss/faiss/impl/index_write.cpp:590: don't know how to serialize this type of index ``` The index was created with the add_faiss_index method ``` train_dataset.add_faiss_index( column='embeddings', index_name='embeddings', string_factory=faiss_index_string, train_size=config.faiss_train_size, device=0, # Use -1 for CPU, or specify GPU device ID faiss_verbose=True ) ``` 2. Athough faiss is written to be compatible on the gpu for searching [https://github.com/facebookresearch/faiss/wiki/Faiss-on-the-GPU](https://github.com/facebookresearch/faiss/wiki/Faiss-on-the-GPU) I am getting error when trying to use the hugggingface code to do the search on gpu. This seems to be caused by this line https://github.com/huggingface/datasets/blob/f9975f636542df7f95c27065ea93147440d690b7/src/datasets/search.py#L376 producing error ``` total_scores, total_examples = self.dataset.get_nearest_examples_batch('embeddings', embeddings, k=self.k) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/datasets/search.py", line 773, in get_nearest_examples_batch total_scores, total_indices = self.search_batch(index_name, queries, k, **kwargs) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/datasets/search.py", line 727, in search_batch return self._indexes[index_name].search_batch(queries, k, **kwargs) File "/users/rubman/.conda/envs/protein_npt_env/lib/python3.10/site-packages/datasets/search.py", line 376, in search_batch if not queries.flags.c_contiguous: AttributeError: 'Tensor' object has no attribute 'flags' ``` ### Steps to reproduce the bug ``` train_dataset.add_faiss_index( column='embeddings', index_name='embeddings', string_factory=faiss_index_string, train_size=config.faiss_train_size, device=0, # Use -1 for CPU, or specify GPU device ID faiss_verbose=True ) Trainer( model=model, args=args, train_dataset=train_dataset, eval_dataset=eval_dataset, data_collator=data_collator, tokenizer=tokenizer ) train_dataset.get_nearest_examples_batch('embeddings', embeddings, k=self.k) ``` ### Expected behavior I would expect the faiss database code to be gpu compatible ### Environment info huggingface Version: 2.16.1
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6643/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6643/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6642
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6642/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6642/comments
https://api.github.com/repos/huggingface/datasets/issues/6642/events
https://github.com/huggingface/datasets/issues/6642
2,119,085,766
I_kwDODunzps5-Tq7G
6,642
Differently dataset object saved than it is loaded.
{ "login": "MFajcik", "id": 31218150, "node_id": "MDQ6VXNlcjMxMjE4MTUw", "avatar_url": "https://avatars.githubusercontent.com/u/31218150?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MFajcik", "html_url": "https://github.com/MFajcik", "followers_url": "https://api.github.com/users/MFajcik/followers", "following_url": "https://api.github.com/users/MFajcik/following{/other_user}", "gists_url": "https://api.github.com/users/MFajcik/gists{/gist_id}", "starred_url": "https://api.github.com/users/MFajcik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MFajcik/subscriptions", "organizations_url": "https://api.github.com/users/MFajcik/orgs", "repos_url": "https://api.github.com/users/MFajcik/repos", "events_url": "https://api.github.com/users/MFajcik/events{/privacy}", "received_events_url": "https://api.github.com/users/MFajcik/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-02-05T17:28:57
2024-02-06T09:50:19
2024-02-06T09:50:19
NONE
null
null
null
### Describe the bug Differently sized object is saved than it is loaded. ### Steps to reproduce the bug Hi, I save dataset in a following way: ``` dataset = load_dataset("json", data_files={ "train": os.path.join(input_folder, f"{task_meta_type}_{task_type}_train.jsonl"), "test": os.path.join(input_folder, f"{task_meta_type}_{task_type}_test.jsonl")}) print(os.path.join(output_folder, f"{task_meta_type}_{task_type}")) print(f"Length of train dataset: {len(dataset['train'])}") print(f"Length of test dataset: {len(dataset['test'])}") dataset.save_to_disk(os.path.join(output_folder, f"{task_meta_type}_{task_type}")) ``` this yields output ``` .data/hf_dataset/propaganda_zanr Length of train dataset: 7642 Length of test dataset: 1000 ``` Everything looks fine. Then I load the dataset ```python from datasets import load_dataset dataset_path = ".data/hf_dataset/propaganda_zanr" dataset = load_dataset(dataset_path) print(f"Length of train dataset: {len(dataset['train'])}") print(f"Length of test dataset: {len(dataset['test'])}") ``` this prints ``` Generating train split: 1 examples [00:00, 72.10 examples/s] Generating test split: 1 examples [00:00, 100.69 examples/s] Length of train dataset: 1 Length of test dataset: 1 ``` I dont' understand :( ### Expected behavior same object is loaded ### Environment info datasets==2.16.1
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6642/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6642/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6641
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6641/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6641/comments
https://api.github.com/repos/huggingface/datasets/issues/6641/events
https://github.com/huggingface/datasets/issues/6641
2,116,963,132
I_kwDODunzps5-Lks8
6,641
unicodedecodeerror: 'utf-8' codec can't decode byte 0xac in position 25: invalid start byte
{ "login": "Hughhuh", "id": 109789057, "node_id": "U_kgDOBos_gQ", "avatar_url": "https://avatars.githubusercontent.com/u/109789057?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Hughhuh", "html_url": "https://github.com/Hughhuh", "followers_url": "https://api.github.com/users/Hughhuh/followers", "following_url": "https://api.github.com/users/Hughhuh/following{/other_user}", "gists_url": "https://api.github.com/users/Hughhuh/gists{/gist_id}", "starred_url": "https://api.github.com/users/Hughhuh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hughhuh/subscriptions", "organizations_url": "https://api.github.com/users/Hughhuh/orgs", "repos_url": "https://api.github.com/users/Hughhuh/repos", "events_url": "https://api.github.com/users/Hughhuh/events{/privacy}", "received_events_url": "https://api.github.com/users/Hughhuh/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
1
2024-02-04T08:49:31
2024-02-06T09:26:07
2024-02-06T09:11:45
NONE
null
null
null
### Describe the bug unicodedecodeerror: 'utf-8' codec can't decode byte 0xac in position 25: invalid start byte ### Steps to reproduce the bug ``` import sys sys.getdefaultencoding() 'utf-8' from datasets import load_dataset print(f"Train dataset size: {len(dataset['train'])}") print(f"Test dataset size: {len(dataset['test'])}") Resolving data files: 100% 159/159 [00:00<00:00, 9909.28it/s] Using custom data configuration samsum-0b1209637541c9e6 Downloading and preparing dataset json/samsum to C:/Users/Administrator/.cache/huggingface/datasets/json/samsum-0b1209637541c9e6/0.0.0/0f7e3662623656454fcd2b650f34e886a7db4b9104504885bd462096cc7a9f51... Downloading data files: 100% 3/3 [00:00<00:00, 119.99it/s] Extracting data files: 100% 3/3 [00:00<00:00, 9.54it/s] Generating train split: 88392/0 [00:15<00:00, 86848.17 examples/s] Generating test split: 0/0 [00:00<?, ? examples/s] --------------------------------------------------------------------------- ArrowInvalid Traceback (most recent call last) File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\datasets\packaged_modules\json\json.py:132, in Json._generate_tables(self, files) 131 try: --> 132 pa_table = paj.read_json( 133 io.BytesIO(batch), read_options=paj.ReadOptions(block_size=block_size) 134 ) 135 break File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\pyarrow\_json.pyx:290, in pyarrow._json.read_json() File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\pyarrow\error.pxi:144, in pyarrow.lib.pyarrow_internal_check_status() File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\pyarrow\error.pxi:100, in pyarrow.lib.check_status() ArrowInvalid: JSON parse error: Invalid value. in row 0 During handling of the above exception, another exception occurred: UnicodeDecodeError Traceback (most recent call last) File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\datasets\builder.py:1819, in ArrowBasedBuilder._prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id) 1818 _time = time.time() -> 1819 for _, table in generator: 1820 if max_shard_size is not None and writer._num_bytes > max_shard_size: File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\datasets\packaged_modules\json\json.py:153, in Json._generate_tables(self, files) 152 with open(file, encoding="utf-8") as f: --> 153 dataset = json.load(f) 154 except json.JSONDecodeError: File ~\AppData\Local\Programs\Python\Python310\lib\json\__init__.py:293, in load(fp, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw) 276 """Deserialize ``fp`` (a ``.read()``-supporting file-like object containing 277 a JSON document) to a Python object. 278 (...) 291 kwarg; otherwise ``JSONDecoder`` is used. 292 """ --> 293 return loads(fp.read(), 294 cls=cls, object_hook=object_hook, 295 parse_float=parse_float, parse_int=parse_int, 296 parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw) File ~\AppData\Local\Programs\Python\Python310\lib\codecs.py:322, in BufferedIncrementalDecoder.decode(self, input, final) 321 data = self.buffer + input --> 322 (result, consumed) = self._buffer_decode(data, self.errors, final) 323 # keep undecoded input until the next call UnicodeDecodeError: 'utf-8' codec can't decode byte 0xac in position 25: invalid start byte The above exception was the direct cause of the following exception: DatasetGenerationError Traceback (most recent call last) Cell In[81], line 5 1 from datasets import load_dataset 3 # Load dataset from the hub 4 #dataset = load_dataset("json",data_files="C:/Users/Administrator/Desktop/samsum/samsum/data/corpus/train.json",field="data") ----> 5 dataset = load_dataset('json',"samsum") 6 #dataset = load_dataset("samsum") 7 print(f"Train dataset size: {len(dataset['train'])}") File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\datasets\load.py:1758, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, num_proc, **config_kwargs) 1755 try_from_hf_gcs = path not in _PACKAGED_DATASETS_MODULES 1757 # Download and prepare data -> 1758 builder_instance.download_and_prepare( 1759 download_config=download_config, 1760 download_mode=download_mode, 1761 ignore_verifications=ignore_verifications, 1762 try_from_hf_gcs=try_from_hf_gcs, 1763 num_proc=num_proc, 1764 ) 1766 # Build dataset for splits 1767 keep_in_memory = ( 1768 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size) 1769 ) File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\datasets\builder.py:860, in DatasetBuilder.download_and_prepare(self, output_dir, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, file_format, max_shard_size, num_proc, storage_options, **download_and_prepare_kwargs) 858 if num_proc is not None: 859 prepare_split_kwargs["num_proc"] = num_proc --> 860 self._download_and_prepare( 861 dl_manager=dl_manager, 862 verify_infos=verify_infos, 863 **prepare_split_kwargs, 864 **download_and_prepare_kwargs, 865 ) 866 # Sync info 867 self.info.dataset_size = sum(split.num_bytes for split in self.info.splits.values()) File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\datasets\builder.py:953, in DatasetBuilder._download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 949 split_dict.add(split_generator.split_info) 951 try: 952 # Prepare split will record examples associated to the split --> 953 self._prepare_split(split_generator, **prepare_split_kwargs) 954 except OSError as e: 955 raise OSError( 956 "Cannot find data file. " 957 + (self.manual_download_instructions or "") 958 + "\nOriginal error:\n" 959 + str(e) 960 ) from None File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\datasets\builder.py:1708, in ArrowBasedBuilder._prepare_split(self, split_generator, file_format, num_proc, max_shard_size) 1706 gen_kwargs = split_generator.gen_kwargs 1707 job_id = 0 -> 1708 for job_id, done, content in self._prepare_split_single( 1709 gen_kwargs=gen_kwargs, job_id=job_id, **_prepare_split_args 1710 ): 1711 if done: 1712 result = content File ~\AppData\Local\Programs\Python\Python310\lib\site-packages\datasets\builder.py:1851, in ArrowBasedBuilder._prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id) 1849 if isinstance(e, SchemaInferenceError) and e.__context__ is not None: 1850 e = e.__context__ -> 1851 raise DatasetGenerationError("An error occurred while generating the dataset") from e 1853 yield job_id, True, (total_num_examples, total_num_bytes, writer._features, num_shards, shard_lengths) DatasetGenerationError: An error occurred while generating the dataset ``` ### Expected behavior can't load dataset ### Environment info dataset:samsum system :win10 gpu:m40 24G
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6641/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6641/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6640
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6640/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6640/comments
https://api.github.com/repos/huggingface/datasets/issues/6640/events
https://github.com/huggingface/datasets/issues/6640
2,115,864,531
I_kwDODunzps5-HYfT
6,640
Sign Language Support
{ "login": "Merterm", "id": 6684795, "node_id": "MDQ6VXNlcjY2ODQ3OTU=", "avatar_url": "https://avatars.githubusercontent.com/u/6684795?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Merterm", "html_url": "https://github.com/Merterm", "followers_url": "https://api.github.com/users/Merterm/followers", "following_url": "https://api.github.com/users/Merterm/following{/other_user}", "gists_url": "https://api.github.com/users/Merterm/gists{/gist_id}", "starred_url": "https://api.github.com/users/Merterm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Merterm/subscriptions", "organizations_url": "https://api.github.com/users/Merterm/orgs", "repos_url": "https://api.github.com/users/Merterm/repos", "events_url": "https://api.github.com/users/Merterm/events{/privacy}", "received_events_url": "https://api.github.com/users/Merterm/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-02-02T21:54:51
2024-02-02T21:54:51
null
NONE
null
null
null
### Feature request Currently, there are only several Sign Language labels, I would like to propose adding all the Signed Languages as new labels which are described in this ISO standard: https://www.evertype.com/standards/iso639/sign-language.html ### Motivation Datasets currently only have labels for several signed languages. There are more signed languages in the world. Furthermore, some signed languages that have a lot of online data cannot be found because of this reason (for instance, German Sign Language, and there is no German Sign Language label on huggingface datasets even though there are a lot of readily available sign language datasets exist for German Sign Language, which are used very frequently in Sign Language Processing papers, and models.) ### Your contribution I can submit a PR for this as well, adding the ISO codes and languages to the labels in datasets.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6640/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6640/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6638
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6638/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6638/comments
https://api.github.com/repos/huggingface/datasets/issues/6638/events
https://github.com/huggingface/datasets/issues/6638
2,113,329,257
I_kwDODunzps599thp
6,638
Cannot download wmt16 dataset
{ "login": "vidyasiv", "id": 81709031, "node_id": "MDQ6VXNlcjgxNzA5MDMx", "avatar_url": "https://avatars.githubusercontent.com/u/81709031?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vidyasiv", "html_url": "https://github.com/vidyasiv", "followers_url": "https://api.github.com/users/vidyasiv/followers", "following_url": "https://api.github.com/users/vidyasiv/following{/other_user}", "gists_url": "https://api.github.com/users/vidyasiv/gists{/gist_id}", "starred_url": "https://api.github.com/users/vidyasiv/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vidyasiv/subscriptions", "organizations_url": "https://api.github.com/users/vidyasiv/orgs", "repos_url": "https://api.github.com/users/vidyasiv/repos", "events_url": "https://api.github.com/users/vidyasiv/events{/privacy}", "received_events_url": "https://api.github.com/users/vidyasiv/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-02-01T19:41:42
2024-02-01T20:07:29
2024-02-01T20:07:29
NONE
null
null
null
### Describe the bug As of this morning (PST) 2/1/2024, seeing the wmt16 dataset is missing from opus , could you suggest an alternative? ``` Downloading data files: 0%| | 0/4 [00:00<?, ?it/s]Traceback (most recent call last): File "test.py", line 2, in <module> raw_datasets = load_dataset("wmt16","ro-en",split="train") File "/usr/local/lib/python3.8/dist-packages/datasets/load.py", line 2153, in load_dataset builder_instance.download_and_prepare( File "/usr/local/lib/python3.8/dist-packages/datasets/builder.py", line 954, in download_and_prepare self._download_and_prepare( File "/usr/local/lib/python3.8/dist-packages/datasets/builder.py", line 1717, in _download_and_prepare super()._download_and_prepare( File "/usr/local/lib/python3.8/dist-packages/datasets/builder.py", line 1027, in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) File "/root/.cache/huggingface/modules/datasets_modules/datasets/wmt16/746749a11d25c02058042da7502d973ff410e73457f3d305fc1177dc0e8c4227/wmt_utils.py", line 754, in _split_generators downloaded_files = dl_manager.download_and_extract(urls_to_download) File "/usr/local/lib/python3.8/dist-packages/datasets/download/download_manager.py", line 565, in download_and_extract return self.extract(self.download(url_or_urls)) File "/usr/local/lib/python3.8/dist-packages/datasets/download/download_manager.py", line 428, in download downloaded_path_or_paths = map_nested( File "/usr/local/lib/python3.8/dist-packages/datasets/utils/py_utils.py", line 464, in map_nested mapped = [ File "/usr/local/lib/python3.8/dist-packages/datasets/utils/py_utils.py", line 465, in <listcomp> _single_map_nested((function, obj, types, None, True, None)) File "/usr/local/lib/python3.8/dist-packages/datasets/utils/py_utils.py", line 384, in _single_map_nested mapped = [_single_map_nested((function, v, types, None, True, None)) for v in pbar] File "/usr/local/lib/python3.8/dist-packages/datasets/utils/py_utils.py", line 384, in <listcomp> mapped = [_single_map_nested((function, v, types, None, True, None)) for v in pbar] File "/usr/local/lib/python3.8/dist-packages/datasets/utils/py_utils.py", line 367, in _single_map_nested return function(data_struct) File "/usr/local/lib/python3.8/dist-packages/datasets/download/download_manager.py", line 454, in _download return cached_path(url_or_filename, download_config=download_config) File "/usr/local/lib/python3.8/dist-packages/datasets/utils/file_utils.py", line 182, in cached_path output_path = get_from_cache( File "/usr/local/lib/python3.8/dist-packages/datasets/utils/file_utils.py", line 596, in get_from_cache raise FileNotFoundError(f"Couldn't find file at {url}") FileNotFoundError: Couldn't find file at https://opus.nlpl.eu/download.php?f=SETIMES/v2/tmx/en-ro.tmx.gz ``` ### Steps to reproduce the bug ``` from datasets import load_dataset raw_datasets = load_dataset("wmt16","ro-en",split="train") ``` ### Expected behavior Expect the dataset to be downloaded/ at least a clean exit with error explaining dataset is missing and a suggestion for next steps ### Environment info - `datasets` version: 2.14.7 - Platform: Linux-5.15.0-92-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - Huggingface_hub version: 0.17.3 - PyArrow version: 15.0.0 - Pandas version: 2.0.1
{ "login": "vidyasiv", "id": 81709031, "node_id": "MDQ6VXNlcjgxNzA5MDMx", "avatar_url": "https://avatars.githubusercontent.com/u/81709031?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vidyasiv", "html_url": "https://github.com/vidyasiv", "followers_url": "https://api.github.com/users/vidyasiv/followers", "following_url": "https://api.github.com/users/vidyasiv/following{/other_user}", "gists_url": "https://api.github.com/users/vidyasiv/gists{/gist_id}", "starred_url": "https://api.github.com/users/vidyasiv/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vidyasiv/subscriptions", "organizations_url": "https://api.github.com/users/vidyasiv/orgs", "repos_url": "https://api.github.com/users/vidyasiv/repos", "events_url": "https://api.github.com/users/vidyasiv/events{/privacy}", "received_events_url": "https://api.github.com/users/vidyasiv/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6638/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6638/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6637
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6637/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6637/comments
https://api.github.com/repos/huggingface/datasets/issues/6637/events
https://github.com/huggingface/datasets/issues/6637
2,113,025,975
I_kwDODunzps598je3
6,637
'with_format' is extremely slow when used together with 'interleave_datasets' or 'shuffle' on IterableDatasets
{ "login": "tobycrisford", "id": 22883190, "node_id": "MDQ6VXNlcjIyODgzMTkw", "avatar_url": "https://avatars.githubusercontent.com/u/22883190?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tobycrisford", "html_url": "https://github.com/tobycrisford", "followers_url": "https://api.github.com/users/tobycrisford/followers", "following_url": "https://api.github.com/users/tobycrisford/following{/other_user}", "gists_url": "https://api.github.com/users/tobycrisford/gists{/gist_id}", "starred_url": "https://api.github.com/users/tobycrisford/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tobycrisford/subscriptions", "organizations_url": "https://api.github.com/users/tobycrisford/orgs", "repos_url": "https://api.github.com/users/tobycrisford/repos", "events_url": "https://api.github.com/users/tobycrisford/events{/privacy}", "received_events_url": "https://api.github.com/users/tobycrisford/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
1
2024-02-01T17:16:54
2024-02-05T10:43:47
null
NONE
null
null
null
### Describe the bug If you: 1. Interleave two iterable datasets together with the interleave_datasets function, or shuffle an iterable dataset 2. Set the output format to torch tensors with .with_format('torch') Then iterating through the dataset becomes over 100x slower than it is if you don't apply the torch formatting. ### Steps to reproduce the bug ```python import datasets import torch from tqdm import tqdm rand_a = torch.randn(3,224,224) rand_b = torch.randn(3,224,224) a = torch.stack([rand_a] * 1000) b = torch.stack([rand_b] * 1000) features = datasets.Features({"tensor": datasets.Array3D(shape=(3,224,224), dtype="float32")}) ds_a = datasets.Dataset.from_dict({"tensor": a}, features=features).to_iterable_dataset() ds_b = datasets.Dataset.from_dict({"tensor": b}, features=features).to_iterable_dataset() # Iterating through either dataset with torch formatting is really fast (2000it/s on my machine) for example in tqdm(ds_a.with_format('torch')): pass # Iterating through either dataset shuffled is also pretty fast (100it/s on my machine) for example in tqdm(ds_a.shuffle()): pass # Iterating through this interleaved dataset is pretty fast (200it/s on my machine) ds_fast = datasets.interleave_datasets([ds_a, ds_b]) for example in tqdm(ds_fast): pass # Iterating through either dataset with torch formatting *after shuffling* is really slow... (<2it/s on my machine) for example in tqdm(ds_a.shuffle().with_format('torch')): pass # Iterating through this torch formatted interleaved dataset is also really slow (<2it/s on my machine)... ds_slow = datasets.interleave_datasets([ds_a, ds_b]).with_format('torch') for example in tqdm(ds_slow): pass # Even doing this is way faster!! (70it/s on my machine) for example in tqdm(ds_fast): test = torch.tensor(example['tensor']) ``` ### Expected behavior Applying torch formatting to the interleaved dataset shouldn't increase the time taken to iterate through the dataset by very much, since even explicitly converting every example is over 70x faster than calling .with_format('torch'). ### Environment info - `datasets` version: 2.16.1 - Platform: Linux-6.5.0-15-generic-x86_64-with-glibc2.38 - Python version: 3.11.6 - `huggingface_hub` version: 0.20.3 - PyArrow version: 15.0.0 - Pandas version: 2.2.0 - `fsspec` version: 2023.10.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6637/reactions", "total_count": 4, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 3 }
https://api.github.com/repos/huggingface/datasets/issues/6637/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6624
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6624/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6624/comments
https://api.github.com/repos/huggingface/datasets/issues/6624/events
https://github.com/huggingface/datasets/issues/6624
2,103,950,718
I_kwDODunzps59Z71-
6,624
How to download the laion-coco dataset
{ "login": "vanpersie32", "id": 15981416, "node_id": "MDQ6VXNlcjE1OTgxNDE2", "avatar_url": "https://avatars.githubusercontent.com/u/15981416?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vanpersie32", "html_url": "https://github.com/vanpersie32", "followers_url": "https://api.github.com/users/vanpersie32/followers", "following_url": "https://api.github.com/users/vanpersie32/following{/other_user}", "gists_url": "https://api.github.com/users/vanpersie32/gists{/gist_id}", "starred_url": "https://api.github.com/users/vanpersie32/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vanpersie32/subscriptions", "organizations_url": "https://api.github.com/users/vanpersie32/orgs", "repos_url": "https://api.github.com/users/vanpersie32/repos", "events_url": "https://api.github.com/users/vanpersie32/events{/privacy}", "received_events_url": "https://api.github.com/users/vanpersie32/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-28T03:56:05
2024-02-06T09:43:31
2024-02-06T09:43:31
NONE
null
null
null
The laion coco dataset is not available now. How to download it https://huggingface.co/datasets/laion/laion-coco
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6624/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6624/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6623
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6623/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6623/comments
https://api.github.com/repos/huggingface/datasets/issues/6623/events
https://github.com/huggingface/datasets/issues/6623
2,103,870,123
I_kwDODunzps59ZoKr
6,623
streaming datasets doesn't work properly with multi-node
{ "login": "rohitgr7", "id": 30778939, "node_id": "MDQ6VXNlcjMwNzc4OTM5", "avatar_url": "https://avatars.githubusercontent.com/u/30778939?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rohitgr7", "html_url": "https://github.com/rohitgr7", "followers_url": "https://api.github.com/users/rohitgr7/followers", "following_url": "https://api.github.com/users/rohitgr7/following{/other_user}", "gists_url": "https://api.github.com/users/rohitgr7/gists{/gist_id}", "starred_url": "https://api.github.com/users/rohitgr7/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rohitgr7/subscriptions", "organizations_url": "https://api.github.com/users/rohitgr7/orgs", "repos_url": "https://api.github.com/users/rohitgr7/repos", "events_url": "https://api.github.com/users/rohitgr7/events{/privacy}", "received_events_url": "https://api.github.com/users/rohitgr7/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
8
2024-01-27T23:46:13
2024-03-08T14:27:08
null
NONE
null
null
null
### Feature request Let’s say I have a dataset with 5 samples with values [1, 2, 3, 4, 5], with 2 GPUs (for DDP) and batch size of 2. This dataset is an `IterableDataset` since I am streaming it. Now I split the dataset using `split_dataset_by_node` to ensure it doesn’t get repeated. And since it’s already splitted, I don’t have to use `DistributedSampler` (also they don't work with iterable datasets anyway)? But in this case I noticed that the: First iteraton: first GPU will get → [1, 2] first GPU will get → [3, 4] Second iteraton: first GPU will get → [5] first GPU will get → Nothing which actually creates an issue since in case of `DistributedSampler`, the samples are repeated internally to ensure non of the GPUs at any iteration is missing any data for gradient sync. So my questions are: 1. Here since splitting is happening before hand, how to make sure each GPU get’s a batch at each iteration to avoid gradient sync issues? 2. Do we need to use `DistributedSampler`? If yes, how? 3. in the docstrings of `split_dataset_by_node`, this is mentioned: *"If the dataset has a number of shards that is a factor of `world_size` (i.e. if `dataset.n_shards % world_size == 0`), then the shards are evenly assigned across the nodes, which is the most optimized. Otherwise, each node keeps 1 example out of `world_size`, skipping the other examples."* Can you explain the last part here? 4. If `dataset.n_shards % world_size != 0`, is it possible to shard the streaming dataset on the fly to avoid the case where data is missing? ### Motivation Somehow streaming datasets should work with DDP since for big LLMs a lot of data is required and DDP/multi-node is mostly used to train such models and streaming can actually help solve the data part of it. ### Your contribution Yes, I can help in submitting the PR once we get mutual understanding on how it should behave.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6623/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6623/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6622
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6622/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6622/comments
https://api.github.com/repos/huggingface/datasets/issues/6622/events
https://github.com/huggingface/datasets/issues/6622
2,103,780,697
I_kwDODunzps59ZSVZ
6,622
multi-GPU map does not work
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-27T20:06:08
2024-02-08T11:18:21
2024-02-08T11:18:21
NONE
null
null
null
### Describe the bug Here is the code for single-GPU processing: https://pastebin.com/bfmEeK2y Here is the code for multi-GPU processing: https://pastebin.com/gQ7i5AQy Here is the video showing that the multi-GPU mapping does not work as expected (there are so many things wrong here, it's better to watch the 3-minute video than explain here): https://youtu.be/RNbdPkSppc4 ### Steps to reproduce the bug - ### Expected behavior - ### Environment info x2 RTX A4000
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6622/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6622/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6621
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6621/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6621/comments
https://api.github.com/repos/huggingface/datasets/issues/6621/events
https://github.com/huggingface/datasets/issues/6621
2,103,675,294
I_kwDODunzps59Y4me
6,621
deleted
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-27T16:59:58
2024-01-27T17:14:43
2024-01-27T17:14:43
NONE
null
null
null
...
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6621/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6621/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6620
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6620/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6620/comments
https://api.github.com/repos/huggingface/datasets/issues/6620/events
https://github.com/huggingface/datasets/issues/6620
2,103,110,536
I_kwDODunzps59WuuI
6,620
wiki_dpr.py error (ID mismatch between lines {id} and vector {vec_id}
{ "login": "kiehls90", "id": 101498700, "node_id": "U_kgDOBgy_TA", "avatar_url": "https://avatars.githubusercontent.com/u/101498700?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kiehls90", "html_url": "https://github.com/kiehls90", "followers_url": "https://api.github.com/users/kiehls90/followers", "following_url": "https://api.github.com/users/kiehls90/following{/other_user}", "gists_url": "https://api.github.com/users/kiehls90/gists{/gist_id}", "starred_url": "https://api.github.com/users/kiehls90/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kiehls90/subscriptions", "organizations_url": "https://api.github.com/users/kiehls90/orgs", "repos_url": "https://api.github.com/users/kiehls90/repos", "events_url": "https://api.github.com/users/kiehls90/events{/privacy}", "received_events_url": "https://api.github.com/users/kiehls90/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
1
2024-01-27T01:00:09
2024-02-06T09:40:19
2024-02-06T09:40:19
NONE
null
null
null
### Describe the bug I'm trying to run a rag example, and the dataset is wiki_dpr. wiki_dpr download and extracting have been completed successfully. However, at the generating train split stage, an error from wiki_dpr.py keeps popping up. Especially in "_generate_examples" : 1. The following error occurs in the line **id, text, title = line.strip().split("\t")** ValueError: not enough values ​​to unpack (expected 3, got 2) -> This part handles exceptions so that even if an error occurs, it passes. 2. **ID mismatch between lines {id} and vector {vec_id}** This error seems to occur at the line " assert int(id) == int(vec_id),". After I handled the exception in the split error, generating train split progressed to 80%, but an id mismatch error occurred at about the 16200000th vector id. Debugging is even more difficult because it takes a long time to download and split wiki_dpr. I need help. thank you in advance!! ### Steps to reproduce the bug Occurs in the generating train split step when running the rag example in the transformers repository. Specifically, it is an error in wiki_dpr.py. ### Expected behavior . ### Environment info python 3.8
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6620/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6620/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6618
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6618/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6618/comments
https://api.github.com/repos/huggingface/datasets/issues/6618/events
https://github.com/huggingface/datasets/issues/6618
2,101,868,198
I_kwDODunzps59R_am
6,618
While importing load_dataset from datasets
{ "login": "Era-cell", "id": 77973415, "node_id": "MDQ6VXNlcjc3OTczNDE1", "avatar_url": "https://avatars.githubusercontent.com/u/77973415?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Era-cell", "html_url": "https://github.com/Era-cell", "followers_url": "https://api.github.com/users/Era-cell/followers", "following_url": "https://api.github.com/users/Era-cell/following{/other_user}", "gists_url": "https://api.github.com/users/Era-cell/gists{/gist_id}", "starred_url": "https://api.github.com/users/Era-cell/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Era-cell/subscriptions", "organizations_url": "https://api.github.com/users/Era-cell/orgs", "repos_url": "https://api.github.com/users/Era-cell/repos", "events_url": "https://api.github.com/users/Era-cell/events{/privacy}", "received_events_url": "https://api.github.com/users/Era-cell/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
5
2024-01-26T09:21:57
2024-07-23T09:31:07
2024-02-06T09:25:54
NONE
null
null
null
### Describe the bug cannot import name 'DEFAULT_CIPHERS' from 'urllib3.util.ssl_' this is the error i received ### Steps to reproduce the bug from datasets import load_dataset ### Expected behavior No errors ### Environment info python 3.11.5
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6618/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6618/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6615
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6615/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6615/comments
https://api.github.com/repos/huggingface/datasets/issues/6615/events
https://github.com/huggingface/datasets/issues/6615
2,098,951,409
I_kwDODunzps59G3Tx
6,615
...
{ "login": "ftkeys", "id": 22179777, "node_id": "MDQ6VXNlcjIyMTc5Nzc3", "avatar_url": "https://avatars.githubusercontent.com/u/22179777?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ftkeys", "html_url": "https://github.com/ftkeys", "followers_url": "https://api.github.com/users/ftkeys/followers", "following_url": "https://api.github.com/users/ftkeys/following{/other_user}", "gists_url": "https://api.github.com/users/ftkeys/gists{/gist_id}", "starred_url": "https://api.github.com/users/ftkeys/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ftkeys/subscriptions", "organizations_url": "https://api.github.com/users/ftkeys/orgs", "repos_url": "https://api.github.com/users/ftkeys/repos", "events_url": "https://api.github.com/users/ftkeys/events{/privacy}", "received_events_url": "https://api.github.com/users/ftkeys/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-24T19:37:03
2024-01-24T19:42:30
2024-01-24T19:40:11
NONE
null
null
null
...
{ "login": "ftkeys", "id": 22179777, "node_id": "MDQ6VXNlcjIyMTc5Nzc3", "avatar_url": "https://avatars.githubusercontent.com/u/22179777?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ftkeys", "html_url": "https://github.com/ftkeys", "followers_url": "https://api.github.com/users/ftkeys/followers", "following_url": "https://api.github.com/users/ftkeys/following{/other_user}", "gists_url": "https://api.github.com/users/ftkeys/gists{/gist_id}", "starred_url": "https://api.github.com/users/ftkeys/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ftkeys/subscriptions", "organizations_url": "https://api.github.com/users/ftkeys/orgs", "repos_url": "https://api.github.com/users/ftkeys/repos", "events_url": "https://api.github.com/users/ftkeys/events{/privacy}", "received_events_url": "https://api.github.com/users/ftkeys/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6615/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6615/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6614
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6614/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6614/comments
https://api.github.com/repos/huggingface/datasets/issues/6614/events
https://github.com/huggingface/datasets/issues/6614
2,098,884,520
I_kwDODunzps59Gm-o
6,614
`datasets/downloads` cleanup tool
{ "login": "stas00", "id": 10676103, "node_id": "MDQ6VXNlcjEwNjc2MTAz", "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stas00", "html_url": "https://github.com/stas00", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "organizations_url": "https://api.github.com/users/stas00/orgs", "repos_url": "https://api.github.com/users/stas00/repos", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "received_events_url": "https://api.github.com/users/stas00/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-01-24T18:52:10
2024-01-24T18:55:09
null
CONTRIBUTOR
null
null
null
### Feature request Splitting off https://github.com/huggingface/huggingface_hub/issues/1997 - currently `huggingface-cli delete-cache` doesn't take care of cleaning `datasets` temp files e.g. I discovered having millions of files under `datasets/downloads` cache, I had to do: ``` sudo find /data/huggingface/datasets/downloads -type f -mtime +3 -exec rm {} \+ sudo find /data/huggingface/datasets/downloads -type d -empty -delete ``` could the cleanup be integrated into `huggingface-cli` or a different tool provided to keep the folders tidy and not consume inodes and space e.g. there were tens of thousands of `.lock` files - I don't know why they never get removed - lock files should be temporary for the duration of the operation requiring the lock and not remain after the operation finished, IMHO. Also I think one should be able to nuke `datasets/downloads` w/o hurting the cache, but I think there are some datasets that rely on files extracted under this dir - or at least they did in the past - which is very difficult to manage since one has no idea what is safe to delete and what not. Thank you @Wauplin (requested to be tagged)
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6614/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6614/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6612
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6612/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6612/comments
https://api.github.com/repos/huggingface/datasets/issues/6612/events
https://github.com/huggingface/datasets/issues/6612
2,098,078,210
I_kwDODunzps59DiIC
6,612
cnn_dailymail repeats itself
{ "login": "KeremZaman", "id": 8274752, "node_id": "MDQ6VXNlcjgyNzQ3NTI=", "avatar_url": "https://avatars.githubusercontent.com/u/8274752?v=4", "gravatar_id": "", "url": "https://api.github.com/users/KeremZaman", "html_url": "https://github.com/KeremZaman", "followers_url": "https://api.github.com/users/KeremZaman/followers", "following_url": "https://api.github.com/users/KeremZaman/following{/other_user}", "gists_url": "https://api.github.com/users/KeremZaman/gists{/gist_id}", "starred_url": "https://api.github.com/users/KeremZaman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/KeremZaman/subscriptions", "organizations_url": "https://api.github.com/users/KeremZaman/orgs", "repos_url": "https://api.github.com/users/KeremZaman/repos", "events_url": "https://api.github.com/users/KeremZaman/events{/privacy}", "received_events_url": "https://api.github.com/users/KeremZaman/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-24T11:38:25
2024-02-01T08:14:50
2024-02-01T08:14:50
NONE
null
null
null
### Describe the bug When I try to load `cnn_dailymail` dataset, it takes longer than usual and when I checked the dataset it's 3x bigger than it's supposed to be. Check https://huggingface.co/datasets/cnn_dailymail: it says 287k rows for train. But when I check length of train split it says 861339. Also I checked data: ``` >>> ds['train']['highlights'][0] "Harry Potter star Daniel Radcliffe gets £20M fortune as he turns 18 Monday . Young actor says he has no plans to fritter his cash away . Radcliffe's earnings from first five Potter films have been held in trust fund ."```` >>> ds['train']['highlights'][0] "Harry Potter star Daniel Radcliffe gets £20M fortune as he turns 18 Monday . Young actor says he has no plans to fritter his cash away . Radcliffe's earnings from first five Potter films have been held in trust fund ."```` >>> ds['train']['highlights'][287113] "Harry Potter star Daniel Radcliffe gets £20M fortune as he turns 18 Monday .\nYoung actor says he has no plans to fritter his cash away .\nRadcliffe's earnings from first five Potter films have been held in trust fund ."```` >>> ds['train']['highlights'][574226] "Harry Potter star Daniel Radcliffe gets £20M fortune as he turns 18 Monday .\nYoung actor says he has no plans to fritter his cash away .\nRadcliffe's earnings from first five Potter films have been held in trust fund ." ``` The datasets seems to be updated 6 days ago to convert it to Parquet. Probably, there is some issue with backward compatability. ### Steps to reproduce the bug 1. ``` from datasets import load_dataset ds = load_dataset('cnn_dailymail', '3.0.0') len(ds['train']) ``` ### Expected behavior It should not repeat itself. ### Environment info datasets==2.13.2 Python==3.7.13
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6612/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6612/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6611
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6611/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6611/comments
https://api.github.com/repos/huggingface/datasets/issues/6611/events
https://github.com/huggingface/datasets/issues/6611
2,096,004,858
I_kwDODunzps587n76
6,611
`load_from_disk` with large dataset from S3 runs into `botocore.exceptions.ClientError`
{ "login": "zotroneneis", "id": 15320635, "node_id": "MDQ6VXNlcjE1MzIwNjM1", "avatar_url": "https://avatars.githubusercontent.com/u/15320635?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zotroneneis", "html_url": "https://github.com/zotroneneis", "followers_url": "https://api.github.com/users/zotroneneis/followers", "following_url": "https://api.github.com/users/zotroneneis/following{/other_user}", "gists_url": "https://api.github.com/users/zotroneneis/gists{/gist_id}", "starred_url": "https://api.github.com/users/zotroneneis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zotroneneis/subscriptions", "organizations_url": "https://api.github.com/users/zotroneneis/orgs", "repos_url": "https://api.github.com/users/zotroneneis/repos", "events_url": "https://api.github.com/users/zotroneneis/events{/privacy}", "received_events_url": "https://api.github.com/users/zotroneneis/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2024-01-23T12:37:57
2024-01-23T12:37:57
null
NONE
null
null
null
### Describe the bug When loading a large dataset (>1000GB) from S3 I run into the following error: ``` Traceback (most recent call last): File "/home/alp/.local/lib/python3.10/site-packages/s3fs/core.py", line 113, in _error_wrapper return await func(*args, **kwargs) File "/home/alp/.local/lib/python3.10/site-packages/aiobotocore/client.py", line 383, in _make_api_call raise error_class(parsed_response, operation_name) botocore.exceptions.ClientError: An error occurred (RequestTimeTooSkewed) when calling the GetObject operation: The difference between the request time and the current time is too large. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/alp/phoneme-classification.monorepo/aws_sagemaker/data_processing/inspect_final_dataset.py", line 13, in <module> dataset = load_from_disk("s3://speech-recognition-processed-data/whisper/de/train_data/", storage_options=storage_options) File "/home/alp/.local/lib/python3.10/site-packages/datasets/load.py", line 1902, in load_from_disk return Dataset.load_from_disk(dataset_path, keep_in_memory=keep_in_memory, storage_options=storage_options) File "/home/alp/.local/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 1686, in load_from_disk fs.download(src_dataset_path, [dest_dataset_path.as](http://dest_dataset_path.as/)_posix(), recursive=True) File "/home/alp/.local/lib/python3.10/site-packages/fsspec/spec.py", line 1480, in download return self.get(rpath, lpath, recursive=recursive, **kwargs) File "/home/alp/.local/lib/python3.10/site-packages/fsspec/asyn.py", line 121, in wrapper return sync(self.loop, func, *args, **kwargs) File "/home/alp/.local/lib/python3.10/site-packages/fsspec/asyn.py", line 106, in sync raise return_result File "/home/alp/.local/lib/python3.10/site-packages/fsspec/asyn.py", line 61, in _runner result[0] = await coro File "/home/alp/.local/lib/python3.10/site-packages/fsspec/asyn.py", line 604, in _get return await _run_coros_in_chunks( File "/home/alp/.local/lib/python3.10/site-packages/fsspec/asyn.py", line 257, in _run_coros_in_chunks await asyncio.gather(*chunk, return_exceptions=return_exceptions), File "/usr/lib/python3.10/asyncio/tasks.py", line 408, in wait_for return await fut File "/home/alp/.local/lib/python3.10/site-packages/s3fs/core.py", line 1193, in _get_file body, content_length = await _open_file(range=0) File "/home/alp/.local/lib/python3.10/site-packages/s3fs/core.py", line 1184, in _open_file resp = await self._call_s3( File "/home/alp/.local/lib/python3.10/site-packages/s3fs/core.py", line 348, in _call_s3 return await _error_wrapper( File "/home/alp/.local/lib/python3.10/site-packages/s3fs/core.py", line 140, in _error_wrapper raise err PermissionError: The difference between the request time and the current time is too large. ``` The usual problem for this error is that the time on my local machine is out of sync with the current time. However, this is not the case here. I checked the time and even reset it with no success. See resources here: - https://stackoverflow.com/questions/4770635/s3-error-the-difference-between-the-request-time-and-the-current-time-is-too-la - https://stackoverflow.com/questions/25964491/aws-s3-upload-fails-requesttimetooskewed The error does not appear when loading a smaller dataset (e.g. our test set) from the same s3 path. ### Steps to reproduce the bug 1. Create large dataset 2. Try loading it from s3 using: ``` dataset = load_from_disk("s3://...", storage_options=storage_options) ``` ### Expected behavior Load dataset without running into this error. ### Environment info - `datasets` version: 2.13.1 - Platform: Linux-5.15.0-91-generic-x86_64-with-glibc2.35 - Python version: 3.10.12 - Huggingface_hub version: 0.19.3 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6611/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6611/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6610
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6610/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6610/comments
https://api.github.com/repos/huggingface/datasets/issues/6610/events
https://github.com/huggingface/datasets/issues/6610
2,095,643,711
I_kwDODunzps586Pw_
6,610
cast_column to Sequence(subfeatures_dict) has err
{ "login": "neiblegy", "id": 16574677, "node_id": "MDQ6VXNlcjE2NTc0Njc3", "avatar_url": "https://avatars.githubusercontent.com/u/16574677?v=4", "gravatar_id": "", "url": "https://api.github.com/users/neiblegy", "html_url": "https://github.com/neiblegy", "followers_url": "https://api.github.com/users/neiblegy/followers", "following_url": "https://api.github.com/users/neiblegy/following{/other_user}", "gists_url": "https://api.github.com/users/neiblegy/gists{/gist_id}", "starred_url": "https://api.github.com/users/neiblegy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neiblegy/subscriptions", "organizations_url": "https://api.github.com/users/neiblegy/orgs", "repos_url": "https://api.github.com/users/neiblegy/repos", "events_url": "https://api.github.com/users/neiblegy/events{/privacy}", "received_events_url": "https://api.github.com/users/neiblegy/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-23T09:32:32
2024-01-25T02:15:23
2024-01-25T02:15:23
NONE
null
null
null
### Describe the bug I am working with the following demo code: ``` from datasets import load_dataset from datasets.features import Sequence, Value, ClassLabel, Features ais_dataset = load_dataset("/data/ryan.gao/ais_dataset_cache/raw/1978/") ais_dataset = ais_dataset["train"] def add_class(example): example["my_labeled_bbox"] = {"bbox": [100,100,200,200], "label": "cat"} return example ais_dataset = ais_dataset.map(add_class, batched=False, num_proc=32) ais_dataset = ais_dataset.cast_column("my_labeled_bbox", Sequence( { "bbox": Sequence(Value(dtype="int64")), "label": ClassLabel(names=["cat", "dog"]) })) print(ais_dataset[0]) ``` However, executing this code results in an error: ``` File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2111, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") TypeError: Couldn't cast array of type int64 to Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None) ``` Upon examining the source code in datasets/table.py at line 2035: ``` if isinstance(feature, Sequence) and isinstance(feature.feature, dict): feature = { name: Sequence(subfeature, length=feature.length) for name, subfeature in feature.feature.items() } ``` I noticed that if subfeature is of type Sequence, the code results in Sequence(Sequence(...), ...) and Sequence(ClassLabel(...), ...), which appears to be the source of the error. ### Steps to reproduce the bug run my demo code ### Expected behavior no exception ### Environment info python 3.9 datasets: 2.16.1
{ "login": "neiblegy", "id": 16574677, "node_id": "MDQ6VXNlcjE2NTc0Njc3", "avatar_url": "https://avatars.githubusercontent.com/u/16574677?v=4", "gravatar_id": "", "url": "https://api.github.com/users/neiblegy", "html_url": "https://github.com/neiblegy", "followers_url": "https://api.github.com/users/neiblegy/followers", "following_url": "https://api.github.com/users/neiblegy/following{/other_user}", "gists_url": "https://api.github.com/users/neiblegy/gists{/gist_id}", "starred_url": "https://api.github.com/users/neiblegy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neiblegy/subscriptions", "organizations_url": "https://api.github.com/users/neiblegy/orgs", "repos_url": "https://api.github.com/users/neiblegy/repos", "events_url": "https://api.github.com/users/neiblegy/events{/privacy}", "received_events_url": "https://api.github.com/users/neiblegy/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6610/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6610/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6609
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6609/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6609/comments
https://api.github.com/repos/huggingface/datasets/issues/6609/events
https://github.com/huggingface/datasets/issues/6609
2,095,085,650
I_kwDODunzps584HhS
6,609
Wrong path for cache directory in offline mode
{ "login": "je-santos", "id": 42117435, "node_id": "MDQ6VXNlcjQyMTE3NDM1", "avatar_url": "https://avatars.githubusercontent.com/u/42117435?v=4", "gravatar_id": "", "url": "https://api.github.com/users/je-santos", "html_url": "https://github.com/je-santos", "followers_url": "https://api.github.com/users/je-santos/followers", "following_url": "https://api.github.com/users/je-santos/following{/other_user}", "gists_url": "https://api.github.com/users/je-santos/gists{/gist_id}", "starred_url": "https://api.github.com/users/je-santos/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/je-santos/subscriptions", "organizations_url": "https://api.github.com/users/je-santos/orgs", "repos_url": "https://api.github.com/users/je-santos/repos", "events_url": "https://api.github.com/users/je-santos/events{/privacy}", "received_events_url": "https://api.github.com/users/je-santos/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false } ]
null
5
2024-01-23T01:47:19
2024-02-06T17:21:25
2024-02-06T17:21:25
NONE
null
null
null
### Describe the bug Dear huggingfacers, I'm trying to use a subset of the-stack dataset. When I run the command the first time ``` dataset = load_dataset( path='bigcode/the-stack', data_dir='data/fortran', split='train' ) ``` It downloads the files and caches them normally. Nevertheless, since my compute nodes are not online (`HF_DATASETS_OFFLINE=1`) . Whenever I try to run the command again, the library is passing the wrong cache path: `Cache directory for the-stack doesn't exist at /Users/user/.cache/huggingface/datasets/bigcode___the-stack/default-data_dir=data%2Ffortran-data_dir=data%2Ffortran` when the right path is: `'/Users/user/.cache/huggingface/datasets/bigcode___the-stack/default-data_dir=data\%2Ffortran` Not sure why those redundancies are included in the path. If I try adding the correct path through the the cache_dir argument it throws an error: ConnectionError: Couldn't reach the Hugging Face Hub for dataset 'bigcode/the-stack': Offline mode is enabled. Your help with this issue is greatly appreciated. Thanks a lot for the great work. ### Steps to reproduce the bug 1: `dataset = load_dataset( path='bigcode/the-stack', data_dir='data/fortran', split='train' )` 2: `HF_DATASETS_OFFLINE=1` 3: `dataset = load_dataset( path='bigcode/the-stack', data_dir='data/fortran', split='train' )` ### Expected behavior being able to use the cached data ### Environment info several different systems
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6609/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6609/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6605
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6605/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6605/comments
https://api.github.com/repos/huggingface/datasets/issues/6605/events
https://github.com/huggingface/datasets/issues/6605
2,090,188,376
I_kwDODunzps58lb5Y
6,605
ELI5 no longer available, but referenced in example code
{ "login": "drdsgvo", "id": 81480344, "node_id": "MDQ6VXNlcjgxNDgwMzQ0", "avatar_url": "https://avatars.githubusercontent.com/u/81480344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/drdsgvo", "html_url": "https://github.com/drdsgvo", "followers_url": "https://api.github.com/users/drdsgvo/followers", "following_url": "https://api.github.com/users/drdsgvo/following{/other_user}", "gists_url": "https://api.github.com/users/drdsgvo/gists{/gist_id}", "starred_url": "https://api.github.com/users/drdsgvo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/drdsgvo/subscriptions", "organizations_url": "https://api.github.com/users/drdsgvo/orgs", "repos_url": "https://api.github.com/users/drdsgvo/repos", "events_url": "https://api.github.com/users/drdsgvo/events{/privacy}", "received_events_url": "https://api.github.com/users/drdsgvo/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-19T10:21:52
2024-02-01T17:58:23
2024-02-01T17:58:22
NONE
null
null
null
Here, an example code is given: https://huggingface.co/docs/transformers/tasks/language_modeling This code + article references the ELI5 dataset. ELI5 is no longer available, as the ELI5 dataset page states: https://huggingface.co/datasets/eli5 "Defunct: Dataset "eli5" is defunct and no longer accessible due to unavailability of the source data. Reddit recently [changed the terms of access](https://www.reddit.com/r/reddit/comments/12qwagm/an_update_regarding_reddits_api/) to its API, making the source data for this dataset unavailable. " Please change the example code to use a different dataset.
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6605/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6605/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6604
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6604/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6604/comments
https://api.github.com/repos/huggingface/datasets/issues/6604/events
https://github.com/huggingface/datasets/issues/6604
2,089,713,945
I_kwDODunzps58joEZ
6,604
Transform fingerprint collisions due to setting fixed random seed
{ "login": "normster", "id": 6687910, "node_id": "MDQ6VXNlcjY2ODc5MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/6687910?v=4", "gravatar_id": "", "url": "https://api.github.com/users/normster", "html_url": "https://github.com/normster", "followers_url": "https://api.github.com/users/normster/followers", "following_url": "https://api.github.com/users/normster/following{/other_user}", "gists_url": "https://api.github.com/users/normster/gists{/gist_id}", "starred_url": "https://api.github.com/users/normster/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/normster/subscriptions", "organizations_url": "https://api.github.com/users/normster/orgs", "repos_url": "https://api.github.com/users/normster/repos", "events_url": "https://api.github.com/users/normster/events{/privacy}", "received_events_url": "https://api.github.com/users/normster/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-19T06:32:25
2024-01-26T15:05:35
2024-01-26T15:05:35
NONE
null
null
null
### Describe the bug The transform fingerprinting logic relies on the `random` library for random bits when the function is not hashable (e.g. bound methods as used in `trl`: https://github.com/huggingface/trl/blob/main/trl/trainer/dpo_trainer.py#L356). This causes collisions when the training code sets a fixed random seed, which is common practice: https://github.com/huggingface/alignment-handbook/blob/main/recipes/zephyr-7b-beta/sft/config_full.yaml#L45. This results in fingerprint collisions which leads to silently loading incorrect cache files corresponding to completely different datasets. ### Steps to reproduce the bug n/a ### Expected behavior Use `uuid` v4 instead of `random.getrandbits()` ### Environment info `datasets` main branch
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6604/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6604/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6603
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6603/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6603/comments
https://api.github.com/repos/huggingface/datasets/issues/6603/events
https://github.com/huggingface/datasets/issues/6603
2,089,230,766
I_kwDODunzps58hyGu
6,603
datasets map `cache_file_name` does not work
{ "login": "ChenchaoZhao", "id": 35147961, "node_id": "MDQ6VXNlcjM1MTQ3OTYx", "avatar_url": "https://avatars.githubusercontent.com/u/35147961?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ChenchaoZhao", "html_url": "https://github.com/ChenchaoZhao", "followers_url": "https://api.github.com/users/ChenchaoZhao/followers", "following_url": "https://api.github.com/users/ChenchaoZhao/following{/other_user}", "gists_url": "https://api.github.com/users/ChenchaoZhao/gists{/gist_id}", "starred_url": "https://api.github.com/users/ChenchaoZhao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ChenchaoZhao/subscriptions", "organizations_url": "https://api.github.com/users/ChenchaoZhao/orgs", "repos_url": "https://api.github.com/users/ChenchaoZhao/repos", "events_url": "https://api.github.com/users/ChenchaoZhao/events{/privacy}", "received_events_url": "https://api.github.com/users/ChenchaoZhao/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
2
2024-01-18T23:08:30
2024-01-28T04:01:15
null
NONE
null
null
null
### Describe the bug In the documentation `datasets.Dataset.map` arg `cache_file_name` is said to be a string, but it doesn't work. ### Steps to reproduce the bug 1. pick a dataset 2. write a map function 3. do `ds.map(..., cache_file_name='some_filename')` 4. it crashes ### Expected behavior It will tell you the filename you specified does not exist or it will generate a new file and tell you the filename does not exist. ### Environment info - `datasets` version: 2.16.0 - Platform: Linux-5.10.201-168.748.amzn2int.x86_64-x86_64-with-glibc2.26 - Python version: 3.10.13 - `huggingface_hub` version: 0.20.2 - PyArrow version: 14.0.2 - Pandas version: 2.1.4 - `fsspec` version: 2023.12.2
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6603/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6603/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6602
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6602/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6602/comments
https://api.github.com/repos/huggingface/datasets/issues/6602/events
https://github.com/huggingface/datasets/issues/6602
2,089,217,483
I_kwDODunzps58hu3L
6,602
Index error when data is large
{ "login": "ChenchaoZhao", "id": 35147961, "node_id": "MDQ6VXNlcjM1MTQ3OTYx", "avatar_url": "https://avatars.githubusercontent.com/u/35147961?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ChenchaoZhao", "html_url": "https://github.com/ChenchaoZhao", "followers_url": "https://api.github.com/users/ChenchaoZhao/followers", "following_url": "https://api.github.com/users/ChenchaoZhao/following{/other_user}", "gists_url": "https://api.github.com/users/ChenchaoZhao/gists{/gist_id}", "starred_url": "https://api.github.com/users/ChenchaoZhao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ChenchaoZhao/subscriptions", "organizations_url": "https://api.github.com/users/ChenchaoZhao/orgs", "repos_url": "https://api.github.com/users/ChenchaoZhao/repos", "events_url": "https://api.github.com/users/ChenchaoZhao/events{/privacy}", "received_events_url": "https://api.github.com/users/ChenchaoZhao/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2024-01-18T23:00:47
2024-01-18T23:00:47
null
NONE
null
null
null
### Describe the bug At `save_to_disk` step, the `max_shard_size` by default is `500MB`. However, one row of the dataset might be larger than `500MB` then the saving will throw an index error. Without looking at the source code, the bug is due to wrong calculation of number of shards which i think is `total_size / min(max_shard_size, row_size)` which should be `total_size / max(max_shard_size, row_size)` The fix is setting a larger `max_shard_size` ### Steps to reproduce the bug 1. create a dataset with large dense tensors per row 2. set a small `max_shard_size` say 1MB 3. `save_to_disk` ### Expected behavior ``` raise IndexError(f"Index {index} out of range for dataset of size {size}.") IndexError: Index 10 out of range for dataset of size 10. ``` ### Environment info - `datasets` version: 2.16.0 - Platform: Linux-5.10.201-168.748.amzn2int.x86_64-x86_64-with-glibc2.26 - Python version: 3.10.13 - `huggingface_hub` version: 0.20.2 - PyArrow version: 14.0.2 - Pandas version: 2.1.4 - `fsspec` version: 2023.12.2
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6602/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6602/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6600
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6600/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6600/comments
https://api.github.com/repos/huggingface/datasets/issues/6600/events
https://github.com/huggingface/datasets/issues/6600
2,088,446,385
I_kwDODunzps58eymx
6,600
Loading CSV exported dataset has unexpected format
{ "login": "OrianeN", "id": 59572247, "node_id": "MDQ6VXNlcjU5NTcyMjQ3", "avatar_url": "https://avatars.githubusercontent.com/u/59572247?v=4", "gravatar_id": "", "url": "https://api.github.com/users/OrianeN", "html_url": "https://github.com/OrianeN", "followers_url": "https://api.github.com/users/OrianeN/followers", "following_url": "https://api.github.com/users/OrianeN/following{/other_user}", "gists_url": "https://api.github.com/users/OrianeN/gists{/gist_id}", "starred_url": "https://api.github.com/users/OrianeN/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/OrianeN/subscriptions", "organizations_url": "https://api.github.com/users/OrianeN/orgs", "repos_url": "https://api.github.com/users/OrianeN/repos", "events_url": "https://api.github.com/users/OrianeN/events{/privacy}", "received_events_url": "https://api.github.com/users/OrianeN/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
2
2024-01-18T14:48:27
2024-01-23T14:42:32
null
NONE
null
null
null
### Describe the bug I wanted to be able to save a HF dataset for translations and load it again in another script, but I'm a bit confused with the documentation and the result I've got so I'm opening this issue to ask if this behavior is as expected. ### Steps to reproduce the bug The documentation I've mainly consulted is https://huggingface.co/docs/datasets/v2.16.1/en/package_reference/loading_methods#datasets.load_dataset and https://huggingface.co/docs/datasets/package_reference/main_classes#datasets.Dataset (where I've found `.to_csv()`) ```python # Load a dataset of translations test_dataset = load_dataset("opus100", name="en-fr", split="test") # Save with .to_csv() test_csv_path = "try_testset_save.csv" test_dataset.to_csv(test_csv_path) # Load dataset from the CSV loaded_dataset = load_dataset("csv", data_files=test_csv_path) print(test_dataset_fromfile[0]["translation"]) print(test_dataset_fromfile[0]["translation"]["en"]) ``` ``` Creating CSV from Arrow format: 100% 2/2 [00:00<00:00, 47.99ba/s] Downloading data files: 100% 1/1 [00:00<00:00, 65.33it/s] Extracting data files: 100% 1/1 [00:00<00:00, 42.10it/s] Generating train split: 2000/0 [00:00<00:00, 47486.09 examples/s] {'en': "She wasn't going to vaccinate her kid against polio, no way.", 'fr': 'Elle ne vaccinerait pas son enfant contre la polio. Pas question.'} --------------------------------------------------------------------------- TypeError Traceback (most recent call last) Cell In[29], line 11 9 loaded_dataset = load_dataset("csv", data_files=test_csv_path) 10 print(test_dataset_fromfile[0]["translation"]) ---> 11 print(test_dataset_fromfile[0]["translation"]["en"]) TypeError: string indices must be integers, not 'str' ``` ### Expected behavior Each translation was saved as a stringified dict like `"{'en': ""She wasn't going to vaccinate her kid against polio, no way."", 'fr': 'Elle ne vaccinerait pas son enfant contre la polio. Pas question.'}"` where I would have expected 2 columns (1st with English segments, and 2nd with French segments), and I was expecting `load_dataset` to infer the type of feature automatically as I haven't seen anything about it in the documentation. Do you have an example of how to effectively save and load datasets of translations ? ### Environment info - `datasets` version: 2.15.0 - Platform: Linux-3.10.0-1160.36.2.el7.x86_64-x86_64-with-glibc2.17 - Python version: 3.11.5 - `huggingface_hub` version: 0.16.4 - PyArrow version: 14.0.2 - Pandas version: 2.1.4 - `fsspec` version: 2023.10.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6600/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6600/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6599
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6599/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6599/comments
https://api.github.com/repos/huggingface/datasets/issues/6599/events
https://github.com/huggingface/datasets/issues/6599
2,086,684,664
I_kwDODunzps58YEf4
6,599
Easy way to segment into 30s snippets given an m4a file and a vtt file
{ "login": "RonanKMcGovern", "id": 78278410, "node_id": "MDQ6VXNlcjc4Mjc4NDEw", "avatar_url": "https://avatars.githubusercontent.com/u/78278410?v=4", "gravatar_id": "", "url": "https://api.github.com/users/RonanKMcGovern", "html_url": "https://github.com/RonanKMcGovern", "followers_url": "https://api.github.com/users/RonanKMcGovern/followers", "following_url": "https://api.github.com/users/RonanKMcGovern/following{/other_user}", "gists_url": "https://api.github.com/users/RonanKMcGovern/gists{/gist_id}", "starred_url": "https://api.github.com/users/RonanKMcGovern/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/RonanKMcGovern/subscriptions", "organizations_url": "https://api.github.com/users/RonanKMcGovern/orgs", "repos_url": "https://api.github.com/users/RonanKMcGovern/repos", "events_url": "https://api.github.com/users/RonanKMcGovern/events{/privacy}", "received_events_url": "https://api.github.com/users/RonanKMcGovern/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-01-17T17:51:40
2024-01-23T10:42:17
2024-01-22T15:35:49
NONE
null
null
null
### Feature request Uploading datasets is straightforward thanks to the ability to push Audio to hub. However, it would be nice if the data (text and audio) could be segmented when being pushed (if not possible already). ### Motivation It's easy to create a vtt file from an audio file. If there could be auto-segmenting, this would make the creation of datasets much faster. ### Your contribution I have made a custom script to do this but it's not all that clean - uses librosa and pydub.
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6599/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6599/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6598
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6598/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6598/comments
https://api.github.com/repos/huggingface/datasets/issues/6598/events
https://github.com/huggingface/datasets/issues/6598
2,084,236,605
I_kwDODunzps58Ou09
6,598
Unexpected keyword argument 'hf' when downloading CSV dataset from S3
{ "login": "dguenms", "id": 5592111, "node_id": "MDQ6VXNlcjU1OTIxMTE=", "avatar_url": "https://avatars.githubusercontent.com/u/5592111?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dguenms", "html_url": "https://github.com/dguenms", "followers_url": "https://api.github.com/users/dguenms/followers", "following_url": "https://api.github.com/users/dguenms/following{/other_user}", "gists_url": "https://api.github.com/users/dguenms/gists{/gist_id}", "starred_url": "https://api.github.com/users/dguenms/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dguenms/subscriptions", "organizations_url": "https://api.github.com/users/dguenms/orgs", "repos_url": "https://api.github.com/users/dguenms/repos", "events_url": "https://api.github.com/users/dguenms/events{/privacy}", "received_events_url": "https://api.github.com/users/dguenms/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
7
2024-01-16T15:16:01
2024-07-23T14:30:10
2024-07-23T14:30:10
NONE
null
null
null
### Describe the bug I receive this error message when using `load_dataset` with "csv" path and `dataset_files=s3://...`: ``` TypeError: Session.__init__() got an unexpected keyword argument 'hf' ``` I found a similar issue here: https://stackoverflow.com/questions/77596258/aws-issue-load-dataset-from-s3-fails-with-unexpected-keyword-argument-error-in Full stacktrace: ``` .../site-packages/datasets/load.py:2549: in load_dataset builder_instance.download_and_prepare( .../site-packages/datasets/builder.py:1005: in download_and_prepare self._download_and_prepare( .../site-packages/datasets/builder.py:1078: in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) .../site-packages/datasets/packaged_modules/csv/csv.py:147: in _split_generators data_files = dl_manager.download_and_extract(self.config.data_files) .../site-packages/datasets/download/download_manager.py:562: in download_and_extract return self.extract(self.download(url_or_urls)) .../site-packages/datasets/download/download_manager.py:426: in download downloaded_path_or_paths = map_nested( .../site-packages/datasets/utils/py_utils.py:466: in map_nested mapped = [ .../site-packages/datasets/utils/py_utils.py:467: in <listcomp> _single_map_nested((function, obj, types, None, True, None)) .../site-packages/datasets/utils/py_utils.py:387: in _single_map_nested mapped = [_single_map_nested((function, v, types, None, True, None)) for v in pbar] .../site-packages/datasets/utils/py_utils.py:387: in <listcomp> mapped = [_single_map_nested((function, v, types, None, True, None)) for v in pbar] .../site-packages/datasets/utils/py_utils.py:370: in _single_map_nested return function(data_struct) .../site-packages/datasets/download/download_manager.py:451: in _download out = cached_path(url_or_filename, download_config=download_config) .../site-packages/datasets/utils/file_utils.py:188: in cached_path output_path = get_from_cache( ...1/site-packages/datasets/utils/file_utils.py:511: in get_from_cache response = fsspec_head(url, storage_options=storage_options) .../site-packages/datasets/utils/file_utils.py:316: in fsspec_head fs, _, paths = fsspec.get_fs_token_paths(url, storage_options=storage_options) .../site-packages/fsspec/core.py:622: in get_fs_token_paths fs = filesystem(protocol, **inkwargs) .../site-packages/fsspec/registry.py:290: in filesystem return cls(**storage_options) .../site-packages/fsspec/spec.py:79: in __call__ obj = super().__call__(*args, **kwargs) .../site-packages/s3fs/core.py:187: in __init__ self.s3 = self.connect() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <s3fs.core.S3FileSystem object at 0x1500a1310>, refresh = True def connect(self, refresh=True): """ Establish S3 connection object. Parameters ---------- refresh : bool Whether to create new session/client, even if a previous one with the same parameters already exists. If False (default), an existing one will be used if possible """ if refresh is False: # back compat: we store whole FS instance now return self.s3 anon, key, secret, kwargs, ckwargs, token, ssl = ( self.anon, self.key, self.secret, self.kwargs, self.client_kwargs, self.token, self.use_ssl) if not self.passed_in_session: > self.session = botocore.session.Session(**self.kwargs) E TypeError: Session.__init__() got an unexpected keyword argument 'hf' ``` ### Steps to reproduce the bug 1. Assuming a valid CSV file located at `s3://bucket/data.csv` 2. Run the below code: ``` storage_options = { "key": "...", "secret": "...", "client_kwargs": { "endpoint_url": "...", } } load_dataset("csv", data_files="s3://bucket/data.csv", storage_options=storage_options) ``` Encountered in version `2.16.1` but also reproduced in `2.16.0` and `2.15.0`. Note: I encountered this in a unit test using a `moto` mock for S3, however since the error occurs before the session is instantiated, it should not be the issue. ### Expected behavior No exception is raised, the boto3 session is created successfully, and the CSV file is downloaded successfully and returned as a dataset. === After some research I found that `DownloadConfig` has a `__post_init__` method that always forces this value to be set in its `storage_options`, even though in case of an S3 location the storage options get passed on to the S3 Session which does not expect this parameter. I assume this parameter is needed when reading from the huggingface hub and should not be set in this context. Unfortunately there is nothing the user can do to work around it. Even if you manually do something like: ``` download_config = DownloadConfig() del download_config.storage_options["hf"] load_dataset("csv", data_files="s3://bucket/data.csv", download_config=download_config) ``` the library will still reinsert this parameter when `download_config = self.download_config.copy()` in line 418 of `download_manager.py` (`DownloadManager.download`). Therefore `load_dataset` currently cannot be used to read a dataset in CSV format from an S3 location. ### Environment info - `datasets` version: 2.16.1 - Platform: macOS-14.2.1-arm64-arm-64bit - Python version: 3.11.7 - `huggingface_hub` version: 0.20.2 - PyArrow version: 14.0.2 - Pandas version: 2.1.4 - `fsspec` version: 2023.10.0
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6598/reactions", "total_count": 9, "+1": 9, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6598/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6597
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6597/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6597/comments
https://api.github.com/repos/huggingface/datasets/issues/6597/events
https://github.com/huggingface/datasets/issues/6597
2,083,708,521
I_kwDODunzps58Mt5p
6,597
Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
6
2024-01-16T11:27:07
2024-02-05T12:29:37
2024-02-05T12:29:37
MEMBER
null
null
null
While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace. ## Steps to reproduce the bug The command: ```python commit_info = ds.push_to_hub( "caner", config_name="default", commit_message="Convert dataset to Parquet", commit_description="Convert dataset to Parquet.", create_pr=True, token=token, ) ``` creates the additional dataset `albertvillanova/caner`.
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6597/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6597/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6595
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6595/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6595/comments
https://api.github.com/repos/huggingface/datasets/issues/6595/events
https://github.com/huggingface/datasets/issues/6595
2,082,896,148
I_kwDODunzps58JnkU
6,595
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
14
2024-01-16T02:03:09
2024-01-27T18:26:33
2024-01-26T02:28:32
NONE
null
null
null
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import datasets from datasets import load_from_disk dataset = load_from_disk("ds") datasets.config.DEFAULT_MAX_BATCH_SIZE = 1 dataset.push_to_hub("kopyl/ds", private=True, max_shard_size="500MB") ``` And i get this error: `pyarrow.lib.ArrowNotImplementedError: Unhandled type for Arrow to Parquet schema conversion: halffloat` Full traceback: ``` >>> dataset.push_to_hub("kopyl/3M_icons_monochrome_only_no_captioning_mapped-for-SDXL-2", private=True, max_shard_size="500MB") Map: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1451/1451 [00:00<00:00, 6827.40 examples/s] Uploading the dataset shards: 0%| | 0/2099 [00:00<?, ?it/s] Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/local/lib/python3.10/dist-packages/datasets/dataset_dict.py", line 1705, in push_to_hub split_additions, uploaded_size, dataset_nbytes = self[split]._push_parquet_shards_to_hub( File "/usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py", line 5208, in _push_parquet_shards_to_hub shard.to_parquet(buffer) File "/usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py", line 4931, in to_parquet return ParquetDatasetWriter(self, path_or_buf, batch_size=batch_size, **parquet_writer_kwargs).write() File "/usr/local/lib/python3.10/dist-packages/datasets/io/parquet.py", line 129, in write written = self._write(file_obj=self.path_or_buf, batch_size=batch_size, **self.parquet_writer_kwargs) File "/usr/local/lib/python3.10/dist-packages/datasets/io/parquet.py", line 141, in _write writer = pq.ParquetWriter(file_obj, schema=schema, **parquet_writer_kwargs) File "/usr/local/lib/python3.10/dist-packages/pyarrow/parquet/core.py", line 1016, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 1869, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Unhandled type for Arrow to Parquet schema conversion: halffloat ``` Smaller datasets with the same way of saving and pushing work wonders. Big ones are not. I'm currently trying to upload dataset like this: `HfApi().upload_folder...` But i'm not sure that in this case "load_dataset" would work well. This setting num_shards does not help too: ``` dataset.push_to_hub("kopyl/3M_icons_monochrome_only_no_captioning_mapped-for-SDXL-2", private=True, num_shards={'train': 500}) ``` Tried 3000, 500, 478, 100 Also do you know if it's possible to push a dataset with multiple processes? It would take an eternity pushing 1TB... ### Steps to reproduce the bug Described above ### Expected behavior Should be able to upload... ### Environment info Total dataset size: 978G Amount of `.arrow` files: 2101 Each `.arrow` file size: 477M (i know 477 megabytes * 2101 does not equal 978G, but i just checked the size of a couple `.arrow` files, i don't know if some might have different size) Some files: - "ds/train/state.json": https://pastebin.com/tJ3ZLGAg - "ds/train/dataset_info.json": https://pastebin.com/JdXMQ5ih
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6595/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6595/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6594
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6594/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6594/comments
https://api.github.com/repos/huggingface/datasets/issues/6594/events
https://github.com/huggingface/datasets/issues/6594
2,082,748,275
I_kwDODunzps58JDdz
6,594
IterableDataset sharding logic needs improvement
{ "login": "rwightman", "id": 5702664, "node_id": "MDQ6VXNlcjU3MDI2NjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/5702664?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rwightman", "html_url": "https://github.com/rwightman", "followers_url": "https://api.github.com/users/rwightman/followers", "following_url": "https://api.github.com/users/rwightman/following{/other_user}", "gists_url": "https://api.github.com/users/rwightman/gists{/gist_id}", "starred_url": "https://api.github.com/users/rwightman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rwightman/subscriptions", "organizations_url": "https://api.github.com/users/rwightman/orgs", "repos_url": "https://api.github.com/users/rwightman/repos", "events_url": "https://api.github.com/users/rwightman/events{/privacy}", "received_events_url": "https://api.github.com/users/rwightman/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2024-01-15T22:22:36
2024-01-15T22:25:10
null
NONE
null
null
null
### Describe the bug The sharding of IterableDatasets with respect to distributed and dataloader worker processes appears problematic with significant performance traps and inconsistencies wrt to distributed train processes vs worker processes. Splitting across num_workers (per train process loader processes) and world_size (distributed training processes) appears inconsistent. * worker split: https://github.com/huggingface/datasets/blob/9d6d16117a30ba345b0236407975f701c5b288d4/src/datasets/iterable_dataset.py#L1266-L1283 * distributed split: https://github.com/huggingface/datasets/blob/9d6d16117a30ba345b0236407975f701c5b288d4/src/datasets/iterable_dataset.py#L1335-L1356 In the case of the distributed split, there is a modulus check that flips between two very different behaviours, why is this different than splitting across the data loader workers? For IterableDatasets the DataLoaders worker processes are independent, so whether it's workers within one train process or across a distributed world the shards should be distributed the same, across `world_size * num_worker` independent workers in either case... Further, the fallback case when the `n_shards % world_size == 0` check fails is a rather extreme change. I argue it is not desirable to do that implicitly, it should be an explicit case for specific scenarios (ie reliable validation). A train scenario would likely be much better handled with improved wrapping / stopping behaviour to eg also fix #6437. Changing from stepping shards to stepping samples means that every single process reads ALL of the shards. This was never an intended default for sharded training, shards gain their performance advantage in large scale distributed training by explicitly avoiding the need to have every process overlapping in the data they read, by default, only the data allocated to each process via their assigned shards should be read in each pass of the dataset. Using a large scale CLIP example, some of the larger datasets have 10-20k shards across 100+TB of data. Training with 1000 GPUs we are switching between reading 100 terabytes per epoch to 100 petabytes if say change 20k % 1000 and drop one gpu-node to 20k % 992. The 'step over samples' case might be worth the overhead in specific validation scenarios where gaurantees of at least/most once samples seen are more important and do not make up a significant portion of train time or are done in smaller world sizes outside of train. ### Steps to reproduce the bug N/A ### Expected behavior We have an iterable dataset with N shards, to split across workers * shuffle shards (same seed across all train processes) * step shard iterator across distributed processes * step shard iterator across dataloader worker processes * shuffle samples in every worker via shuffle buffer (different seed in each worker, but ideally controllable (based on base seed + worker id + epoch). * end up with (possibly uneven) number of shards per worker but each shard only ever accessed by 1 worker per pass (epoch) ### Environment info N/A
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6594/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6594/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6592
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6592/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6592/comments
https://api.github.com/repos/huggingface/datasets/issues/6592/events
https://github.com/huggingface/datasets/issues/6592
2,082,410,257
I_kwDODunzps58Hw8R
6,592
Logs are delayed when doing .map when `docker logs`
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-15T17:05:21
2024-02-12T17:35:21
2024-02-12T17:35:21
NONE
null
null
null
### Describe the bug When I run my SD training in a Docker image and then listen to logs like `docker logs train -f`, the progress bar is delayed. It's updating every few percent. When you have a large dataset that has to be mapped (like 1+ million samples), it's crucial to see the updates in real-time, not every couple hours to make sure nothing got frozen or broken ### Steps to reproduce the bug 1. Run any huge dataset processing as a Docker image 2. `docker logs image_name` to it ### Expected behavior ... ### Environment info ...
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6592/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6592/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6591
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6591/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6591/comments
https://api.github.com/repos/huggingface/datasets/issues/6591/events
https://github.com/huggingface/datasets/issues/6591
2,082,378,957
I_kwDODunzps58HpTN
6,591
The datasets models housed in Dropbox can't support a lot of users downloading them
{ "login": "RDaneelOlivav", "id": 4933774, "node_id": "MDQ6VXNlcjQ5MzM3NzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/4933774?v=4", "gravatar_id": "", "url": "https://api.github.com/users/RDaneelOlivav", "html_url": "https://github.com/RDaneelOlivav", "followers_url": "https://api.github.com/users/RDaneelOlivav/followers", "following_url": "https://api.github.com/users/RDaneelOlivav/following{/other_user}", "gists_url": "https://api.github.com/users/RDaneelOlivav/gists{/gist_id}", "starred_url": "https://api.github.com/users/RDaneelOlivav/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/RDaneelOlivav/subscriptions", "organizations_url": "https://api.github.com/users/RDaneelOlivav/orgs", "repos_url": "https://api.github.com/users/RDaneelOlivav/repos", "events_url": "https://api.github.com/users/RDaneelOlivav/events{/privacy}", "received_events_url": "https://api.github.com/users/RDaneelOlivav/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-15T16:43:38
2024-01-22T23:18:09
2024-01-22T23:18:09
NONE
null
null
null
### Describe the bug I'm using the datasets ``` from datasets import load_dataset, Audio dataset = load_dataset("PolyAI/minds14", name="en-US", split="train") ``` And it seems that sometimes when I imagine a lot of users are accessing the same resources, the Dropbox host fails: `raise ConnectionError(f"Couldn't reach {url} (error {response.status_code})") ConnectionError: Couldn't reach https://www.dropbox.com/s/e2us0hcs3ilr20e/MInDS-14.zip?dl=1 (error 429)` My question is if we can somehow host these files elsewhere or can you change the limit of simultaneous users accessing those resources or any other solution? Also, has anyone had this issue before? Thanks ### Steps to reproduce the bug 1: Create a python script like so: ``` from datasets import load_dataset, Audio dataset = load_dataset("PolyAI/minds14", name="en-US", split="train") ``` 2: Execute this by a certain number of users at the same time ### Expected behavior I woudl expect that this shouldnt happen unless its a huge amount of users, which it is not the case ### Environment info This was done in an Ubuntu 22 environment.
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6591/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6591/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6590
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6590/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6590/comments
https://api.github.com/repos/huggingface/datasets/issues/6590/events
https://github.com/huggingface/datasets/issues/6590
2,082,000,084
I_kwDODunzps58GMzU
6,590
Feature request: Multi-GPU dataset mapping for SDXL training
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-01-15T13:06:06
2024-01-15T13:07:07
null
NONE
null
null
null
### Feature request We need to speed up SDXL dataset pre-process. Please make it possible to use multiple GPUs for the [official SDXL trainer](https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_sdxl.py) :) ### Motivation Pre-computing 3 million of images takes around 2 days. Would be nice to be able to be able to do multi-GPU (or even better – multi-GPU + multi-node) vae and embedding precompute... ### Your contribution I'm not sure i can wrap my head around the multi-GPU mapping... Plus it's too expensive for me to take x2 A100 and spend a day just figuring out the staff since I don't have a job right now.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6590/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6590/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6589
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6589/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6589/comments
https://api.github.com/repos/huggingface/datasets/issues/6589/events
https://github.com/huggingface/datasets/issues/6589
2,081,358,619
I_kwDODunzps58DwMb
6,589
After `2.16.0` version, there are `PermissionError` when users use shared cache_dir
{ "login": "minhopark-neubla", "id": 106717516, "node_id": "U_kgDOBlxhTA", "avatar_url": "https://avatars.githubusercontent.com/u/106717516?v=4", "gravatar_id": "", "url": "https://api.github.com/users/minhopark-neubla", "html_url": "https://github.com/minhopark-neubla", "followers_url": "https://api.github.com/users/minhopark-neubla/followers", "following_url": "https://api.github.com/users/minhopark-neubla/following{/other_user}", "gists_url": "https://api.github.com/users/minhopark-neubla/gists{/gist_id}", "starred_url": "https://api.github.com/users/minhopark-neubla/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/minhopark-neubla/subscriptions", "organizations_url": "https://api.github.com/users/minhopark-neubla/orgs", "repos_url": "https://api.github.com/users/minhopark-neubla/repos", "events_url": "https://api.github.com/users/minhopark-neubla/events{/privacy}", "received_events_url": "https://api.github.com/users/minhopark-neubla/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-15T06:46:27
2024-02-02T07:55:38
2024-01-30T15:28:38
NONE
null
null
null
### Describe the bug - We use shared `cache_dir` using `HF_HOME="{shared_directory}"` - After dataset version 2.16.0, datasets uses `filelock` package for file locking #6445 - But, `filelock` package make `.lock` file with `644` permission - Dataset is not available to other users except the user who created the lock file via `load_dataset`. ### Steps to reproduce the bug 1. `pip install datasets==2.16.0` 2. `export HF_HOME="{shared_directory}"` 3. download dataset with `load_dataset` 4. logout and login another user 5. `pip install datasets==2.16.0` 6. `export HF_HOME="{shared_directory}"` 7. download dataset with `load_dataset` 8. `PermissionError` occurs ### Expected behavior - Users can share `cache_dir` using environment variable `HF_HOME` ### Environment info - python == 3.9.10 - datasets == 2.16.0 - ubuntu 22.04 - shared_directory has ACL ![image (1)](https://github.com/huggingface/datasets/assets/106717516/5ca759db-ad0c-4883-9a97-9c8fccd00d8a) - users are same group (developers)
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6589/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6589/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6588
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6588/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6588/comments
https://api.github.com/repos/huggingface/datasets/issues/6588/events
https://github.com/huggingface/datasets/issues/6588
2,081,284,253
I_kwDODunzps58DeCd
6,588
fix os.listdir return name is empty string
{ "login": "d710055071", "id": 12895488, "node_id": "MDQ6VXNlcjEyODk1NDg4", "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "gravatar_id": "", "url": "https://api.github.com/users/d710055071", "html_url": "https://github.com/d710055071", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url": "https://api.github.com/users/d710055071/gists{/gist_id}", "starred_url": "https://api.github.com/users/d710055071/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/d710055071/subscriptions", "organizations_url": "https://api.github.com/users/d710055071/orgs", "repos_url": "https://api.github.com/users/d710055071/repos", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "received_events_url": "https://api.github.com/users/d710055071/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-15T05:34:36
2024-01-24T10:08:29
2024-01-24T10:08:29
CONTRIBUTOR
null
null
null
### Describe the bug xlistdir return name is empty string Overloaded os.listdir ### Steps to reproduce the bug ```python from datasets.download.streaming_download_manager import xjoin from datasets.download.streaming_download_manager import xlistdir config = DownloadConfig(storage_options=options) manger = StreamingDownloadManager("ILSVRC2012",download_config=config) input_path = "lakefs://datalab/main/imagenet/ILSVRC2012.zip" download_files = manger.download_and_extract(input_path) current_dir = xjoin(download_files,"ILSVRC2012/Images/ILSVRC2012_img_train") folder_list = xlistdir(current_dir) ``` in xlistdir function Obj ["name"] ends with "/" last return "" ### Expected behavior Obj ["name"] ends with "/" return folder name ### Environment info no
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6588/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6588/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6585
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6585/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6585/comments
https://api.github.com/repos/huggingface/datasets/issues/6585/events
https://github.com/huggingface/datasets/issues/6585
2,078,874,005
I_kwDODunzps576RmV
6,585
losing DatasetInfo in Dataset.map when num_proc > 1
{ "login": "JochenSiegWork", "id": 135010976, "node_id": "U_kgDOCAwaoA", "avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JochenSiegWork", "html_url": "https://github.com/JochenSiegWork", "followers_url": "https://api.github.com/users/JochenSiegWork/followers", "following_url": "https://api.github.com/users/JochenSiegWork/following{/other_user}", "gists_url": "https://api.github.com/users/JochenSiegWork/gists{/gist_id}", "starred_url": "https://api.github.com/users/JochenSiegWork/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JochenSiegWork/subscriptions", "organizations_url": "https://api.github.com/users/JochenSiegWork/orgs", "repos_url": "https://api.github.com/users/JochenSiegWork/repos", "events_url": "https://api.github.com/users/JochenSiegWork/events{/privacy}", "received_events_url": "https://api.github.com/users/JochenSiegWork/received_events", "type": "User", "site_admin": false }
[]
open
false
{ "login": "JochenSiegWork", "id": 135010976, "node_id": "U_kgDOCAwaoA", "avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JochenSiegWork", "html_url": "https://github.com/JochenSiegWork", "followers_url": "https://api.github.com/users/JochenSiegWork/followers", "following_url": "https://api.github.com/users/JochenSiegWork/following{/other_user}", "gists_url": "https://api.github.com/users/JochenSiegWork/gists{/gist_id}", "starred_url": "https://api.github.com/users/JochenSiegWork/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JochenSiegWork/subscriptions", "organizations_url": "https://api.github.com/users/JochenSiegWork/orgs", "repos_url": "https://api.github.com/users/JochenSiegWork/repos", "events_url": "https://api.github.com/users/JochenSiegWork/events{/privacy}", "received_events_url": "https://api.github.com/users/JochenSiegWork/received_events", "type": "User", "site_admin": false }
[ { "login": "JochenSiegWork", "id": 135010976, "node_id": "U_kgDOCAwaoA", "avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JochenSiegWork", "html_url": "https://github.com/JochenSiegWork", "followers_url": "https://api.github.com/users/JochenSiegWork/followers", "following_url": "https://api.github.com/users/JochenSiegWork/following{/other_user}", "gists_url": "https://api.github.com/users/JochenSiegWork/gists{/gist_id}", "starred_url": "https://api.github.com/users/JochenSiegWork/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JochenSiegWork/subscriptions", "organizations_url": "https://api.github.com/users/JochenSiegWork/orgs", "repos_url": "https://api.github.com/users/JochenSiegWork/repos", "events_url": "https://api.github.com/users/JochenSiegWork/events{/privacy}", "received_events_url": "https://api.github.com/users/JochenSiegWork/received_events", "type": "User", "site_admin": false } ]
null
2
2024-01-12T13:39:19
2024-01-12T14:08:24
null
CONTRIBUTOR
null
null
null
### Describe the bug Hello and thanks for developing this package! When I process a Dataset with the map function using multiple processors some set attributes of the DatasetInfo get lost and are None in the resulting Dataset. ### Steps to reproduce the bug ```python from datasets import Dataset, DatasetInfo def run_map(num_proc): dataset = Dataset.from_dict( {"col1": [0, 1], "col2": [3, 4]}, info=DatasetInfo( dataset_name="my_dataset", ), ) ds = dataset.map(lambda x: x, num_proc=num_proc) print(ds.info.dataset_name) run_map(1) run_map(2) ``` This puts out: ```bash Map: 100%|██████████| 2/2 [00:00<00:00, 724.66 examples/s] my_dataset Map (num_proc=2): 100%|██████████| 2/2 [00:00<00:00, 18.25 examples/s] None ``` ### Expected behavior I expect the DatasetInfo to be kept as it was and there should be no difference in the output of running map with num_proc=1 and num_proc=2. Expected output: ```bash Map: 100%|██████████| 2/2 [00:00<00:00, 724.66 examples/s] my_dataset Map (num_proc=2): 100%|██████████| 2/2 [00:00<00:00, 18.25 examples/s] my_dataset ``` ### Environment info - `datasets` version: 2.16.1 - Platform: Linux-5.10.102.1-microsoft-standard-WSL2-x86_64-with-glibc2.17 - Python version: 3.8.18 - `huggingface_hub` version: 0.20.2 - PyArrow version: 12.0.1 - Pandas version: 2.0.3 - `fsspec` version: 2023.9.2
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6585/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6585/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6584
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6584/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6584/comments
https://api.github.com/repos/huggingface/datasets/issues/6584/events
https://github.com/huggingface/datasets/issues/6584
2,078,454,878
I_kwDODunzps574rRe
6,584
np.fromfile not supported
{ "login": "d710055071", "id": 12895488, "node_id": "MDQ6VXNlcjEyODk1NDg4", "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "gravatar_id": "", "url": "https://api.github.com/users/d710055071", "html_url": "https://github.com/d710055071", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url": "https://api.github.com/users/d710055071/gists{/gist_id}", "starred_url": "https://api.github.com/users/d710055071/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/d710055071/subscriptions", "organizations_url": "https://api.github.com/users/d710055071/orgs", "repos_url": "https://api.github.com/users/d710055071/repos", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "received_events_url": "https://api.github.com/users/d710055071/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
6
2024-01-12T09:46:17
2024-01-15T05:20:50
null
CONTRIBUTOR
null
null
null
How to do np.fromfile to use it like np.load ```python def xnumpy_fromfile(filepath_or_buffer, *args, download_config: Optional[DownloadConfig] = None, **kwargs): import numpy as np if hasattr(filepath_or_buffer, "read"): return np.fromfile(filepath_or_buffer, *args, **kwargs) else: filepath_or_buffer = str(filepath_or_buffer) return np.fromfile(xopen(filepath_or_buffer, "rb", download_config=download_config).read(), *args, **kwargs) ``` this is not work
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6584/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6584/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6580
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6580/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6580/comments
https://api.github.com/repos/huggingface/datasets/issues/6580/events
https://github.com/huggingface/datasets/issues/6580
2,075,645,042
I_kwDODunzps57t9Ry
6,580
dataset cache only stores one config of the dataset in parquet dir, and uses that for all other configs resulting in showing same data in all configs.
{ "login": "kartikgupta321", "id": 78641018, "node_id": "MDQ6VXNlcjc4NjQxMDE4", "avatar_url": "https://avatars.githubusercontent.com/u/78641018?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kartikgupta321", "html_url": "https://github.com/kartikgupta321", "followers_url": "https://api.github.com/users/kartikgupta321/followers", "following_url": "https://api.github.com/users/kartikgupta321/following{/other_user}", "gists_url": "https://api.github.com/users/kartikgupta321/gists{/gist_id}", "starred_url": "https://api.github.com/users/kartikgupta321/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kartikgupta321/subscriptions", "organizations_url": "https://api.github.com/users/kartikgupta321/orgs", "repos_url": "https://api.github.com/users/kartikgupta321/repos", "events_url": "https://api.github.com/users/kartikgupta321/events{/privacy}", "received_events_url": "https://api.github.com/users/kartikgupta321/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-11T03:14:18
2024-01-20T12:46:16
2024-01-20T12:46:16
NONE
null
null
null
### Describe the bug ds = load_dataset("ai2_arc", "ARC-Easy"), i have tried to force redownload, delete cache and changing the cache dir. ### Steps to reproduce the bug dataset = [] dataset_name = "ai2_arc" possible_configs = [ 'ARC-Challenge', 'ARC-Easy' ] for config in possible_configs: dataset_slice = load_dataset(dataset_name, config,ignore_verifications=True,cache_dir='ai2_arc_files') dataset.append(dataset_slice) ### Expected behavior all configs should get saved in cache with their respective names. ### Environment info ai2_arc
{ "login": "kartikgupta321", "id": 78641018, "node_id": "MDQ6VXNlcjc4NjQxMDE4", "avatar_url": "https://avatars.githubusercontent.com/u/78641018?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kartikgupta321", "html_url": "https://github.com/kartikgupta321", "followers_url": "https://api.github.com/users/kartikgupta321/followers", "following_url": "https://api.github.com/users/kartikgupta321/following{/other_user}", "gists_url": "https://api.github.com/users/kartikgupta321/gists{/gist_id}", "starred_url": "https://api.github.com/users/kartikgupta321/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kartikgupta321/subscriptions", "organizations_url": "https://api.github.com/users/kartikgupta321/orgs", "repos_url": "https://api.github.com/users/kartikgupta321/repos", "events_url": "https://api.github.com/users/kartikgupta321/events{/privacy}", "received_events_url": "https://api.github.com/users/kartikgupta321/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6580/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6580/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6579
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6579/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6579/comments
https://api.github.com/repos/huggingface/datasets/issues/6579/events
https://github.com/huggingface/datasets/issues/6579
2,075,407,473
I_kwDODunzps57tDRx
6,579
Unable to load `eli5` dataset with streaming
{ "login": "haok1402", "id": 89672451, "node_id": "MDQ6VXNlcjg5NjcyNDUx", "avatar_url": "https://avatars.githubusercontent.com/u/89672451?v=4", "gravatar_id": "", "url": "https://api.github.com/users/haok1402", "html_url": "https://github.com/haok1402", "followers_url": "https://api.github.com/users/haok1402/followers", "following_url": "https://api.github.com/users/haok1402/following{/other_user}", "gists_url": "https://api.github.com/users/haok1402/gists{/gist_id}", "starred_url": "https://api.github.com/users/haok1402/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/haok1402/subscriptions", "organizations_url": "https://api.github.com/users/haok1402/orgs", "repos_url": "https://api.github.com/users/haok1402/repos", "events_url": "https://api.github.com/users/haok1402/events{/privacy}", "received_events_url": "https://api.github.com/users/haok1402/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-10T23:44:20
2024-01-11T09:19:18
2024-01-11T09:19:17
NONE
null
null
null
### Describe the bug Unable to load `eli5` dataset with streaming. ### Steps to reproduce the bug This fails with FileNotFoundError: https://files.pushshift.io/reddit/submissions ``` from datasets import load_dataset load_dataset("eli5", streaming=True) ``` This works correctly. ``` from datasets import load_dataset load_dataset("eli5") ``` ### Expected behavior - Loading `eli5` dataset should not raise an error under the streaming mode. - Or at the very least, show a warning that streaming mode is not supported with `eli5` dataset. ### Environment info - `datasets` version: 2.16.1 - Platform: Linux-6.2.0-39-generic-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.19.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3 - `fsspec` version: 2023.6.0
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6579/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6579/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6577
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6577/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6577/comments
https://api.github.com/repos/huggingface/datasets/issues/6577/events
https://github.com/huggingface/datasets/issues/6577
2,074,790,848
I_kwDODunzps57qsvA
6,577
502 Server Errors when streaming large dataset
{ "login": "sanchit-gandhi", "id": 93869735, "node_id": "U_kgDOBZhWpw", "avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sanchit-gandhi", "html_url": "https://github.com/sanchit-gandhi", "followers_url": "https://api.github.com/users/sanchit-gandhi/followers", "following_url": "https://api.github.com/users/sanchit-gandhi/following{/other_user}", "gists_url": "https://api.github.com/users/sanchit-gandhi/gists{/gist_id}", "starred_url": "https://api.github.com/users/sanchit-gandhi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sanchit-gandhi/subscriptions", "organizations_url": "https://api.github.com/users/sanchit-gandhi/orgs", "repos_url": "https://api.github.com/users/sanchit-gandhi/repos", "events_url": "https://api.github.com/users/sanchit-gandhi/events{/privacy}", "received_events_url": "https://api.github.com/users/sanchit-gandhi/received_events", "type": "User", "site_admin": false }
[ { "id": 3287858981, "node_id": "MDU6TGFiZWwzMjg3ODU4OTgx", "url": "https://api.github.com/repos/huggingface/datasets/labels/streaming", "name": "streaming", "color": "fef2c0", "default": false, "description": "" } ]
closed
false
null
[]
null
6
2024-01-10T16:59:36
2024-02-12T11:46:03
2024-01-15T16:05:44
CONTRIBUTOR
null
null
null
### Describe the bug When streaming a [large ASR dataset](https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set) from the Hug (~3TB) I often encounter 502 Server Errors seemingly randomly during streaming: ``` huggingface_hub.utils._errors.HfHubHTTPError: 502 Server Error: Bad Gateway for url: https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set/resolve/7d2acc5c59de848e456e951a76e805304d6fb350/train/train-00288-of-07135.parquet ``` This is despite the parquet file definitely existing on the Hub: https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set/blob/main/train/train-00228-of-07135.parquet And having the correct commit id: [7d2acc5c59de848e456e951a76e805304d6fb350](https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set/commits/main/train) I’m wondering whether this is coming from datasets? Or from the Hub side? ### Steps to reproduce the bug Reproducer: ```python from datasets import load_dataset from torch.utils.data import DataLoader from tqdm import tqdm NUM_EPOCHS = 20 dataset = load_dataset("sanchit-gandhi/concatenated-train-set", "train", streaming=True) dataset = dataset.with_format("torch") dataloader = DataLoader(dataset["train"], batch_size=256, drop_last=True, pin_memory=True, num_workers=16) for epoch in tqdm(range(NUM_EPOCHS), desc="Epoch", position=0): for batch in tqdm(dataloader, desc="Batch", position=1): continue ``` Running the above script tends to fail within about 2 hours with a traceback like the following: <details> <summary> Traceback: </summary> ```python 1029 for batch in train_loader: 1030 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 630, in __next__ 1031 data = self._next_data() 1032 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 1325, in _next_data 1033 return self._process_data(data) 1034 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 1371, in _process_data 1035 data.reraise() 1036 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/torch/_utils.py", line 694, in reraise 1037 raise exception 1038 huggingface_hub.utils._errors.HfHubHTTPError: Caught HfHubHTTPError in DataLoader worker process 10. 1039 Original Traceback (most recent call last): 1040 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/huggingface_hub/utils/_errors.py", line 286, in hf_raise_for_status 1041 response.raise_for_status() 1042 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/requests/models.py", line 1021, in raise_for_status 1043 raise HTTPError(http_error_msg, response=self) 1044 requests.exceptions.HTTPError: 502 Server Error: Bad Gateway for url: https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set/resolve/7d2acc5c59de848e456e951a76e805304d6fb350/train/train-00288-of-07135.parquet 1045 The above exception was the direct cause of the following exception: 1046 Traceback (most recent call last): 1047 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/torch/utils/data/_utils/worker.py", line 308, in _worker_loop 1048 data = fetcher.fetch(index) 1049 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 32, in fetch 1050 data.append(next(self.dataset_iter)) 1051 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 1363, in __iter__ 1052 yield from self._iter_pytorch() 1053 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 1298, in _iter_pytorch 1054 for key, example in ex_iterable: 1055 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 983, in __iter__ 1056 for x in self.ex_iterable: 1057 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 863, in __iter__ 1058 yield from self._iter() 1059 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 900, in _iter 1060 for key, example in iterator: 1061 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 679, in __iter__ 1062 yield from self._iter() 1063 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 741, in _iter 1064 for key, example in iterator: 1065 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 863, in __iter__ 1066 yield from self._iter() 1067 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 900, in _iter 1068 for key, example in iterator: 1069 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 1115, in __iter__ 1070 for key, example in self.ex_iterable: 1071 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 679, in __iter__ 1072 yield from self._iter() 1073 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 741, in _iter 1074 for key, example in iterator: 1075 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 1115, in __iter__ 1076 for key, example in self.ex_iterable: 1077 File "/home/sanchitgandhi/datasets/src/datasets/iterable_dataset.py", line 282, in __iter__ 1078 for key, pa_table in self.generate_tables_fn(**self.kwargs): 1079 File "/home/sanchitgandhi/datasets/src/datasets/packaged_modules/parquet/parquet.py", line 87, in _generate_tables 1080 for batch_idx, record_batch in enumerate( 1081 File "pyarrow/_parquet.pyx", line 1367, in iter_batches 1082 File "pyarrow/types.pxi", line 88, in pyarrow.lib._datatype_to_pep3118 1083 File "/home/sanchitgandhi/datasets/src/datasets/download/streaming_download_manager.py", line 341, in read_with_retries 1084 out = read(*args, **kwargs) 1085 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/fsspec/spec.py", line 1856, in read 1086 out = self.cache._fetch(self.loc, self.loc + length) 1087 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/fsspec/caching.py", line 189, in _fetch 1088 self.cache = self.fetcher(start, end) # new block replaces old 1089 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/huggingface_hub/hf_file_system.py", line 626, in _fetch_range 1090 hf_raise_for_status(r) 1091 File "/home/sanchitgandhi/hf/lib/python3.8/site-packages/huggingface_hub/utils/_errors.py", line 333, in hf_raise_for_status 1092 raise HfHubHTTPError(str(e), response=response) from e 1093 huggingface_hub.utils._errors.HfHubHTTPError: 502 Server Error: Bad Gateway for url: https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set/resolve/7d2acc5c59de848e456e951a76e805304d6fb350/train/train-00288-of-07135.parquet ``` </details> ### Expected behavior Should be able to stream the dataset without any 502 error. ### Environment info - `datasets` version: 2.16.2.dev0 - Platform: Linux-5.13.0-1023-gcp-x86_64-with-glibc2.29 - Python version: 3.8.10 - `huggingface_hub` version: 0.20.1 - PyArrow version: 14.0.2 - Pandas version: 2.0.3 - `fsspec` version: 2023.10.0
{ "login": "Wauplin", "id": 11801849, "node_id": "MDQ6VXNlcjExODAxODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Wauplin", "html_url": "https://github.com/Wauplin", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "repos_url": "https://api.github.com/users/Wauplin/repos", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6577/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6577/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6576
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6576/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6576/comments
https://api.github.com/repos/huggingface/datasets/issues/6576/events
https://github.com/huggingface/datasets/issues/6576
2,073,710,124
I_kwDODunzps57mk4s
6,576
document page 404 not found after redirection
{ "login": "annahung31", "id": 39179888, "node_id": "MDQ6VXNlcjM5MTc5ODg4", "avatar_url": "https://avatars.githubusercontent.com/u/39179888?v=4", "gravatar_id": "", "url": "https://api.github.com/users/annahung31", "html_url": "https://github.com/annahung31", "followers_url": "https://api.github.com/users/annahung31/followers", "following_url": "https://api.github.com/users/annahung31/following{/other_user}", "gists_url": "https://api.github.com/users/annahung31/gists{/gist_id}", "starred_url": "https://api.github.com/users/annahung31/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/annahung31/subscriptions", "organizations_url": "https://api.github.com/users/annahung31/orgs", "repos_url": "https://api.github.com/users/annahung31/repos", "events_url": "https://api.github.com/users/annahung31/events{/privacy}", "received_events_url": "https://api.github.com/users/annahung31/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-10T06:48:14
2024-01-17T14:01:31
2024-01-17T14:01:31
NONE
null
null
null
### Describe the bug The redirected page encountered 404 not found. ### Steps to reproduce the bug 1. In this tutorial: https://huggingface.co/learn/nlp-course/chapter5/4?fw=pt original md: https://github.com/huggingface/course/blob/2c733c2246b8b7e0e6f19a9e5d15bb12df43b2a3/chapters/en/chapter5/4.mdx#L49 ``` By default, 🤗 Datasets will decompress the files needed to load a dataset. If you want to preserve hard drive space, you can pass `DownloadConfig(delete_extracted=True)` to the `download_config` argument of `load_dataset()`. See the [documentation](https://huggingface.co/docs/datasets/package_reference/builder_classes.html?#datasets.utils.DownloadConfig) for more details. ``` The documentation points to `https://huggingface.co/docs/datasets/package_reference/builder_classes.html?#datasets.utils.DownloadConfig` it shows `The documentation page PACKAGE_REFERENCE/BUILDER_CLASSES.HTML doesn’t exist in v2.16.1, but exists on the main version. Click [here](https://huggingface.co/docs/datasets/main/en/package_reference/builder_classes.html) to redirect to the main version of the documentation.` But the redirected website `https://huggingface.co/docs/datasets/main/en/package_reference/builder_classes.html` is 404 not found. ### Expected behavior I Guess the redirected webisite should be `https://huggingface.co/docs/datasets/main/en/package_reference/builder_classes` (without `.html`) or `https://huggingface.co/docs/datasets/main/en/package_reference/builder_classes#datasets.DownloadConfig`. ### Environment info Datasets main
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6576/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6576/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6571
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6571/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6571/comments
https://api.github.com/repos/huggingface/datasets/issues/6571/events
https://github.com/huggingface/datasets/issues/6571
2,072,111,000
I_kwDODunzps57geeY
6,571
Make DatasetDict.column_names return a list instead of dict
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-01-09T10:45:17
2024-01-09T10:45:17
null
MEMBER
null
null
null
Currently, `DatasetDict.column_names` returns a dict, with each split name as keys and the corresponding list of column names as values. However, by construction, all splits have the same column names. I think it makes more sense to return a single list with the column names, which is the same for all the split keys.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6571/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6571/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6570
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6570/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6570/comments
https://api.github.com/repos/huggingface/datasets/issues/6570/events
https://github.com/huggingface/datasets/issues/6570
2,071,805,265
I_kwDODunzps57fT1R
6,570
No online docs for 2.16 release
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
null
[]
null
7
2024-01-09T07:43:30
2024-01-09T16:45:50
2024-01-09T16:45:50
MEMBER
null
null
null
We do not have the online docs for the latest minor release 2.16 (2.16.0 nor 2.16.1). In the online docs, the latest version appearing is 2.15.0: https://huggingface.co/docs/datasets/index ![Screenshot from 2024-01-09 08-43-08](https://github.com/huggingface/datasets/assets/8515462/83613222-867f-41f4-8833-7a4a76582f44)
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6570/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6570/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6569
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6569/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6569/comments
https://api.github.com/repos/huggingface/datasets/issues/6569/events
https://github.com/huggingface/datasets/issues/6569
2,070,251,122
I_kwDODunzps57ZYZy
6,569
WebDataset ignores features defined in YAML or passed to load_dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false } ]
null
0
2024-01-08T11:24:21
2024-01-11T16:11:06
2024-01-11T16:11:05
MEMBER
null
null
null
we should not override if the features exist already https://github.com/huggingface/datasets/blob/d26abadce0b884db32382b92422d8a6aa997d40a/src/datasets/packaged_modules/webdataset/webdataset.py#L78-L85
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6569/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6569/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6568
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6568/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6568/comments
https://api.github.com/repos/huggingface/datasets/issues/6568/events
https://github.com/huggingface/datasets/issues/6568
2,069,922,151
I_kwDODunzps57YIFn
6,568
keep_in_memory=True does not seem to work
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
6
2024-01-08T08:03:58
2024-01-13T04:53:04
null
NONE
null
null
null
UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :(
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6568/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6568/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6567
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6567/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6567/comments
https://api.github.com/repos/huggingface/datasets/issues/6567/events
https://github.com/huggingface/datasets/issues/6567
2,069,808,842
I_kwDODunzps57XsbK
6,567
AttributeError: 'str' object has no attribute 'to'
{ "login": "andysingal", "id": 20493493, "node_id": "MDQ6VXNlcjIwNDkzNDkz", "avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4", "gravatar_id": "", "url": "https://api.github.com/users/andysingal", "html_url": "https://github.com/andysingal", "followers_url": "https://api.github.com/users/andysingal/followers", "following_url": "https://api.github.com/users/andysingal/following{/other_user}", "gists_url": "https://api.github.com/users/andysingal/gists{/gist_id}", "starred_url": "https://api.github.com/users/andysingal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andysingal/subscriptions", "organizations_url": "https://api.github.com/users/andysingal/orgs", "repos_url": "https://api.github.com/users/andysingal/repos", "events_url": "https://api.github.com/users/andysingal/events{/privacy}", "received_events_url": "https://api.github.com/users/andysingal/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
3
2024-01-08T06:40:21
2024-01-08T11:56:19
2024-01-08T10:03:17
NONE
null
null
null
### Describe the bug ``` -------------------------------------------------------------------------- AttributeError Traceback (most recent call last) [<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>() 8 report_to="wandb") 9 ---> 10 trainer = Trainer( 11 model=model, 12 args=training_args, 1 frames [/usr/local/lib/python3.10/dist-packages/transformers/trainer.py](https://localhost:8080/#) in _move_model_to_device(self, model, device) 688 689 def _move_model_to_device(self, model, device): --> 690 model = model.to(device) 691 # Moving a model to an XLA device disconnects the tied weights, so we have to retie them. 692 if self.args.parallel_mode == ParallelMode.TPU and hasattr(model, "tie_weights"): AttributeError: 'str' object has no attribute 'to' ``` ### Steps to reproduce the bug here is the notebook: ``` https://colab.research.google.com/drive/10JDBNsLlYrQdnI2FWfDK3F5M8wvVUDXG?usp=sharing ``` ### Expected behavior run the Training ### Environment info Colab Notebook , T4
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6567/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6567/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6566
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6566/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6566/comments
https://api.github.com/repos/huggingface/datasets/issues/6566/events
https://github.com/huggingface/datasets/issues/6566
2,069,495,429
I_kwDODunzps57Wf6F
6,566
I train controlnet_sdxl in bf16 datatype, got unsupported ERROR in datasets
{ "login": "HelloWorldBeginner", "id": 25008090, "node_id": "MDQ6VXNlcjI1MDA4MDkw", "avatar_url": "https://avatars.githubusercontent.com/u/25008090?v=4", "gravatar_id": "", "url": "https://api.github.com/users/HelloWorldBeginner", "html_url": "https://github.com/HelloWorldBeginner", "followers_url": "https://api.github.com/users/HelloWorldBeginner/followers", "following_url": "https://api.github.com/users/HelloWorldBeginner/following{/other_user}", "gists_url": "https://api.github.com/users/HelloWorldBeginner/gists{/gist_id}", "starred_url": "https://api.github.com/users/HelloWorldBeginner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/HelloWorldBeginner/subscriptions", "organizations_url": "https://api.github.com/users/HelloWorldBeginner/orgs", "repos_url": "https://api.github.com/users/HelloWorldBeginner/repos", "events_url": "https://api.github.com/users/HelloWorldBeginner/events{/privacy}", "received_events_url": "https://api.github.com/users/HelloWorldBeginner/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-01-08T02:37:03
2024-06-02T14:24:39
2024-05-17T09:40:14
NONE
null
null
null
### Describe the bug ``` Traceback (most recent call last): File "train_controlnet_sdxl.py", line 1252, in <module> main(args) File "train_controlnet_sdxl.py", line 1013, in main train_dataset = train_dataset.map(compute_embeddings_fn, batched=True, new_fingerprint=new_fingerprint) File "/home/miniconda3/envs/mhh_df/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 592, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/miniconda3/envs/mhh_df/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/miniconda3/envs/mhh_df/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 3093, in map for rank, done, content in Dataset._map_single(**dataset_kwargs): File "/home/miniconda3/envs/mhh_df/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 3489, in _map_single writer.write_batch(batch) File "/home/miniconda3/envs/mhh_df/lib/python3.8/site-packages/datasets/arrow_writer.py", line 557, in write_batch arrays.append(pa.array(typed_sequence)) File "pyarrow/array.pxi", line 248, in pyarrow.lib.array File "pyarrow/array.pxi", line 113, in pyarrow.lib._handle_arrow_array_protocol File "/home/miniconda3/envs/mhh_df/lib/python3.8/site-packages/datasets/arrow_writer.py", line 191, in __arrow_array__ out = pa.array(cast_to_python_objects(data, only_1d_for_numpy=True)) File "/home/miniconda3/envs/mhh_df/lib/python3.8/site-packages/datasets/features/features.py", line 447, in cast_to_python_objects return _cast_to_python_objects( File "/home/miniconda3/envs/mhh_df/lib/python3.8/site-packages/datasets/features/features.py", line 324, in _cast_to_python_objects for x in obj.detach().cpu().numpy() TypeError: Got unsupported ScalarType BFloat16 ``` ### Steps to reproduce the bug Here is my train script I use BF16 type,I use diffusers train my model ``` export MODEL_DIR="/home/mhh/sd_models/stable-diffusion-xl-base-1.0" export OUTPUT_DIR="./control_net" export VAE_NAME="/home/mhh/sd_models/sdxl-vae-fp16-fix" accelerate launch train_controlnet_sdxl.py \ --pretrained_model_name_or_path=$MODEL_DIR \ --output_dir=$OUTPUT_DIR \ --pretrained_vae_model_name_or_path=$VAE_NAME \ --dataset_name=/home/mhh/sd_datasets/fusing/fill50k \ --mixed_precision="bf16" \ --resolution=1024 \ --learning_rate=1e-5 \ --max_train_steps=200 \ --validation_image "/home/mhh/sd_datasets/controlnet_image/conditioning_image_1.png" "/home/mhh/sd_datasets/controlnet_image/conditioning_image_2.png" \ --validation_prompt "red circle with blue background" "cyan circle with brown floral background" \ --validation_steps=50 \ --train_batch_size=1 \ --gradient_accumulation_steps=4 \ --report_to="wandb" \ --seed=42 \ ``` ### Expected behavior When I changed the data type to fp16, it worked. ### Environment info datasets 2.16.1 numpy 1.24.4
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6566/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6566/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6565
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6565/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6565/comments
https://api.github.com/repos/huggingface/datasets/issues/6565/events
https://github.com/huggingface/datasets/issues/6565
2,068,939,670
I_kwDODunzps57UYOW
6,565
`drop_last_batch=True` for IterableDataset map function is ignored with multiprocessing DataLoader
{ "login": "naba89", "id": 12119806, "node_id": "MDQ6VXNlcjEyMTE5ODA2", "avatar_url": "https://avatars.githubusercontent.com/u/12119806?v=4", "gravatar_id": "", "url": "https://api.github.com/users/naba89", "html_url": "https://github.com/naba89", "followers_url": "https://api.github.com/users/naba89/followers", "following_url": "https://api.github.com/users/naba89/following{/other_user}", "gists_url": "https://api.github.com/users/naba89/gists{/gist_id}", "starred_url": "https://api.github.com/users/naba89/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/naba89/subscriptions", "organizations_url": "https://api.github.com/users/naba89/orgs", "repos_url": "https://api.github.com/users/naba89/repos", "events_url": "https://api.github.com/users/naba89/events{/privacy}", "received_events_url": "https://api.github.com/users/naba89/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-07T02:46:50
2024-01-11T16:10:31
2024-01-11T16:10:31
NONE
null
null
null
### Describe the bug Scenario: - Interleaving two iterable datasets of unequal lengths (`all_exhausted`), followed by a batch mapping with batch size 2 to effectively merge the two datasets and get a sample from each dataset in a single batch, with `drop_last_batch=True` to skip the last batch in case it doesn't have two samples. What works: - Using DataLoader with `num_workers=0` What does not work: - Using DataLoader with `num_workers=1`, errors in the last batch. Basically, `drop_last_batch=True` is ignored when using multiple dataloading workers. Please take a look at the minimal repro script below. ### Steps to reproduce the bug ```python from datasets import Dataset, interleave_datasets from torch.utils.data import DataLoader def merge_samples(batch): assert len(batch['a']) == 2, "Batch size must be 2" batch['c'] = [batch['a'][0]] batch['d'] = [batch['a'][1]] return batch def gen1(): for ii in range(1, 8385): yield {"a": ii} def gen2(): for ii in range(1, 5302): yield {"a": ii} if __name__ == '__main__': dataset1 = Dataset.from_generator(gen1).to_iterable_dataset(num_shards=1024) dataset2 = Dataset.from_generator(gen2).to_iterable_dataset(num_shards=1024) interleaved = interleave_datasets([dataset1, dataset2], stopping_strategy="all_exhausted") mapped = interleaved.map(merge_samples, batched=True, batch_size=2, remove_columns=interleaved.column_names, drop_last_batch=True) # Works loader = DataLoader(mapped, batch_size=32, num_workers=0) i = 0 for b in loader: print(i, b['c'].shape, b['d'].shape) i += 1 print("DataLoader with num_workers=0 works") # Doesn't work loader = DataLoader(mapped, batch_size=32, num_workers=1) i = 0 for b in loader: print(i, b['c'].shape, b['d'].shape) i += 1 ``` ### Expected behavior `drop_last_batch=True` should have same behaviour for `num_workers=0` and `num_workers>=1` ### Environment info - `datasets` version: 2.16.1 - Platform: macOS-10.16-x86_64-i386-64bit - Python version: 3.10.12 - `huggingface_hub` version: 0.20.2 - PyArrow version: 12.0.1 - Pandas version: 2.0.3 - `fsspec` version: 2023.6.0 I have also tested on Linux and got the same behavior.
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6565/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6565/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6564
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6564/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6564/comments
https://api.github.com/repos/huggingface/datasets/issues/6564/events
https://github.com/huggingface/datasets/issues/6564
2,068,893,194
I_kwDODunzps57UM4K
6,564
`Dataset.filter` missing `with_rank` parameter
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-06T23:48:13
2024-01-29T16:36:55
2024-01-29T16:36:54
NONE
null
null
null
### Describe the bug The issue shall be open: https://github.com/huggingface/datasets/issues/6435 When i try to pass `with_rank` to `Dataset.filter()`, i get this: `Dataset.filter() got an unexpected keyword argument 'with_rank'` ### Steps to reproduce the bug Run notebook: https://colab.research.google.com/drive/1WUNKph8BdP0on5ve3gQnh_PE0cFLQqTn?usp=sharing ### Expected behavior Should work? ### Environment info NVIDIA RTX 4090
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6564/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6564/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6563
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6563/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6563/comments
https://api.github.com/repos/huggingface/datasets/issues/6563/events
https://github.com/huggingface/datasets/issues/6563
2,068,302,402
I_kwDODunzps57R8pC
6,563
`ImportError`: cannot import name 'insecure_hashlib' from 'huggingface_hub.utils' (.../huggingface_hub/utils/__init__.py)
{ "login": "wasertech", "id": 79070834, "node_id": "MDQ6VXNlcjc5MDcwODM0", "avatar_url": "https://avatars.githubusercontent.com/u/79070834?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wasertech", "html_url": "https://github.com/wasertech", "followers_url": "https://api.github.com/users/wasertech/followers", "following_url": "https://api.github.com/users/wasertech/following{/other_user}", "gists_url": "https://api.github.com/users/wasertech/gists{/gist_id}", "starred_url": "https://api.github.com/users/wasertech/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wasertech/subscriptions", "organizations_url": "https://api.github.com/users/wasertech/orgs", "repos_url": "https://api.github.com/users/wasertech/repos", "events_url": "https://api.github.com/users/wasertech/events{/privacy}", "received_events_url": "https://api.github.com/users/wasertech/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
7
2024-01-06T02:28:54
2024-03-14T02:59:42
2024-01-06T16:13:27
NONE
null
null
null
### Describe the bug Yep its not [there](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/utils/__init__.py) anymore. ```text + python /home/trainer/sft_train.py --model_name cognitivecomputations/dolphin-2.2.1-mistral-7b --dataset_name wasertech/OneOS --load_in_4bit --use_peft --batch_size 4 --num_train_epochs 1 --learning_rate 1.41e-5 --gradient_accumulation_steps 8 --seq_length 4096 --output_dir output --log_with wandb Traceback (most recent call last): File "/home/trainer/sft_train.py", line 22, in <module> from datasets import load_dataset File "/home/trainer/llm-train/lib/python3.8/site-packages/datasets/__init__.py", line 22, in <module> from .arrow_dataset import Dataset File "/home/trainer/llm-train/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 66, in <module> from .arrow_reader import ArrowReader File "/home/trainer/llm-train/lib/python3.8/site-packages/datasets/arrow_reader.py", line 30, in <module> from .download.download_config import DownloadConfig File "/home/trainer/llm-train/lib/python3.8/site-packages/datasets/download/__init__.py", line 9, in <module> from .download_manager import DownloadManager, DownloadMode File "/home/trainer/llm-train/lib/python3.8/site-packages/datasets/download/download_manager.py", line 31, in <module> from ..utils import tqdm as hf_tqdm File "/home/trainer/llm-train/lib/python3.8/site-packages/datasets/utils/__init__.py", line 19, in <module> from .info_utils import VerificationMode File "/home/trainer/llm-train/lib/python3.8/site-packages/datasets/utils/info_utils.py", line 5, in <module> from huggingface_hub.utils import insecure_hashlib ImportError: cannot import name 'insecure_hashlib' from 'huggingface_hub.utils' (/home/trainer/llm-train/lib/python3.8/site-packages/huggingface_hub/utils/__init__.py) ``` ### Steps to reproduce the bug Using `datasets==2.16.1` and `huggingface_hub== 0.17.3`, load a dataset with `load_dataset`. ### Expected behavior The dataset should be (downloaded - if needed - and) returned. ### Environment info ```text trainer@a311ae86939e:/mnt$ pip show datasets Name: datasets Version: 2.16.1 Summary: HuggingFace community-driven open-source library of datasets Home-page: https://github.com/huggingface/datasets Author: HuggingFace Inc. Author-email: thomas@huggingface.co License: Apache 2.0 Location: /home/trainer/llm-train/lib/python3.8/site-packages Requires: packaging, pyyaml, multiprocess, pyarrow-hotfix, pandas, pyarrow, xxhash, dill, numpy, aiohttp, tqdm, fsspec, requests, filelock, huggingface-hub Required-by: trl, lm-eval, evaluate trainer@a311ae86939e:/mnt$ pip show huggingface_hub Name: huggingface-hub Version: 0.17.3 Summary: Client library to download and publish models, datasets and other repos on the huggingface.co hub Home-page: https://github.com/huggingface/huggingface_hub Author: Hugging Face, Inc. Author-email: julien@huggingface.co License: Apache Location: /home/trainer/llm-train/lib/python3.8/site-packages Requires: requests, pyyaml, packaging, typing-extensions, tqdm, filelock, fsspec Required-by: transformers, tokenizers, peft, evaluate, datasets, accelerate trainer@a311ae86939e:/mnt$ huggingface-cli env Copy-and-paste the text below in your GitHub issue. - huggingface_hub version: 0.17.3 - Platform: Linux-6.5.13-7-MANJARO-x86_64-with-glibc2.29 - Python version: 3.8.10 - Running in iPython ?: No - Running in notebook ?: No - Running in Google Colab ?: No - Token path ?: /home/trainer/.cache/huggingface/token - Has saved token ?: True - Who am I ?: wasertech - Configured git credential helpers: - FastAI: N/A - Tensorflow: N/A - Torch: 2.1.2 - Jinja2: 3.1.2 - Graphviz: N/A - Pydot: N/A - Pillow: 10.2.0 - hf_transfer: N/A - gradio: N/A - tensorboard: N/A - numpy: 1.24.4 - pydantic: N/A - aiohttp: 3.9.1 - ENDPOINT: https://huggingface.co - HUGGINGFACE_HUB_CACHE: /home/trainer/.cache/huggingface/hub - HUGGINGFACE_ASSETS_CACHE: /home/trainer/.cache/huggingface/assets - HF_TOKEN_PATH: /home/trainer/.cache/huggingface/token - HF_HUB_OFFLINE: False - HF_HUB_DISABLE_TELEMETRY: False - HF_HUB_DISABLE_PROGRESS_BARS: None - HF_HUB_DISABLE_SYMLINKS_WARNING: False - HF_HUB_DISABLE_EXPERIMENTAL_WARNING: False - HF_HUB_DISABLE_IMPLICIT_TOKEN: False - HF_HUB_ENABLE_HF_TRANSFER: False ```
{ "login": "wasertech", "id": 79070834, "node_id": "MDQ6VXNlcjc5MDcwODM0", "avatar_url": "https://avatars.githubusercontent.com/u/79070834?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wasertech", "html_url": "https://github.com/wasertech", "followers_url": "https://api.github.com/users/wasertech/followers", "following_url": "https://api.github.com/users/wasertech/following{/other_user}", "gists_url": "https://api.github.com/users/wasertech/gists{/gist_id}", "starred_url": "https://api.github.com/users/wasertech/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wasertech/subscriptions", "organizations_url": "https://api.github.com/users/wasertech/orgs", "repos_url": "https://api.github.com/users/wasertech/repos", "events_url": "https://api.github.com/users/wasertech/events{/privacy}", "received_events_url": "https://api.github.com/users/wasertech/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6563/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6563/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6562
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6562/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6562/comments
https://api.github.com/repos/huggingface/datasets/issues/6562/events
https://github.com/huggingface/datasets/issues/6562
2,067,904,504
I_kwDODunzps57Qbf4
6,562
datasets.DownloadMode.FORCE_REDOWNLOAD use cache to download dataset features with load_dataset function
{ "login": "LsTam91", "id": 73234162, "node_id": "MDQ6VXNlcjczMjM0MTYy", "avatar_url": "https://avatars.githubusercontent.com/u/73234162?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LsTam91", "html_url": "https://github.com/LsTam91", "followers_url": "https://api.github.com/users/LsTam91/followers", "following_url": "https://api.github.com/users/LsTam91/following{/other_user}", "gists_url": "https://api.github.com/users/LsTam91/gists{/gist_id}", "starred_url": "https://api.github.com/users/LsTam91/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LsTam91/subscriptions", "organizations_url": "https://api.github.com/users/LsTam91/orgs", "repos_url": "https://api.github.com/users/LsTam91/repos", "events_url": "https://api.github.com/users/LsTam91/events{/privacy}", "received_events_url": "https://api.github.com/users/LsTam91/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2024-01-05T19:10:25
2024-01-05T19:10:25
null
NONE
null
null
null
### Describe the bug I have updated my dataset by adding a new feature, and push it to the hub. When I want to download it on my machine which contain the old version by using `datasets.load_dataset("your_dataset_name", download_mode=datasets.DownloadMode.FORCE_REDOWNLOAD)` I get an error (paste bellow). Seems that the load_dataset function still use the old features schema instead of downloading everything new from the HUB. I find a way to go around this issue by manually deleting the old dataset cache. But from my understanding of `datasets.DownloadMode.FORCE_REDOWNLOAD` option, the dataset cache should be ignored. ### Steps to reproduce the bug 1. Download your dataset in your machine using `datasets.load_dataset` 2. Create a new feature in your dataset and push it to the hub 3. On the same machine redownload your dataset using `datasets.load_dataset("your_dataset_name", download_mode=datasets.DownloadMode.FORCE_REDOWNLOAD)` ### Expected behavior ` ValueError: Couldn't cast id: string level: string context: list<element: string> child 0, element: string type: string answer: string question: string supporting_facts: list<element: string> child 0, element: string fra_answer: string fra_question: string -- schema metadata -- huggingface: '{"info": {"features": {"id": {"dtype": "string", "_type": "' + 490 to {'id': Value(dtype='string', id=None), 'level': Value(dtype='string', id=None), 'context': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'type': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'question': Value(dtype='string', id=None), 'supporting_facts': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)} because column names don't match The above exception was the direct cause of the following exception: DatasetGenerationError ... DatasetGenerationError: An error occurred while generating the dataset` ### Environment info datasets-2.16.1 huggingface-hub-0.20.2
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6562/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6562/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6561
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6561/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6561/comments
https://api.github.com/repos/huggingface/datasets/issues/6561/events
https://github.com/huggingface/datasets/issues/6561
2,067,404,951
I_kwDODunzps57OhiX
6,561
Document YAML configuration with "data_dir"
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "organizations_url": "https://api.github.com/users/severo/orgs", "repos_url": "https://api.github.com/users/severo/repos", "events_url": "https://api.github.com/users/severo/events{/privacy}", "received_events_url": "https://api.github.com/users/severo/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
open
false
null
[]
null
1
2024-01-05T14:03:33
2024-01-05T14:06:18
null
CONTRIBUTOR
null
null
null
See https://huggingface.co/datasets/uonlp/CulturaX/discussions/15#6597e83f185db94370d6bf50 for reference
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6561/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6561/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6560
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6560/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6560/comments
https://api.github.com/repos/huggingface/datasets/issues/6560/events
https://github.com/huggingface/datasets/issues/6560
2,065,637,625
I_kwDODunzps57HyD5
6,560
Support Video
{ "login": "yuvalkirstain", "id": 57996478, "node_id": "MDQ6VXNlcjU3OTk2NDc4", "avatar_url": "https://avatars.githubusercontent.com/u/57996478?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yuvalkirstain", "html_url": "https://github.com/yuvalkirstain", "followers_url": "https://api.github.com/users/yuvalkirstain/followers", "following_url": "https://api.github.com/users/yuvalkirstain/following{/other_user}", "gists_url": "https://api.github.com/users/yuvalkirstain/gists{/gist_id}", "starred_url": "https://api.github.com/users/yuvalkirstain/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yuvalkirstain/subscriptions", "organizations_url": "https://api.github.com/users/yuvalkirstain/orgs", "repos_url": "https://api.github.com/users/yuvalkirstain/repos", "events_url": "https://api.github.com/users/yuvalkirstain/events{/privacy}", "received_events_url": "https://api.github.com/users/yuvalkirstain/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892865, "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate", "name": "duplicate", "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists" }, { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-01-04T13:10:58
2024-08-23T09:51:27
2024-08-23T09:51:27
NONE
null
null
null
### Feature request HF datasets are awesome in supporting text and images. Will be great to see such a support in videos :) ### Motivation Video generation :) ### Your contribution Will probably be limited to raising this feature request ;)
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "organizations_url": "https://api.github.com/users/severo/orgs", "repos_url": "https://api.github.com/users/severo/repos", "events_url": "https://api.github.com/users/severo/events{/privacy}", "received_events_url": "https://api.github.com/users/severo/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6560/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6560/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6559
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6559/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6559/comments
https://api.github.com/repos/huggingface/datasets/issues/6559/events
https://github.com/huggingface/datasets/issues/6559
2,065,118,332
I_kwDODunzps57FzR8
6,559
Latest version 2.16.1, when load dataset error occurs. ValueError: BuilderConfig 'allenai--c4' not found. Available: ['default']
{ "login": "zhulinJulia24", "id": 145004780, "node_id": "U_kgDOCKSY7A", "avatar_url": "https://avatars.githubusercontent.com/u/145004780?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zhulinJulia24", "html_url": "https://github.com/zhulinJulia24", "followers_url": "https://api.github.com/users/zhulinJulia24/followers", "following_url": "https://api.github.com/users/zhulinJulia24/following{/other_user}", "gists_url": "https://api.github.com/users/zhulinJulia24/gists{/gist_id}", "starred_url": "https://api.github.com/users/zhulinJulia24/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zhulinJulia24/subscriptions", "organizations_url": "https://api.github.com/users/zhulinJulia24/orgs", "repos_url": "https://api.github.com/users/zhulinJulia24/repos", "events_url": "https://api.github.com/users/zhulinJulia24/events{/privacy}", "received_events_url": "https://api.github.com/users/zhulinJulia24/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
8
2024-01-04T07:04:48
2024-04-03T10:40:53
2024-01-05T01:26:25
NONE
null
null
null
### Describe the bug python script is: ``` from datasets import load_dataset cache_dir = 'path/to/your/cache/directory' dataset = load_dataset('allenai/c4','allenai--c4', data_files={'train': 'en/c4-train.00000-of-01024.json.gz'}, split='train', use_auth_token=False, cache_dir=cache_dir) ``` the script success when datasets version is 2.14.7. when using 2.16.1, error occurs ` ValueError: BuilderConfig 'allenai--c4' not found. Available: ['default']` ### Steps to reproduce the bug 1. pip install datasets==2.16.1 2. run python script: ``` from datasets import load_dataset cache_dir = 'path/to/your/cache/directory' dataset = load_dataset('allenai/c4','allenai--c4', data_files={'train': 'en/c4-train.00000-of-01024.json.gz'}, split='train', use_auth_token=False, cache_dir=cache_dir) ``` ### Expected behavior the dataset should be loaded successful in the latest version. ### Environment info datasets 2.16.1
{ "login": "zhulinJulia24", "id": 145004780, "node_id": "U_kgDOCKSY7A", "avatar_url": "https://avatars.githubusercontent.com/u/145004780?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zhulinJulia24", "html_url": "https://github.com/zhulinJulia24", "followers_url": "https://api.github.com/users/zhulinJulia24/followers", "following_url": "https://api.github.com/users/zhulinJulia24/following{/other_user}", "gists_url": "https://api.github.com/users/zhulinJulia24/gists{/gist_id}", "starred_url": "https://api.github.com/users/zhulinJulia24/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zhulinJulia24/subscriptions", "organizations_url": "https://api.github.com/users/zhulinJulia24/orgs", "repos_url": "https://api.github.com/users/zhulinJulia24/repos", "events_url": "https://api.github.com/users/zhulinJulia24/events{/privacy}", "received_events_url": "https://api.github.com/users/zhulinJulia24/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6559/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6559/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6558
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6558/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6558/comments
https://api.github.com/repos/huggingface/datasets/issues/6558/events
https://github.com/huggingface/datasets/issues/6558
2,064,885,984
I_kwDODunzps57E6jg
6,558
OSError: image file is truncated (1 bytes not processed) #28323
{ "login": "andysingal", "id": 20493493, "node_id": "MDQ6VXNlcjIwNDkzNDkz", "avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4", "gravatar_id": "", "url": "https://api.github.com/users/andysingal", "html_url": "https://github.com/andysingal", "followers_url": "https://api.github.com/users/andysingal/followers", "following_url": "https://api.github.com/users/andysingal/following{/other_user}", "gists_url": "https://api.github.com/users/andysingal/gists{/gist_id}", "starred_url": "https://api.github.com/users/andysingal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andysingal/subscriptions", "organizations_url": "https://api.github.com/users/andysingal/orgs", "repos_url": "https://api.github.com/users/andysingal/repos", "events_url": "https://api.github.com/users/andysingal/events{/privacy}", "received_events_url": "https://api.github.com/users/andysingal/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-04T02:15:13
2024-02-21T00:38:12
2024-02-21T00:38:12
NONE
null
null
null
### Describe the bug ``` --------------------------------------------------------------------------- OSError Traceback (most recent call last) Cell In[24], line 28 23 return example 25 # Filter the dataset 26 # filtered_dataset = dataset.filter(contains_number) 27 # Add the 'label' field in the dataset ---> 28 labeled_dataset = dataset.filter(contains_number).map(add_label) 29 # View the structure of the updated dataset 30 print(labeled_dataset) File /usr/local/lib/python3.10/dist-packages/datasets/dataset_dict.py:975, in DatasetDict.filter(self, function, with_indices, input_columns, batched, batch_size, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, fn_kwargs, num_proc, desc) 972 if cache_file_names is None: 973 cache_file_names = {k: None for k in self} 974 return DatasetDict( --> 975 { 976 k: dataset.filter( 977 function=function, 978 with_indices=with_indices, 979 input_columns=input_columns, 980 batched=batched, 981 batch_size=batch_size, 982 keep_in_memory=keep_in_memory, 983 load_from_cache_file=load_from_cache_file, 984 cache_file_name=cache_file_names[k], 985 writer_batch_size=writer_batch_size, 986 fn_kwargs=fn_kwargs, 987 num_proc=num_proc, 988 desc=desc, 989 ) 990 for k, dataset in self.items() 991 } 992 ) File /usr/local/lib/python3.10/dist-packages/datasets/dataset_dict.py:976, in <dictcomp>(.0) 972 if cache_file_names is None: 973 cache_file_names = {k: None for k in self} 974 return DatasetDict( 975 { --> 976 k: dataset.filter( 977 function=function, 978 with_indices=with_indices, 979 input_columns=input_columns, 980 batched=batched, 981 batch_size=batch_size, 982 keep_in_memory=keep_in_memory, 983 load_from_cache_file=load_from_cache_file, 984 cache_file_name=cache_file_names[k], 985 writer_batch_size=writer_batch_size, 986 fn_kwargs=fn_kwargs, 987 num_proc=num_proc, 988 desc=desc, 989 ) 990 for k, dataset in self.items() 991 } 992 ) File /usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py:557, in transmit_format.<locals>.wrapper(*args, **kwargs) 550 self_format = { 551 "type": self._format_type, 552 "format_kwargs": self._format_kwargs, 553 "columns": self._format_columns, 554 "output_all_columns": self._output_all_columns, 555 } 556 # apply actual function --> 557 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 558 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 559 # re-apply format to the output File /usr/local/lib/python3.10/dist-packages/datasets/fingerprint.py:481, in fingerprint_transform.<locals>._fingerprint.<locals>.wrapper(*args, **kwargs) 477 validate_fingerprint(kwargs[fingerprint_name]) 479 # Call actual function --> 481 out = func(dataset, *args, **kwargs) 483 # Update fingerprint of in-place transforms + update in-place history of transforms 485 if inplace: # update after calling func so that the fingerprint doesn't change if the function fails File /usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py:3623, in Dataset.filter(self, function, with_indices, input_columns, batched, batch_size, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 3620 if len(self) == 0: 3621 return self -> 3623 indices = self.map( 3624 function=partial( 3625 get_indices_from_mask_function, function, batched, with_indices, input_columns, self._indices 3626 ), 3627 with_indices=True, 3628 features=Features({"indices": Value("uint64")}), 3629 batched=True, 3630 batch_size=batch_size, 3631 remove_columns=self.column_names, 3632 keep_in_memory=keep_in_memory, 3633 load_from_cache_file=load_from_cache_file, 3634 cache_file_name=cache_file_name, 3635 writer_batch_size=writer_batch_size, 3636 fn_kwargs=fn_kwargs, 3637 num_proc=num_proc, 3638 suffix_template=suffix_template, 3639 new_fingerprint=new_fingerprint, 3640 input_columns=input_columns, 3641 desc=desc or "Filter", 3642 ) 3643 new_dataset = copy.deepcopy(self) 3644 new_dataset._indices = indices.data File /usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py:592, in transmit_tasks.<locals>.wrapper(*args, **kwargs) 590 self: "Dataset" = kwargs.pop("self") 591 # apply actual function --> 592 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 593 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 594 for dataset in datasets: 595 # Remove task templates if a column mapping of the template is no longer valid File /usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py:557, in transmit_format.<locals>.wrapper(*args, **kwargs) 550 self_format = { 551 "type": self._format_type, 552 "format_kwargs": self._format_kwargs, 553 "columns": self._format_columns, 554 "output_all_columns": self._output_all_columns, 555 } 556 # apply actual function --> 557 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 558 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 559 # re-apply format to the output File /usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py:3093, in Dataset.map(self, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 3087 if transformed_dataset is None: 3088 with hf_tqdm( 3089 unit=" examples", 3090 total=pbar_total, 3091 desc=desc or "Map", 3092 ) as pbar: -> 3093 for rank, done, content in Dataset._map_single(**dataset_kwargs): 3094 if done: 3095 shards_done += 1 File /usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py:3470, in Dataset._map_single(shard, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset) 3466 indices = list( 3467 range(*(slice(i, i + batch_size).indices(shard.num_rows))) 3468 ) # Something simpler? 3469 try: -> 3470 batch = apply_function_on_filtered_inputs( 3471 batch, 3472 indices, 3473 check_same_num_examples=len(shard.list_indexes()) > 0, 3474 offset=offset, 3475 ) 3476 except NumExamplesMismatchError: 3477 raise DatasetTransformationNotAllowedError( 3478 "Using `.map` in batched mode on a dataset with attached indexes is allowed only if it doesn't create or remove existing examples. You can first run `.drop_index() to remove your index and then re-add it." 3479 ) from None File /usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py:3349, in Dataset._map_single.<locals>.apply_function_on_filtered_inputs(pa_inputs, indices, check_same_num_examples, offset) 3347 if with_rank: 3348 additional_args += (rank,) -> 3349 processed_inputs = function(*fn_args, *additional_args, **fn_kwargs) 3350 if isinstance(processed_inputs, LazyDict): 3351 processed_inputs = { 3352 k: v for k, v in processed_inputs.data.items() if k not in processed_inputs.keys_to_format 3353 } File /usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py:6212, in get_indices_from_mask_function(function, batched, with_indices, input_columns, indices_mapping, *args, **fn_kwargs) 6209 if input_columns is None: 6210 # inputs only contains a batch of examples 6211 batch: dict = inputs[0] -> 6212 num_examples = len(batch[next(iter(batch.keys()))]) 6213 for i in range(num_examples): 6214 example = {key: batch[key][i] for key in batch} File /usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py:272, in LazyDict.__getitem__(self, key) 270 value = self.data[key] 271 if key in self.keys_to_format: --> 272 value = self.format(key) 273 self.data[key] = value 274 self.keys_to_format.remove(key) File /usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py:375, in LazyBatch.format(self, key) 374 def format(self, key): --> 375 return self.formatter.format_column(self.pa_table.select([key])) File /usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py:442, in PythonFormatter.format_column(self, pa_table) 440 def format_column(self, pa_table: pa.Table) -> list: 441 column = self.python_arrow_extractor().extract_column(pa_table) --> 442 column = self.python_features_decoder.decode_column(column, pa_table.column_names[0]) 443 return column File /usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py:218, in PythonFeaturesDecoder.decode_column(self, column, column_name) 217 def decode_column(self, column: list, column_name: str) -> list: --> 218 return self.features.decode_column(column, column_name) if self.features else column File /usr/local/lib/python3.10/dist-packages/datasets/features/features.py:1951, in Features.decode_column(self, column, column_name) 1938 def decode_column(self, column: list, column_name: str): 1939 """Decode column with custom feature decoding. 1940 1941 Args: (...) 1948 `list[Any]` 1949 """ 1950 return ( -> 1951 [decode_nested_example(self[column_name], value) if value is not None else None for value in column] 1952 if self._column_requires_decoding[column_name] 1953 else column 1954 ) File /usr/local/lib/python3.10/dist-packages/datasets/features/features.py:1951, in <listcomp>(.0) 1938 def decode_column(self, column: list, column_name: str): 1939 """Decode column with custom feature decoding. 1940 1941 Args: (...) 1948 `list[Any]` 1949 """ 1950 return ( -> 1951 [decode_nested_example(self[column_name], value) if value is not None else None for value in column] 1952 if self._column_requires_decoding[column_name] 1953 else column 1954 ) File /usr/local/lib/python3.10/dist-packages/datasets/features/features.py:1339, in decode_nested_example(schema, obj, token_per_repo_id) 1336 elif isinstance(schema, (Audio, Image)): 1337 # we pass the token to read and decode files from private repositories in streaming mode 1338 if obj is not None and schema.decode: -> 1339 return schema.decode_example(obj, token_per_repo_id=token_per_repo_id) 1340 return obj File /usr/local/lib/python3.10/dist-packages/datasets/features/image.py:185, in Image.decode_example(self, value, token_per_repo_id) 183 else: 184 image = PIL.Image.open(BytesIO(bytes_)) --> 185 image.load() # to avoid "Too many open files" errors 186 return image File /usr/local/lib/python3.10/dist-packages/PIL/ImageFile.py:254, in ImageFile.load(self) 252 break 253 else: --> 254 raise OSError( 255 "image file is truncated " 256 f"({len(b)} bytes not processed)" 257 ) 259 b = b + s 260 n, err_code = decoder.decode(b) OSError: image file is truncated (1 bytes not processed) ``` ### Steps to reproduce the bug ``` from datasets import load_dataset dataset = load_dataset("mehul7/captioned_military_aircraft") from transformers import AutoImageProcessor checkpoint = "microsoft/resnet-50" image_processor = AutoImageProcessor.from_pretrained(checkpoint) import re from PIL import Image import io def contains_number(example): try: image = Image.open(io.BytesIO(example["image"]['bytes'])) t = image_processor(images=image, return_tensors="pt")['pixel_values'] except Exception as e: print(f"Error processing image:{example['text']}") return False return bool(re.search(r'\d', example['text'])) # Define a function to add the 'label' field def add_label(example): lab = example['text'].split() temp = 'NOT' for item in lab: if str(item[-1]).isdigit(): temp = item break example['label'] = temp return example # Filter the dataset # filtered_dataset = dataset.filter(contains_number) # Add the 'label' field in the dataset labeled_dataset = dataset.filter(contains_number).map(add_label) # View the structure of the updated dataset print(labeled_dataset) ``` ### Expected behavior needs to form labels same as : https://www.kaggle.com/code/jiabaowangts/dataset-air/notebook ### Environment info Kaggle notebook P100
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6558/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6558/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6554
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6554/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6554/comments
https://api.github.com/repos/huggingface/datasets/issues/6554/events
https://github.com/huggingface/datasets/issues/6554
2,063,839,916
I_kwDODunzps57A7Ks
6,554
Parquet exports are used even if revision is passed
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
1
2024-01-03T11:32:26
2024-02-02T10:35:29
2024-02-02T10:35:29
MEMBER
null
null
null
We should not used Parquet exports if `revision` is passed. I think this is a regression.
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6554/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6554/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6553
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6553/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6553/comments
https://api.github.com/repos/huggingface/datasets/issues/6553/events
https://github.com/huggingface/datasets/issues/6553
2,063,474,183
I_kwDODunzps56_h4H
6,553
Cannot import name 'load_dataset' from .... module ‘datasets’
{ "login": "ciaoyizhen", "id": 83450192, "node_id": "MDQ6VXNlcjgzNDUwMTky", "avatar_url": "https://avatars.githubusercontent.com/u/83450192?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ciaoyizhen", "html_url": "https://github.com/ciaoyizhen", "followers_url": "https://api.github.com/users/ciaoyizhen/followers", "following_url": "https://api.github.com/users/ciaoyizhen/following{/other_user}", "gists_url": "https://api.github.com/users/ciaoyizhen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ciaoyizhen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ciaoyizhen/subscriptions", "organizations_url": "https://api.github.com/users/ciaoyizhen/orgs", "repos_url": "https://api.github.com/users/ciaoyizhen/repos", "events_url": "https://api.github.com/users/ciaoyizhen/events{/privacy}", "received_events_url": "https://api.github.com/users/ciaoyizhen/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-03T08:18:21
2024-02-21T00:38:24
2024-02-21T00:38:24
NONE
null
null
null
### Describe the bug use python -m pip install datasets to install ### Steps to reproduce the bug from datasets import load_dataset ### Expected behavior it doesn't work ### Environment info datasets version==2.15.0 python == 3.10.12 linux version I don't know??
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6553/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6553/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6552
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6552/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6552/comments
https://api.github.com/repos/huggingface/datasets/issues/6552/events
https://github.com/huggingface/datasets/issues/6552
2,063,157,187
I_kwDODunzps56-UfD
6,552
Loading a dataset from Google Colab hangs at "Resolving data files".
{ "login": "KelSolaar", "id": 99779, "node_id": "MDQ6VXNlcjk5Nzc5", "avatar_url": "https://avatars.githubusercontent.com/u/99779?v=4", "gravatar_id": "", "url": "https://api.github.com/users/KelSolaar", "html_url": "https://github.com/KelSolaar", "followers_url": "https://api.github.com/users/KelSolaar/followers", "following_url": "https://api.github.com/users/KelSolaar/following{/other_user}", "gists_url": "https://api.github.com/users/KelSolaar/gists{/gist_id}", "starred_url": "https://api.github.com/users/KelSolaar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/KelSolaar/subscriptions", "organizations_url": "https://api.github.com/users/KelSolaar/orgs", "repos_url": "https://api.github.com/users/KelSolaar/repos", "events_url": "https://api.github.com/users/KelSolaar/events{/privacy}", "received_events_url": "https://api.github.com/users/KelSolaar/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-03T02:18:17
2024-01-08T10:09:04
2024-01-08T10:09:04
NONE
null
null
null
### Describe the bug Hello, I'm trying to load a dataset from Google Colab but the process hangs at `Resolving data files`: ![image](https://github.com/huggingface/datasets/assets/99779/7175ad85-e571-46ed-9f87-92653985777d) It is happening when the `_get_origin_metadata` definition is invoked: ```python def _get_origin_metadata( data_files: List[str], max_workers=64, download_config: Optional[DownloadConfig] = None, ) -> Tuple[str]: return thread_map( partial(_get_single_origin_metadata, download_config=download_config), data_files, max_workers=max_workers, tqdm_class=hf_tqdm, desc="Resolving data files", disable=len(data_files) <= 16, ``` The thread is then stuck at `waiter.acquire()` in the builtin `threading.py` file. I can load the dataset just fine on my machine. Cheers, Thomas ### Steps to reproduce the bug In Google Colab: ```python !pip install datasets from datasets import load_dataset dataset = load_dataset("colour-science/color-checker-detection-dataset") ``` ### Expected behavior The dataset should be loaded. ### Environment info - `datasets` version: 2.16.1 - Platform: Linux-6.1.58+-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.20.1 - PyArrow version: 10.0.1 - Pandas version: 1.5.3 - `fsspec` version: 2023.6.0
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6552/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6552/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6549
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6549/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6549/comments
https://api.github.com/repos/huggingface/datasets/issues/6549/events
https://github.com/huggingface/datasets/issues/6549
2,062,420,259
I_kwDODunzps567gkj
6,549
Loading from hf hub with clearer error message
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "https://api.github.com/users/thomwolf/gists{/gist_id}", "starred_url": "https://api.github.com/users/thomwolf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomwolf/subscriptions", "organizations_url": "https://api.github.com/users/thomwolf/orgs", "repos_url": "https://api.github.com/users/thomwolf/repos", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "received_events_url": "https://api.github.com/users/thomwolf/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
1
2024-01-02T13:26:34
2024-01-02T14:06:49
null
MEMBER
null
null
null
### Feature request Shouldn't this kinda work ? ``` Dataset.from_json("hf://datasets/HuggingFaceTB/eval_data/resolve/main/eval_data_context_and_answers.json") ``` I got an error ``` File ~/miniconda3/envs/datatrove/lib/python3.10/site-packages/datasets/data_files.py:380, in resolve_pattern(pattern, base_path, allowed_extensions, download_config) 378 if allowed_extensions is not None: 379 error_msg += f" with any supported extension {list(allowed_extensions)}" --> 380 raise FileNotFoundError(error_msg) 381 return out FileNotFoundError: Unable to find 'hf://datasets/HuggingFaceTB/eval_data/resolve/main/eval_data_context_and_answers.json' (I'm logged in) ``` Fix: the correct path is ``` hf://datasets/HuggingFaceTB/eval_data/eval_data_context_and_answers.json ``` Proposal: raise a clearer error ### Motivation Clearer error message ### Your contribution Can open a PR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6549/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6549/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6548
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6548/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6548/comments
https://api.github.com/repos/huggingface/datasets/issues/6548/events
https://github.com/huggingface/datasets/issues/6548
2,061,047,984
I_kwDODunzps562Riw
6,548
Skip if a dataset has issues
{ "login": "hadianasliwa", "id": 143214684, "node_id": "U_kgDOCIlIXA", "avatar_url": "https://avatars.githubusercontent.com/u/143214684?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hadianasliwa", "html_url": "https://github.com/hadianasliwa", "followers_url": "https://api.github.com/users/hadianasliwa/followers", "following_url": "https://api.github.com/users/hadianasliwa/following{/other_user}", "gists_url": "https://api.github.com/users/hadianasliwa/gists{/gist_id}", "starred_url": "https://api.github.com/users/hadianasliwa/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hadianasliwa/subscriptions", "organizations_url": "https://api.github.com/users/hadianasliwa/orgs", "repos_url": "https://api.github.com/users/hadianasliwa/repos", "events_url": "https://api.github.com/users/hadianasliwa/events{/privacy}", "received_events_url": "https://api.github.com/users/hadianasliwa/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
1
2023-12-31T12:41:26
2024-01-02T10:33:17
null
NONE
null
null
null
### Describe the bug Hello everyone, I'm using **load_datasets** from **huggingface** to download the datasets and I'm facing an issue, the download starts but it reaches some state and then fails with the following error: Couldn't reach https://huggingface.co/datasets/wikimedia/wikipedia/resolve/4cb9b0d719291f1a10f96f67d609c5d442980dc9/20231101.ext/train-00000-of-00001.parquet Failed to resolve \'huggingface.co\' ([Errno -3] Temporary failure in name resolution)"))'))) ![image](https://github.com/huggingface/datasets/assets/143214684/8847d9cb-529e-4eda-9c76-282713dfa3af) so I was wondering is there a parameter to be passed to load_dataset() to skip files that can't be downloaded?? ### Steps to reproduce the bug Parameter to be passed to load_dataset() of huggingface to skip files that can't be downloaded?? ### Expected behavior load_dataset() finishes without error ### Environment info None
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6548/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6548/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6545
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6545/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6545/comments
https://api.github.com/repos/huggingface/datasets/issues/6545/events
https://github.com/huggingface/datasets/issues/6545
2,060,789,507
I_kwDODunzps561ScD
6,545
`image` column not automatically inferred if image dataset only contains 1 image
{ "login": "apolinario", "id": 788417, "node_id": "MDQ6VXNlcjc4ODQxNw==", "avatar_url": "https://avatars.githubusercontent.com/u/788417?v=4", "gravatar_id": "", "url": "https://api.github.com/users/apolinario", "html_url": "https://github.com/apolinario", "followers_url": "https://api.github.com/users/apolinario/followers", "following_url": "https://api.github.com/users/apolinario/following{/other_user}", "gists_url": "https://api.github.com/users/apolinario/gists{/gist_id}", "starred_url": "https://api.github.com/users/apolinario/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/apolinario/subscriptions", "organizations_url": "https://api.github.com/users/apolinario/orgs", "repos_url": "https://api.github.com/users/apolinario/repos", "events_url": "https://api.github.com/users/apolinario/events{/privacy}", "received_events_url": "https://api.github.com/users/apolinario/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-30T16:17:29
2024-01-09T13:06:31
2024-01-09T13:06:31
NONE
null
null
null
### Describe the bug By default, the standard Image Dataset maps out `file_name` to `image` when loading an Image Dataset. However, if the dataset contains only 1 image, this does not take place ### Steps to reproduce the bug Input (dataset with one image `multimodalart/repro_1_image`) ```py from datasets import load_dataset dataset = load_dataset("multimodalart/repro_1_image") dataset ``` Output: ```py DatasetDict({ train: Dataset({ features: ['file_name', 'prompt'], num_rows: 1 }) }) ``` Input (dataset with 2+ images `multimodalart/repro_2_image`) ```py from datasets import load_dataset dataset = load_dataset("multimodalart/repro_2_image") dataset ``` Output: ```py DatasetDict({ train: Dataset({ features: ['image', 'prompt'], num_rows: 2 }) }) ``` ### Expected behavior Expected to map `file_name` → `image` for all dataset sizes, including 1. ### Environment info Both latest main and 2.16.0
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6545/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6545/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6542
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6542/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6542/comments
https://api.github.com/repos/huggingface/datasets/issues/6542/events
https://github.com/huggingface/datasets/issues/6542
2,059,198,575
I_kwDODunzps56vOBv
6,542
Datasets : wikipedia 20220301.en error
{ "login": "ppx666", "id": 53203620, "node_id": "MDQ6VXNlcjUzMjAzNjIw", "avatar_url": "https://avatars.githubusercontent.com/u/53203620?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ppx666", "html_url": "https://github.com/ppx666", "followers_url": "https://api.github.com/users/ppx666/followers", "following_url": "https://api.github.com/users/ppx666/following{/other_user}", "gists_url": "https://api.github.com/users/ppx666/gists{/gist_id}", "starred_url": "https://api.github.com/users/ppx666/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ppx666/subscriptions", "organizations_url": "https://api.github.com/users/ppx666/orgs", "repos_url": "https://api.github.com/users/ppx666/repos", "events_url": "https://api.github.com/users/ppx666/events{/privacy}", "received_events_url": "https://api.github.com/users/ppx666/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-12-29T08:34:51
2024-01-02T13:21:06
2024-01-02T13:20:30
NONE
null
null
null
### Describe the bug When I used load_dataset to download this data set, the following error occurred. The main problem was that the target data did not exist. ### Steps to reproduce the bug 1.I tried downloading directly. ```python wiki_dataset = load_dataset("wikipedia", "20220301.en") ``` An exception occurred ``` MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to run on large-scale data processing tools like Dataflow, Spark, etc. More information about Apache Beam runners at https://beam.apache.org/documentation/runners/capability-matrix/ If you really want to run it locally because you feel like the Dataset is small enough, you can use the local beam runner called `DirectRunner` (you may run out of memory). Example of usage: `load_dataset('wikipedia', '20220301.en', beam_runner='DirectRunner')` ``` 2.I modified the code as prompted. ```python wiki_dataset = load_dataset('wikipedia', '20220301.en', beam_runner='DirectRunner') ``` An exception occurred: ``` FileNotFoundError: Couldn't find file at https://dumps.wikimedia.org/enwiki/20220301/dumpstatus.json ``` ### Expected behavior I searched in the parent directory of the corresponding URL, but there was no corresponding "20220301" directory. I really need this data set and hope to provide a download method. ### Environment info python 3.8 datasets 2.16.0 apache-beam 2.52.0 dill 0.3.7
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6542/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6542/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6541
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6541/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6541/comments
https://api.github.com/repos/huggingface/datasets/issues/6541/events
https://github.com/huggingface/datasets/issues/6541
2,058,983,826
I_kwDODunzps56uZmS
6,541
Dataset not loading successfully.
{ "login": "hi-sushanta", "id": 93595990, "node_id": "U_kgDOBZQpVg", "avatar_url": "https://avatars.githubusercontent.com/u/93595990?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hi-sushanta", "html_url": "https://github.com/hi-sushanta", "followers_url": "https://api.github.com/users/hi-sushanta/followers", "following_url": "https://api.github.com/users/hi-sushanta/following{/other_user}", "gists_url": "https://api.github.com/users/hi-sushanta/gists{/gist_id}", "starred_url": "https://api.github.com/users/hi-sushanta/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hi-sushanta/subscriptions", "organizations_url": "https://api.github.com/users/hi-sushanta/orgs", "repos_url": "https://api.github.com/users/hi-sushanta/repos", "events_url": "https://api.github.com/users/hi-sushanta/events{/privacy}", "received_events_url": "https://api.github.com/users/hi-sushanta/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
4
2023-12-29T01:35:47
2024-01-17T00:40:46
2024-01-17T00:40:45
NONE
null
null
null
### Describe the bug When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning' I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099) ### Steps to reproduce the bug ## Reproduction Hi, please check this line of code, when I run Show attribute error. ``` from datasets import load_dataset from transformers import WhisperProcessor, WhisperForConditionalGeneration # Select an audio file and read it: ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation") audio_sample = ds[0]["audio"] waveform = audio_sample["array"] sampling_rate = audio_sample["sampling_rate"] # Load the Whisper model in Hugging Face format: processor = WhisperProcessor.from_pretrained("openai/whisper-tiny.en") model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-tiny.en") # Use the model and processor to transcribe the audio: input_features = processor( waveform, sampling_rate=sampling_rate, return_tensors="pt" ).input_features # Generate token ids predicted_ids = model.generate(input_features) # Decode token ids to text transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True) transcription[0] ``` **Attribute Error** ``` AttributeError Traceback (most recent call last) Cell In[9], line 6 4 # Select an audio file and read it: 5 ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation") ----> 6 audio_sample = ds[0]["audio"] 7 waveform = audio_sample["array"] 8 sampling_rate = audio_sample["sampling_rate"] File /opt/pytorch/lib/python3.8/site-packages/datasets/arrow_dataset.py:2795, in Dataset.__getitem__(self, key) 2793 def __getitem__(self, key): # noqa: F811 2794 """Can be used to index columns (by string names) or rows (by integer index or iterable of indices or bools).""" -> 2795 return self._getitem(key) File /opt/pytorch/lib/python3.8/site-packages/datasets/arrow_dataset.py:2780, in Dataset._getitem(self, key, **kwargs) 2778 formatter = get_formatter(format_type, features=self._info.features, **format_kwargs) 2779 pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) -> 2780 formatted_output = format_table( 2781 pa_subtable, key, formatter=formatter, format_columns=format_columns, output_all_columns=output_all_columns 2782 ) 2783 return formatted_output File /opt/pytorch/lib/python3.8/site-packages/datasets/formatting/formatting.py:629, in format_table(table, key, formatter, format_columns, output_all_columns) 627 python_formatter = PythonFormatter(features=formatter.features) 628 if format_columns is None: --> 629 return formatter(pa_table, query_type=query_type) 630 elif query_type == "column": 631 if key in format_columns: File /opt/pytorch/lib/python3.8/site-packages/datasets/formatting/formatting.py:396, in Formatter.__call__(self, pa_table, query_type) 394 def __call__(self, pa_table: pa.Table, query_type: str) -> Union[RowFormat, ColumnFormat, BatchFormat]: 395 if query_type == "row": --> 396 return self.format_row(pa_table) 397 elif query_type == "column": 398 return self.format_column(pa_table) File /opt/pytorch/lib/python3.8/site-packages/datasets/formatting/formatting.py:437, in PythonFormatter.format_row(self, pa_table) 435 return LazyRow(pa_table, self) 436 row = self.python_arrow_extractor().extract_row(pa_table) --> 437 row = self.python_features_decoder.decode_row(row) 438 return row File /opt/pytorch/lib/python3.8/site-packages/datasets/formatting/formatting.py:215, in PythonFeaturesDecoder.decode_row(self, row) 214 def decode_row(self, row: dict) -> dict: --> 215 return self.features.decode_example(row) if self.features else row File /opt/pytorch/lib/python3.8/site-packages/datasets/features/features.py:1917, in Features.decode_example(self, example, token_per_repo_id) 1903 def decode_example(self, example: dict, token_per_repo_id: Optional[Dict[str, Union[str, bool, None]]] = None): 1904 """Decode example with custom feature decoding. 1905 1906 Args: (...) 1914 `dict[str, Any]` 1915 """ -> 1917 return { 1918 column_name: decode_nested_example(feature, value, token_per_repo_id=token_per_repo_id) 1919 if self._column_requires_decoding[column_name] 1920 else value 1921 for column_name, (feature, value) in zip_dict( 1922 {key: value for key, value in self.items() if key in example}, example 1923 ) 1924 } File /opt/pytorch/lib/python3.8/site-packages/datasets/features/features.py:1918, in <dictcomp>(.0) 1903 def decode_example(self, example: dict, token_per_repo_id: Optional[Dict[str, Union[str, bool, None]]] = None): 1904 """Decode example with custom feature decoding. 1905 1906 Args: (...) 1914 `dict[str, Any]` 1915 """ 1917 return { -> 1918 column_name: decode_nested_example(feature, value, token_per_repo_id=token_per_repo_id) 1919 if self._column_requires_decoding[column_name] 1920 else value 1921 for column_name, (feature, value) in zip_dict( 1922 {key: value for key, value in self.items() if key in example}, example 1923 ) 1924 } File /opt/pytorch/lib/python3.8/site-packages/datasets/features/features.py:1339, in decode_nested_example(schema, obj, token_per_repo_id) 1336 elif isinstance(schema, (Audio, Image)): 1337 # we pass the token to read and decode files from private repositories in streaming mode 1338 if obj is not None and schema.decode: -> 1339 return schema.decode_example(obj, token_per_repo_id=token_per_repo_id) 1340 return obj File /opt/pytorch/lib/python3.8/site-packages/datasets/features/audio.py:191, in Audio.decode_example(self, value, token_per_repo_id) 189 array = array.T 190 if self.mono: --> 191 array = librosa.to_mono(array) 192 if self.sampling_rate and self.sampling_rate != sampling_rate: 193 array = librosa.resample(array, orig_sr=sampling_rate, target_sr=self.sampling_rate) File /opt/pytorch/lib/python3.8/site-packages/lazy_loader/__init__.py:78, in attach.<locals>.__getattr__(name) 76 submod_path = f"{package_name}.{attr_to_modules[name]}" 77 submod = importlib.import_module(submod_path) ---> 78 attr = getattr(submod, name) 80 # If the attribute lives in a file (module) with the same 81 # name as the attribute, ensure that the attribute and *not* 82 # the module is accessible on the package. 83 if name == attr_to_modules[name]: File /opt/pytorch/lib/python3.8/site-packages/lazy_loader/__init__.py:77, in attach.<locals>.__getattr__(name) 75 elif name in attr_to_modules: 76 submod_path = f"{package_name}.{attr_to_modules[name]}" ---> 77 submod = importlib.import_module(submod_path) 78 attr = getattr(submod, name) 80 # If the attribute lives in a file (module) with the same 81 # name as the attribute, ensure that the attribute and *not* 82 # the module is accessible on the package. File /usr/lib/python3.8/importlib/__init__.py:127, in import_module(name, package) 125 break 126 level += 1 --> 127 return _bootstrap._gcd_import(name[level:], package, level) File <frozen importlib._bootstrap>:1014, in _gcd_import(name, package, level) File <frozen importlib._bootstrap>:991, in _find_and_load(name, import_) File <frozen importlib._bootstrap>:975, in _find_and_load_unlocked(name, import_) File <frozen importlib._bootstrap>:671, in _load_unlocked(spec) File <frozen importlib._bootstrap_external>:848, in exec_module(self, module) File <frozen importlib._bootstrap>:219, in _call_with_frames_removed(f, *args, **kwds) File /opt/pytorch/lib/python3.8/site-packages/librosa/core/audio.py:13 11 import audioread 12 import numpy as np ---> 13 import scipy.signal 14 import soxr 15 import lazy_loader as lazy File /opt/pytorch/lib/python3.8/site-packages/scipy/signal/__init__.py:323 314 from ._spline import ( # noqa: F401 315 cspline2d, 316 qspline2d, (...) 319 symiirorder2, 320 ) 322 from ._bsplines import * --> 323 from ._filter_design import * 324 from ._fir_filter_design import * 325 from ._ltisys import * File /opt/pytorch/lib/python3.8/site-packages/scipy/signal/_filter_design.py:16 13 from numpy.polynomial.polynomial import polyval as npp_polyval 14 from numpy.polynomial.polynomial import polyvalfromroots ---> 16 from scipy import special, optimize, fft as sp_fft 17 from scipy.special import comb 18 from scipy._lib._util import float_factorial File /opt/pytorch/lib/python3.8/site-packages/scipy/optimize/__init__.py:405 1 """ 2 ===================================================== 3 Optimization and root finding (:mod:`scipy.optimize`) (...) 401 402 """ 404 from ._optimize import * --> 405 from ._minimize import * 406 from ._root import * 407 from ._root_scalar import * File /opt/pytorch/lib/python3.8/site-packages/scipy/optimize/_minimize.py:26 24 from ._trustregion_krylov import _minimize_trust_krylov 25 from ._trustregion_exact import _minimize_trustregion_exact ---> 26 from ._trustregion_constr import _minimize_trustregion_constr 28 # constrained minimization 29 from ._lbfgsb_py import _minimize_lbfgsb File /opt/pytorch/lib/python3.8/site-packages/scipy/optimize/_trustregion_constr/__init__.py:4 1 """This module contains the equality constrained SQP solver.""" ----> 4 from .minimize_trustregion_constr import _minimize_trustregion_constr 6 __all__ = ['_minimize_trustregion_constr'] File /opt/pytorch/lib/python3.8/site-packages/scipy/optimize/_trustregion_constr/minimize_trustregion_constr.py:5 3 from scipy.sparse.linalg import LinearOperator 4 from .._differentiable_functions import VectorFunction ----> 5 from .._constraints import ( 6 NonlinearConstraint, LinearConstraint, PreparedConstraint, strict_bounds) 7 from .._hessian_update_strategy import BFGS 8 from .._optimize import OptimizeResult File /opt/pytorch/lib/python3.8/site-packages/scipy/optimize/_constraints.py:8 6 from ._optimize import OptimizeWarning 7 from warnings import warn, catch_warnings, simplefilter ----> 8 from numpy.testing import suppress_warnings 9 from scipy.sparse import issparse 12 def _arr_to_scalar(x): 13 # If x is a numpy array, return x.item(). This will 14 # fail if the array has more than one element. File /opt/pytorch/lib/python3.8/site-packages/numpy/testing/__init__.py:11 8 from unittest import TestCase 10 from . import _private ---> 11 from ._private.utils import * 12 from ._private.utils import (_assert_valid_refcount, _gen_alignment_data) 13 from ._private import extbuild, decorators as dec File /opt/pytorch/lib/python3.8/site-packages/numpy/testing/_private/utils.py:480 476 pprint.pprint(desired, msg) 477 raise AssertionError(msg.getvalue()) --> 480 @np._no_nep50_warning() 481 def assert_almost_equal(actual,desired,decimal=7,err_msg='',verbose=True): 482 """ 483 Raises an AssertionError if two items are not equal up to desired 484 precision. (...) 548 549 """ 550 __tracebackhide__ = True # Hide traceback for py.test File /opt/pytorch/lib/python3.8/site-packages/numpy/__init__.py:313, in __getattr__(attr) 305 raise AttributeError(__former_attrs__[attr]) 307 # Importing Tester requires importing all of UnitTest which is not a 308 # cheap import Since it is mainly used in test suits, we lazy import it 309 # here to save on the order of 10 ms of import time for most users 310 # 311 # The previous way Tester was imported also had a side effect of adding 312 # the full `numpy.testing` namespace --> 313 if attr == 'testing': 314 import numpy.testing as testing 315 return testing AttributeError: module 'numpy' has no attribute '_no_nep50_warning' ``` ### Expected behavior ``` ' Mr. Quilter is the apostle of the middle classes, and we are glad to welcome his gospel.' ``` Also, make sure this script is provided for your official website so please update: [script](https://huggingface.co/docs/transformers/model_doc/whisper) ### Environment info **System Info** * transformers -> 4.36.1 * datasets -> 2.15.0 * huggingface_hub -> 0.19.4 * python -> 3.8.10 * accelerate -> 0.25.0 * pytorch -> 2.0.1+cpu * Using GPU in Script -> No
{ "login": "hi-sushanta", "id": 93595990, "node_id": "U_kgDOBZQpVg", "avatar_url": "https://avatars.githubusercontent.com/u/93595990?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hi-sushanta", "html_url": "https://github.com/hi-sushanta", "followers_url": "https://api.github.com/users/hi-sushanta/followers", "following_url": "https://api.github.com/users/hi-sushanta/following{/other_user}", "gists_url": "https://api.github.com/users/hi-sushanta/gists{/gist_id}", "starred_url": "https://api.github.com/users/hi-sushanta/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hi-sushanta/subscriptions", "organizations_url": "https://api.github.com/users/hi-sushanta/orgs", "repos_url": "https://api.github.com/users/hi-sushanta/repos", "events_url": "https://api.github.com/users/hi-sushanta/events{/privacy}", "received_events_url": "https://api.github.com/users/hi-sushanta/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6541/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6541/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6540
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6540/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6540/comments
https://api.github.com/repos/huggingface/datasets/issues/6540/events
https://github.com/huggingface/datasets/issues/6540
2,058,965,157
I_kwDODunzps56uVCl
6,540
Extreme inefficiency for `save_to_disk` when merging datasets
{ "login": "KatarinaYuan", "id": 43512683, "node_id": "MDQ6VXNlcjQzNTEyNjgz", "avatar_url": "https://avatars.githubusercontent.com/u/43512683?v=4", "gravatar_id": "", "url": "https://api.github.com/users/KatarinaYuan", "html_url": "https://github.com/KatarinaYuan", "followers_url": "https://api.github.com/users/KatarinaYuan/followers", "following_url": "https://api.github.com/users/KatarinaYuan/following{/other_user}", "gists_url": "https://api.github.com/users/KatarinaYuan/gists{/gist_id}", "starred_url": "https://api.github.com/users/KatarinaYuan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/KatarinaYuan/subscriptions", "organizations_url": "https://api.github.com/users/KatarinaYuan/orgs", "repos_url": "https://api.github.com/users/KatarinaYuan/repos", "events_url": "https://api.github.com/users/KatarinaYuan/events{/privacy}", "received_events_url": "https://api.github.com/users/KatarinaYuan/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
1
2023-12-29T00:44:35
2023-12-30T15:05:48
null
NONE
null
null
null
### Describe the bug Hi, I tried to merge in total 22M sequences of data, where each sequence is of maximum length 2000. I found that merging these datasets and then `save_to_disk` is extremely slow because of flattening the indices. Wondering if you have any suggestions or guidance on this. Thank you very much! ### Steps to reproduce the bug The source data is too big to demonstrate ### Expected behavior The source data is too big to demonstrate ### Environment info python 3.9.0 datasets 2.7.0 pytorch 2.0.0 tokenizers 0.13.1 transformers 4.31.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6540/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6540/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6539
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6539/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6539/comments
https://api.github.com/repos/huggingface/datasets/issues/6539/events
https://github.com/huggingface/datasets/issues/6539
2,058,493,960
I_kwDODunzps56siAI
6,539
'Repo card metadata block was not found' when loading a pragmeval dataset
{ "login": "lambdaofgod", "id": 3647577, "node_id": "MDQ6VXNlcjM2NDc1Nzc=", "avatar_url": "https://avatars.githubusercontent.com/u/3647577?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lambdaofgod", "html_url": "https://github.com/lambdaofgod", "followers_url": "https://api.github.com/users/lambdaofgod/followers", "following_url": "https://api.github.com/users/lambdaofgod/following{/other_user}", "gists_url": "https://api.github.com/users/lambdaofgod/gists{/gist_id}", "starred_url": "https://api.github.com/users/lambdaofgod/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lambdaofgod/subscriptions", "organizations_url": "https://api.github.com/users/lambdaofgod/orgs", "repos_url": "https://api.github.com/users/lambdaofgod/repos", "events_url": "https://api.github.com/users/lambdaofgod/events{/privacy}", "received_events_url": "https://api.github.com/users/lambdaofgod/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2023-12-28T14:18:25
2023-12-28T14:18:37
null
NONE
null
null
null
### Describe the bug I can't load dataset subsets of 'pragmeval'. The funny thing is I ran the dataset author's [colab notebook](https://colab.research.google.com/drive/1sg--LF4z7XR1wxAOfp0-3d4J6kQ9nj_A?usp=sharing) and it works just fine. I tried to install exactly the same packages that are installed on colab using poetry, so my environment info only differs from the one from colab in linux version - I still get the same bug outside colab. ### Steps to reproduce the bug Install dependencies with poetry pyproject.toml ``` [tool.poetry] name = "project" version = "0.1.0" description = "" authors = [] [tool.poetry.dependencies] python = "^3.10" datasets = "2.16.0" pandas = "1.5.3" pyarrow = "10.0.1" huggingface-hub = "0.19.4" fsspec = "2023.6.0" [build-system] requires = ["poetry-core"] build-backend = "poetry.core.masonry.api" ``` `poetry run python -c "import datasets; print(datasets.get_dataset_config_names('pragmeval'))` prints ['default'] ### Expected behavior The command should print ``` ['emergent', 'emobank-arousal', 'emobank-dominance', 'emobank-valence', 'gum', 'mrda', 'pdtb', 'persuasiveness-claimtype', 'persuasiveness-eloquence', 'persuasiveness-premisetype', 'persuasiveness-relevance', 'persuasiveness-specificity', 'persuasiveness-strength', 'sarcasm', 'squinky-formality', 'squinky-implicature', 'squinky-informativeness', 'stac', 'switchboard', 'verifiability'] ``` ### Environment info - `datasets` version: 2.16.0 - Platform: Linux-6.2.0-37-generic-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.19.4 - PyArrow version: 10.0.1 - Pandas version: 1.5.3 - `fsspec` version: 2023.6.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6539/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6539/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6538
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6538/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6538/comments
https://api.github.com/repos/huggingface/datasets/issues/6538/events
https://github.com/huggingface/datasets/issues/6538
2,057,377,630
I_kwDODunzps56oRde
6,538
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
{ "login": "Sonali-Behera-TRT", "id": 131662185, "node_id": "U_kgDOB9kBaQ", "avatar_url": "https://avatars.githubusercontent.com/u/131662185?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Sonali-Behera-TRT", "html_url": "https://github.com/Sonali-Behera-TRT", "followers_url": "https://api.github.com/users/Sonali-Behera-TRT/followers", "following_url": "https://api.github.com/users/Sonali-Behera-TRT/following{/other_user}", "gists_url": "https://api.github.com/users/Sonali-Behera-TRT/gists{/gist_id}", "starred_url": "https://api.github.com/users/Sonali-Behera-TRT/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Sonali-Behera-TRT/subscriptions", "organizations_url": "https://api.github.com/users/Sonali-Behera-TRT/orgs", "repos_url": "https://api.github.com/users/Sonali-Behera-TRT/repos", "events_url": "https://api.github.com/users/Sonali-Behera-TRT/events{/privacy}", "received_events_url": "https://api.github.com/users/Sonali-Behera-TRT/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
15
2023-12-27T13:31:16
2024-01-03T10:06:47
2024-01-03T10:04:58
NONE
null
null
null
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, logging ) from peft import LoraConfig, PeftModel from trl import SFTTrainer from huggingface_hub import login import pandas as pd ``` Error: ```` --------------------------------------------------------------------------- ImportError Traceback (most recent call last) Cell In[5], line 14 4 from transformers import ( 5 AutoModelForCausalLM, 6 AutoTokenizer, (...) 11 logging 12 ) 13 from peft import LoraConfig, PeftModel ---> 14 from trl import SFTTrainer 15 from huggingface_hub import login 16 import pandas as pd File /opt/conda/lib/python3.10/site-packages/trl/__init__.py:21 8 from .import_utils import ( 9 is_diffusers_available, 10 is_npu_available, (...) 13 is_xpu_available, 14 ) 15 from .models import ( 16 AutoModelForCausalLMWithValueHead, 17 AutoModelForSeq2SeqLMWithValueHead, 18 PreTrainedModelWrapper, 19 create_reference_model, 20 ) ---> 21 from .trainer import ( 22 DataCollatorForCompletionOnlyLM, 23 DPOTrainer, 24 IterativeSFTTrainer, 25 PPOConfig, 26 PPOTrainer, 27 RewardConfig, 28 RewardTrainer, 29 SFTTrainer, 30 ) 33 if is_diffusers_available(): 34 from .models import ( 35 DDPOPipelineOutput, 36 DDPOSchedulerOutput, 37 DDPOStableDiffusionPipeline, 38 DefaultDDPOStableDiffusionPipeline, 39 ) File /opt/conda/lib/python3.10/site-packages/trl/trainer/__init__.py:44 42 from .ppo_trainer import PPOTrainer 43 from .reward_trainer import RewardTrainer, compute_accuracy ---> 44 from .sft_trainer import SFTTrainer 45 from .training_configs import RewardConfig File /opt/conda/lib/python3.10/site-packages/trl/trainer/sft_trainer.py:23 21 import torch.nn as nn 22 from datasets import Dataset ---> 23 from datasets.arrow_writer import SchemaInferenceError 24 from datasets.builder import DatasetGenerationError 25 from transformers import ( 26 AutoModelForCausalLM, 27 AutoTokenizer, (...) 33 TrainingArguments, 34 ) ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py ```` transformers version: 4.36.2 python version: 3.10.12 datasets version: 2.16.1 ### Steps to reproduce the bug 1. Install packages ``` !pip install -U datasets trl accelerate peft bitsandbytes transformers trl huggingface_hub ``` 2. import packages ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, logging ) from peft import LoraConfig, PeftModel from trl import SFTTrainer from huggingface_hub import login import pandas as pd ``` ### Expected behavior No error while importing ### Environment info - `datasets` version: 2.16.0 - Platform: Linux-5.15.133+-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.20.1 - PyArrow version: 11.0.0 - Pandas version: 2.1.4 - `fsspec` version: 2023.10.0
{ "login": "Sonali-Behera-TRT", "id": 131662185, "node_id": "U_kgDOB9kBaQ", "avatar_url": "https://avatars.githubusercontent.com/u/131662185?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Sonali-Behera-TRT", "html_url": "https://github.com/Sonali-Behera-TRT", "followers_url": "https://api.github.com/users/Sonali-Behera-TRT/followers", "following_url": "https://api.github.com/users/Sonali-Behera-TRT/following{/other_user}", "gists_url": "https://api.github.com/users/Sonali-Behera-TRT/gists{/gist_id}", "starred_url": "https://api.github.com/users/Sonali-Behera-TRT/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Sonali-Behera-TRT/subscriptions", "organizations_url": "https://api.github.com/users/Sonali-Behera-TRT/orgs", "repos_url": "https://api.github.com/users/Sonali-Behera-TRT/repos", "events_url": "https://api.github.com/users/Sonali-Behera-TRT/events{/privacy}", "received_events_url": "https://api.github.com/users/Sonali-Behera-TRT/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6538/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6538/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6537
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6537/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6537/comments
https://api.github.com/repos/huggingface/datasets/issues/6537/events
https://github.com/huggingface/datasets/issues/6537
2,057,132,173
I_kwDODunzps56nViN
6,537
Adding support for netCDF (*.nc) files
{ "login": "shermansiu", "id": 12627125, "node_id": "MDQ6VXNlcjEyNjI3MTI1", "avatar_url": "https://avatars.githubusercontent.com/u/12627125?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shermansiu", "html_url": "https://github.com/shermansiu", "followers_url": "https://api.github.com/users/shermansiu/followers", "following_url": "https://api.github.com/users/shermansiu/following{/other_user}", "gists_url": "https://api.github.com/users/shermansiu/gists{/gist_id}", "starred_url": "https://api.github.com/users/shermansiu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shermansiu/subscriptions", "organizations_url": "https://api.github.com/users/shermansiu/orgs", "repos_url": "https://api.github.com/users/shermansiu/repos", "events_url": "https://api.github.com/users/shermansiu/events{/privacy}", "received_events_url": "https://api.github.com/users/shermansiu/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
3
2023-12-27T09:27:29
2023-12-27T20:46:53
null
NONE
null
null
null
### Feature request netCDF (*.nc) is a file format for storing multidimensional scientific data, which is used by packages like `xarray` (labelled multi-dimensional arrays in Python). It would be nice to have native support for netCDF in `datasets`. ### Motivation When uploading *.nc files onto Huggingface Hub through the `datasets` API, I would like to be able to preview the dataset without converting it to another format. ### Your contribution I can submit a PR, provided I have the time.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6537/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6537/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6536
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6536/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6536/comments
https://api.github.com/repos/huggingface/datasets/issues/6536/events
https://github.com/huggingface/datasets/issues/6536
2,056,863,239
I_kwDODunzps56mT4H
6,536
datasets.load_dataset raises FileNotFoundError for datasets==2.16.0
{ "login": "ArvinZhuang", "id": 46237844, "node_id": "MDQ6VXNlcjQ2MjM3ODQ0", "avatar_url": "https://avatars.githubusercontent.com/u/46237844?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ArvinZhuang", "html_url": "https://github.com/ArvinZhuang", "followers_url": "https://api.github.com/users/ArvinZhuang/followers", "following_url": "https://api.github.com/users/ArvinZhuang/following{/other_user}", "gists_url": "https://api.github.com/users/ArvinZhuang/gists{/gist_id}", "starred_url": "https://api.github.com/users/ArvinZhuang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArvinZhuang/subscriptions", "organizations_url": "https://api.github.com/users/ArvinZhuang/orgs", "repos_url": "https://api.github.com/users/ArvinZhuang/repos", "events_url": "https://api.github.com/users/ArvinZhuang/events{/privacy}", "received_events_url": "https://api.github.com/users/ArvinZhuang/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false } ]
null
2
2023-12-27T03:15:48
2023-12-30T18:58:04
2023-12-30T15:54:00
NONE
null
null
null
### Describe the bug Seems `datasets.load_dataset` raises FileNotFoundError for some hub datasets with the latest `datasets==2.16.0` ### Steps to reproduce the bug For example `pip install datasets==2.16.0` then ```python import datasets datasets.load_dataset("wentingzhao/anthropic-hh-first-prompt", cache_dir='cache1')["train"] ``` This will raise: ```bash Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/load.py", line 2545, in load_dataset builder_instance.download_and_prepare( File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/builder.py", line 1003, in download_and_prepare self._download_and_prepare( File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/builder.py", line 1076, in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/packaged_modules/parquet/parquet.py", line 43, in _split_generators data_files = dl_manager.download_and_extract(self.config.data_files) File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/download/download_manager.py", line 566, in download_and_extract return self.extract(self.download(url_or_urls)) File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/download/download_manager.py", line 539, in extract extracted_paths = map_nested( File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 466, in map_nested mapped = [ File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 467, in <listcomp> _single_map_nested((function, obj, types, None, True, None)) File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 387, in _single_map_nested mapped = [_single_map_nested((function, v, types, None, True, None)) for v in pbar] File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 387, in <listcomp> mapped = [_single_map_nested((function, v, types, None, True, None)) for v in pbar] File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 370, in _single_map_nested return function(data_struct) File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/download/download_manager.py", line 451, in _download out = cached_path(url_or_filename, download_config=download_config) File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/utils/file_utils.py", line 188, in cached_path output_path = get_from_cache( File "/Users/xxx/miniconda3/envs/env/lib/python3.9/site-packages/datasets/utils/file_utils.py", line 570, in get_from_cache raise FileNotFoundError(f"Couldn't find file at {url}") FileNotFoundError: Couldn't find file at https://huggingface.co/datasets/wentingzhao/anthropic-hh-first-prompt/resolve/11b393a5545f706a357ebcd4a5285d93db176715/cache1/downloads/87d66c365626feca116cba323c4856c9aae056e4503f09f23e34aa085eb9de15 ``` However, seems it works fine for some datasets, for example, if works fine for `datasets.load_dataset("ag_news", cache_dir='cache2')["test"]` But the dataset works fine for datasets==2.15.0, for example `pip install datasets==2.15.0`, then ```python import datasets datasets.load_dataset("wentingzhao/anthropic-hh-first-prompt", cache_dir='cache3')["train"] Dataset({ features: ['user', 'system', 'source'], num_rows: 8552 }) ``` ### Expected behavior 2.16.0 should work as same as 2.15.0 for all datasets ### Environment info python3.9 conda env tested on MacOS and Linux
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6536/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6536/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6535
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6535/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6535/comments
https://api.github.com/repos/huggingface/datasets/issues/6535/events
https://github.com/huggingface/datasets/issues/6535
2,056,264,339
I_kwDODunzps56kBqT
6,535
IndexError: Invalid key: 47682 is out of bounds for size 0 while using PEFT
{ "login": "MahavirDabas18", "id": 57484266, "node_id": "MDQ6VXNlcjU3NDg0MjY2", "avatar_url": "https://avatars.githubusercontent.com/u/57484266?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MahavirDabas18", "html_url": "https://github.com/MahavirDabas18", "followers_url": "https://api.github.com/users/MahavirDabas18/followers", "following_url": "https://api.github.com/users/MahavirDabas18/following{/other_user}", "gists_url": "https://api.github.com/users/MahavirDabas18/gists{/gist_id}", "starred_url": "https://api.github.com/users/MahavirDabas18/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MahavirDabas18/subscriptions", "organizations_url": "https://api.github.com/users/MahavirDabas18/orgs", "repos_url": "https://api.github.com/users/MahavirDabas18/repos", "events_url": "https://api.github.com/users/MahavirDabas18/events{/privacy}", "received_events_url": "https://api.github.com/users/MahavirDabas18/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
3
2023-12-26T10:14:33
2024-02-05T08:42:31
null
NONE
null
null
null
### Describe the bug I am trying to fine-tune the t5 model on the paraphrasing task. While running the same code without- model = get_peft_model(model, config) the model trains without any issues. However, using the model returned from get_peft_model raises the following error due to datasets- IndexError: Invalid key: 47682 is out of bounds for size 0. I had raised this in https://github.com/huggingface/peft/issues/1299#issue-2056173386 and they suggested that I raise it here. Here is the complete error- IndexError Traceback (most recent call last) in <cell line: 1>() ----> 1 trainer.train() 11 frames [/usr/local/lib/python3.10/dist-packages/transformers/trainer.py](https://localhost:8080/#) in train(self, resume_from_checkpoint, trial, ignore_keys_for_eval, **kwargs) 1553 hf_hub_utils.enable_progress_bars() 1554 else: -> 1555 return inner_training_loop( 1556 args=args, 1557 resume_from_checkpoint=resume_from_checkpoint, [/usr/local/lib/python3.10/dist-packages/transformers/trainer.py](https://localhost:8080/#) in _inner_training_loop(self, batch_size, args, resume_from_checkpoint, trial, ignore_keys_for_eval) 1836 1837 step = -1 -> 1838 for step, inputs in enumerate(epoch_iterator): 1839 total_batched_samples += 1 1840 if rng_to_sync: [/usr/local/lib/python3.10/dist-packages/accelerate/data_loader.py](https://localhost:8080/#) in iter(self) 446 # We iterate one batch ahead to check when we are at the end 447 try: --> 448 current_batch = next(dataloader_iter) 449 except StopIteration: 450 yield [/usr/local/lib/python3.10/dist-packages/torch/utils/data/dataloader.py](https://localhost:8080/#) in next(self) 628 # TODO(https://github.com/pytorch/pytorch/issues/76750) 629 self._reset() # type: ignore[call-arg] --> 630 data = self._next_data() 631 self._num_yielded += 1 632 if self._dataset_kind == _DatasetKind.Iterable and \ [/usr/local/lib/python3.10/dist-packages/torch/utils/data/dataloader.py](https://localhost:8080/#) in _next_data(self) 672 def _next_data(self): 673 index = self._next_index() # may raise StopIteration --> 674 data = self._dataset_fetcher.fetch(index) # may raise StopIteration 675 if self._pin_memory: 676 data = _utils.pin_memory.pin_memory(data, self._pin_memory_device) [/usr/local/lib/python3.10/dist-packages/torch/utils/data/_utils/fetch.py](https://localhost:8080/#) in fetch(self, possibly_batched_index) 47 if self.auto_collation: 48 if hasattr(self.dataset, "getitems") and self.dataset.getitems: ---> 49 data = self.dataset.getitems(possibly_batched_index) 50 else: 51 data = [self.dataset[idx] for idx in possibly_batched_index] [/usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in getitems(self, keys) 2802 def getitems(self, keys: List) -> List: 2803 """Can be used to get a batch using a list of integers indices.""" -> 2804 batch = self.getitem(keys) 2805 n_examples = len(batch[next(iter(batch))]) 2806 return [{col: array[i] for col, array in batch.items()} for i in range(n_examples)] [/usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in getitem(self, key) 2798 def getitem(self, key): # noqa: F811 2799 """Can be used to index columns (by string names) or rows (by integer index or iterable of indices or bools).""" -> 2800 return self._getitem(key) 2801 2802 def getitems(self, keys: List) -> List: [/usr/local/lib/python3.10/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in _getitem(self, key, **kwargs) 2782 format_kwargs = format_kwargs if format_kwargs is not None else {} 2783 formatter = get_formatter(format_type, features=self._info.features, **format_kwargs) -> 2784 pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) 2785 formatted_output = format_table( 2786 pa_subtable, key, formatter=formatter, format_columns=format_columns, output_all_columns=output_all_columns [/usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py](https://localhost:8080/#) in query_table(table, key, indices) 581 else: 582 size = indices.num_rows if indices is not None else table.num_rows --> 583 _check_valid_index_key(key, size) 584 # Query the main table 585 if indices is None: [/usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py](https://localhost:8080/#) in _check_valid_index_key(key, size) 534 elif isinstance(key, Iterable): 535 if len(key) > 0: --> 536 _check_valid_index_key(int(max(key)), size=size) 537 _check_valid_index_key(int(min(key)), size=size) 538 else: [/usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py](https://localhost:8080/#) in _check_valid_index_key(key, size) 524 if isinstance(key, int): 525 if (key < 0 and key + size < 0) or (key >= size): --> 526 raise IndexError(f"Invalid key: {key} is out of bounds for size {size}") 527 return 528 elif isinstance(key, slice): IndexError: Invalid key: 47682 is out of bounds for size 0 ### Steps to reproduce the bug device = "cuda:0" if torch.cuda.is_available() else "cpu" #defining model name for tokenizer and model loading model_name= "t5-small" #loading the tokenizer tokenizer = AutoTokenizer.from_pretrained(model_name) def preprocess_function(data, tokenizer): inputs = [f"Paraphrase this sentence: {doc}" for doc in data["text"]] model_inputs = tokenizer(inputs, max_length=150, truncation=True) labels = [ast.literal_eval(i)[0] for i in data['paraphrases']] labels = tokenizer(labels, max_length=150, truncation=True) model_inputs["labels"] = labels["input_ids"] return model_inputs train_dataset = load_dataset("humarin/chatgpt-paraphrases", split="train").shuffle(seed=42).select(range(50000)) val_dataset = load_dataset("humarin/chatgpt-paraphrases", split="train").shuffle(seed=42).select(range(50000,55000)) tokenized_train = train_dataset.map(lambda batch: preprocess_function(batch, tokenizer), batched=True) tokenized_val = val_dataset.map(lambda batch: preprocess_function(batch, tokenizer), batched=True) def print_trainable_parameters(model): """ Prints the number of trainable parameters in the model. """ trainable_params = 0 all_param = 0 for _, param in model.named_parameters(): all_param += param.numel() if param.requires_grad: trainable_params += param.numel() print( f"trainable params: {trainable_params} || all params: {all_param} || trainable%: {100 * trainable_params / all_param}" ) config = LoraConfig( r=16, #attention heads lora_alpha=32, #alpha scaling lora_dropout=0.05, bias="none", task_type="Seq2Seq" ) #loading the model model = AutoModelForSeq2SeqLM.from_pretrained(model_name).to(device) model = get_peft_model(model, config) print_trainable_parameters(model) #loading the data collator data_collator = DataCollatorForSeq2Seq( tokenizer=tokenizer, model=model, label_pad_token_id=-100, padding="longest" ) #defining the training arguments training_args = Seq2SeqTrainingArguments( output_dir=os.getcwd(), evaluation_strategy="epoch", save_strategy="epoch", learning_rate=2e-5, per_device_train_batch_size=16, per_device_eval_batch_size=16, weight_decay=1e-3, save_total_limit=3, load_best_model_at_end=True, num_train_epochs=1, predict_with_generate=True ) def compute_metric_with_extra(tokenizer): def compute_metrics(eval_preds): metric = evaluate.load('rouge') preds, labels = eval_preds # decode preds and labels labels = np.where(labels != -100, labels, tokenizer.pad_token_id) decoded_preds = tokenizer.batch_decode(preds, skip_special_tokens=True) decoded_labels = tokenizer.batch_decode(labels, skip_special_tokens=True) # rougeLSum expects newline after each sentence decoded_preds = ["\n".join(nltk.sent_tokenize(pred.strip())) for pred in decoded_preds] decoded_labels = ["\n".join(nltk.sent_tokenize(label.strip())) for label in decoded_labels] result = metric.compute(predictions=decoded_preds, references=decoded_labels, use_stemmer=True) return result return compute_metrics trainer = Seq2SeqTrainer( model=model, args=training_args, train_dataset=tokenized_train, eval_dataset=tokenized_val, tokenizer=tokenizer, data_collator=data_collator, compute_metrics= compute_metric_with_extra(tokenizer) ) trainer.train() ### Expected behavior I would want the trainer to train normally as it was before I used- model = get_peft_model(model, config) ### Environment info datasets version- 2.16.0 peft version- 0.7.1 transformers version- 4.35.2 accelerate version- 0.25.0 python- 3.10.12 enviroment- google colab
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6535/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6535/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6534
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6534/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6534/comments
https://api.github.com/repos/huggingface/datasets/issues/6534/events
https://github.com/huggingface/datasets/issues/6534
2,056,002,548
I_kwDODunzps56jBv0
6,534
How to configure multiple folders in the same zip package
{ "login": "d710055071", "id": 12895488, "node_id": "MDQ6VXNlcjEyODk1NDg4", "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "gravatar_id": "", "url": "https://api.github.com/users/d710055071", "html_url": "https://github.com/d710055071", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url": "https://api.github.com/users/d710055071/gists{/gist_id}", "starred_url": "https://api.github.com/users/d710055071/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/d710055071/subscriptions", "organizations_url": "https://api.github.com/users/d710055071/orgs", "repos_url": "https://api.github.com/users/d710055071/repos", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "received_events_url": "https://api.github.com/users/d710055071/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
1
2023-12-26T03:56:20
2023-12-26T06:31:16
null
CONTRIBUTOR
null
null
null
How should I write "config" in readme when all the data, such as train test, is in a zip file train floder and test floder in data.zip
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6534/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6534/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6533
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6533/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6533/comments
https://api.github.com/repos/huggingface/datasets/issues/6533/events
https://github.com/huggingface/datasets/issues/6533
2,055,929,101
I_kwDODunzps56iv0N
6,533
ted_talks_iwslt | Error: Config name is missing
{ "login": "rayliuca", "id": 35850903, "node_id": "MDQ6VXNlcjM1ODUwOTAz", "avatar_url": "https://avatars.githubusercontent.com/u/35850903?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rayliuca", "html_url": "https://github.com/rayliuca", "followers_url": "https://api.github.com/users/rayliuca/followers", "following_url": "https://api.github.com/users/rayliuca/following{/other_user}", "gists_url": "https://api.github.com/users/rayliuca/gists{/gist_id}", "starred_url": "https://api.github.com/users/rayliuca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rayliuca/subscriptions", "organizations_url": "https://api.github.com/users/rayliuca/orgs", "repos_url": "https://api.github.com/users/rayliuca/repos", "events_url": "https://api.github.com/users/rayliuca/events{/privacy}", "received_events_url": "https://api.github.com/users/rayliuca/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false } ]
null
2
2023-12-26T00:38:18
2023-12-30T18:58:21
2023-12-30T16:09:50
NONE
null
null
null
### Describe the bug Running load_dataset using the newest `datasets` library like below on the ted_talks_iwslt using year pair data will throw an error "Config name is missing" see also: https://huggingface.co/datasets/ted_talks_iwslt/discussions/3 likely caused by #6493, where the `and not config_kwargs` part in the if logic was removed https://github.com/huggingface/datasets/blob/ef3b5dd3633995c95d77f35fb17f89ff44990bc4/src/datasets/builder.py#L512 ### Steps to reproduce the bug run: ```python load_dataset("ted_talks_iwslt", language_pair=("ja", "en"), year="2015") ``` ### Expected behavior Load the data without error ### Environment info datasets 2.16.0
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6533/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6533/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6532
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6532/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6532/comments
https://api.github.com/repos/huggingface/datasets/issues/6532/events
https://github.com/huggingface/datasets/issues/6532
2,055,631,201
I_kwDODunzps56hnFh
6,532
[Feature request] Indexing datasets by a customly-defined id field to enable random access dataset items via the id
{ "login": "Yu-Shi", "id": 3377221, "node_id": "MDQ6VXNlcjMzNzcyMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/3377221?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Yu-Shi", "html_url": "https://github.com/Yu-Shi", "followers_url": "https://api.github.com/users/Yu-Shi/followers", "following_url": "https://api.github.com/users/Yu-Shi/following{/other_user}", "gists_url": "https://api.github.com/users/Yu-Shi/gists{/gist_id}", "starred_url": "https://api.github.com/users/Yu-Shi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Yu-Shi/subscriptions", "organizations_url": "https://api.github.com/users/Yu-Shi/orgs", "repos_url": "https://api.github.com/users/Yu-Shi/repos", "events_url": "https://api.github.com/users/Yu-Shi/events{/privacy}", "received_events_url": "https://api.github.com/users/Yu-Shi/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
6
2023-12-25T11:37:10
2024-08-03T11:00:30
null
NONE
null
null
null
### Feature request Some datasets may contain an id-like field, for example the `id` field in [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) and the `_id` field in [BeIR/dbpedia-entity](https://huggingface.co/datasets/BeIR/dbpedia-entity). HF datasets support efficient random access via row, but not via this kinds of id fields. I wonder if it is possible to add support for indexing by a custom "id-like" field to enable random access via such ids. The ids may be numbers or strings. ### Motivation In some cases, especially during inference/evaluation, I may want to find out the item that has a specified id, defined by the dataset itself. For example, in a typical re-ranking setting in information retrieval, the user may want to re-rank the set of candidate documents of each query. The input is usually presented in a TREC-style run file, with the following format: ``` <qid> Q0 <docno> <rank> <score> <tag> ``` The re-ranking program should be able to fetch the queries and documents according to the `<qid>` and `<docno>`, which are the original id defined in the query/document datasets. To accomplish this, I have to iterate over the whole HF dataset to get the mapping from real ids to row ids every time I start the program, which is time-consuming. Thus I want HF dataset to provide options for users to index by a custom id column, not by row. ### Your contribution I'm not an expert in this project and I'm afraid that I'm not able to make contributions on the code.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6532/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6532/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6530
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6530/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6530/comments
https://api.github.com/repos/huggingface/datasets/issues/6530/events
https://github.com/huggingface/datasets/issues/6530
2,054,817,609
I_kwDODunzps56egdJ
6,530
Impossible to save a mapped dataset to disk
{ "login": "kopyl", "id": 17604849, "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kopyl", "html_url": "https://github.com/kopyl", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "organizations_url": "https://api.github.com/users/kopyl/orgs", "repos_url": "https://api.github.com/users/kopyl/repos", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "received_events_url": "https://api.github.com/users/kopyl/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
1
2023-12-23T15:18:27
2023-12-24T09:40:30
null
NONE
null
null
null
### Describe the bug I want to play around with different hyperparameters when training but don't want to re-map my dataset with 3 million samples each time for tens of hours when I [fully fine-tune SDXL](https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_sdxl.py). After I do the mapping like this: ``` train_dataset = train_dataset.map(compute_embeddings_fn, batched=True) train_dataset = train_dataset.map( compute_vae_encodings_fn, batched=True, batch_size=16, ) ``` and try to save it like this: `train_dataset.save_to_disk("test")` i get this error ([full traceback](https://pastebin.com/kq3vt739)): ``` TypeError: Object of type function is not JSON serializable The format kwargs must be JSON serializable, but key 'transform' isn't. ``` But what is interesting is that pushing to hub works like that: `train_dataset.push_to_hub("kopyl/mapped-833-icons-sdxl-1024-dataset", token=True)` Here is the link of the pushed dataset: https://huggingface.co/datasets/kopyl/mapped-833-icons-sdxl-1024-dataset ### Steps to reproduce the bug Here is the self-contained notebook: https://colab.research.google.com/drive/1RtCsEMVcwWcMwlWURk_cj_9xUBHz065M?usp=sharing ### Expected behavior It should be easily saved to disk ### Environment info NVIDIA A100, Linux (NC24ads A100 v4 from Azure), CUDA 12.2. [pip freeze](https://pastebin.com/QTNb6iru)
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6530/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6530/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6529
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6529/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6529/comments
https://api.github.com/repos/huggingface/datasets/issues/6529/events
https://github.com/huggingface/datasets/issues/6529
2,054,209,449
I_kwDODunzps56cL-p
6,529
Impossible to only download a test split
{ "login": "ysig", "id": 28439529, "node_id": "MDQ6VXNlcjI4NDM5NTI5", "avatar_url": "https://avatars.githubusercontent.com/u/28439529?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ysig", "html_url": "https://github.com/ysig", "followers_url": "https://api.github.com/users/ysig/followers", "following_url": "https://api.github.com/users/ysig/following{/other_user}", "gists_url": "https://api.github.com/users/ysig/gists{/gist_id}", "starred_url": "https://api.github.com/users/ysig/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ysig/subscriptions", "organizations_url": "https://api.github.com/users/ysig/orgs", "repos_url": "https://api.github.com/users/ysig/repos", "events_url": "https://api.github.com/users/ysig/events{/privacy}", "received_events_url": "https://api.github.com/users/ysig/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
2
2023-12-22T16:56:32
2024-02-02T00:05:04
null
NONE
null
null
null
I've spent a significant amount of time trying to locate the split object inside my _split_generators() custom function. Then after diving [in the code](https://github.com/huggingface/datasets/blob/5ff3670c18ed34fa8ddfa70a9aa403ae6cc9ad54/src/datasets/load.py#L2558) I realized that `download_and_prepare` is executed before! split is passed to the dataset builder in `as_dataset`. If I'm not missing something, this seems like bad design, for the following use case: > Imagine there is a huge dataset that has an evaluation test set and you want to just download and run just to compare your method. Is there a current workaround that can help me achieve the same result? Thank you,
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6529/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/huggingface/datasets/issues/6529/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6524
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6524/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6524/comments
https://api.github.com/repos/huggingface/datasets/issues/6524/events
https://github.com/huggingface/datasets/issues/6524
2,053,076,311
I_kwDODunzps56X3VX
6,524
Streaming the Pile: Missing Files
{ "login": "FelixLabelle", "id": 23347756, "node_id": "MDQ6VXNlcjIzMzQ3NzU2", "avatar_url": "https://avatars.githubusercontent.com/u/23347756?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FelixLabelle", "html_url": "https://github.com/FelixLabelle", "followers_url": "https://api.github.com/users/FelixLabelle/followers", "following_url": "https://api.github.com/users/FelixLabelle/following{/other_user}", "gists_url": "https://api.github.com/users/FelixLabelle/gists{/gist_id}", "starred_url": "https://api.github.com/users/FelixLabelle/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/FelixLabelle/subscriptions", "organizations_url": "https://api.github.com/users/FelixLabelle/orgs", "repos_url": "https://api.github.com/users/FelixLabelle/repos", "events_url": "https://api.github.com/users/FelixLabelle/events{/privacy}", "received_events_url": "https://api.github.com/users/FelixLabelle/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
1
2023-12-21T21:25:09
2023-12-22T09:17:05
2023-12-22T09:17:05
NONE
null
null
null
### Describe the bug The pile does not stream, a "File not Found error" is returned. It looks like the Pile's files have been moved. ### Steps to reproduce the bug To reproduce run the following code: ``` from datasets import load_dataset dataset = load_dataset('EleutherAI/pile', 'en', split='train', streaming=True) next(iter(dataset)) ``` I get the following error: `FileNotFoundError: https://the-eye.eu/public/AI/pile/train/00.jsonl.zst` ### Expected behavior Return the data in a stream. ### Environment info - `datasets` version: 2.12.0 - Platform: Windows-10-10.0.22621-SP0 - Python version: 3.11.5 - Huggingface_hub version: 0.15.1 - PyArrow version: 11.0.0 - Pandas version: 2.0.3
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6524/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6524/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6522
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6522/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6522/comments
https://api.github.com/repos/huggingface/datasets/issues/6522/events
https://github.com/huggingface/datasets/issues/6522
2,052,332,528
I_kwDODunzps56VBvw
6,522
Loading HF Hub Dataset (private org repo) fails to load all features
{ "login": "versipellis", "id": 6579034, "node_id": "MDQ6VXNlcjY1NzkwMzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/6579034?v=4", "gravatar_id": "", "url": "https://api.github.com/users/versipellis", "html_url": "https://github.com/versipellis", "followers_url": "https://api.github.com/users/versipellis/followers", "following_url": "https://api.github.com/users/versipellis/following{/other_user}", "gists_url": "https://api.github.com/users/versipellis/gists{/gist_id}", "starred_url": "https://api.github.com/users/versipellis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/versipellis/subscriptions", "organizations_url": "https://api.github.com/users/versipellis/orgs", "repos_url": "https://api.github.com/users/versipellis/repos", "events_url": "https://api.github.com/users/versipellis/events{/privacy}", "received_events_url": "https://api.github.com/users/versipellis/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2023-12-21T12:26:35
2023-12-21T13:24:31
null
NONE
null
null
null
### Describe the bug When pushing a `Dataset` with multiple `Features` (`input`, `output`, `tags`) to Huggingface Hub (private org repo), and later downloading the `Dataset`, only `input` and `output` load - I believe the expected behavior is for all `Features` to be loaded by default? ### Steps to reproduce the bug Pushing the data. `data_concat` is a `list` of `dict`s. ```python for datum in data_concat: datum_tags = {d["key"]: d["value"] for d in datum["tags"]} split_fraction = # some logic that generates a train/test split number if split_faction < test_fraction: data_test.append(datum) else: data_train.append(datum) dataset = DatasetDict( { "train": Dataset.from_list(data_train), "test": Dataset.from_list(data_test), "full": Dataset.from_list(data_concat), }, ) dataset_shuffled = dataset.shuffle(seed=shuffle_seed) dataset_shuffled.push_to_hub( repo_id=hf_repo_id, private=True, config_name=m, revision=revision, token=hf_token, ) ``` Loading it later: ```python dataset = datasets.load_dataset( path=hf_repo_id, name=name, token=hf_token, ) ``` Produces: ``` DatasetDict({ train: Dataset({ features: ['input', 'output'], num_rows: <obfuscated> }) test: Dataset({ features: ['input', 'output'], num_rows: <obfuscated> }) full: Dataset({ features: ['input', 'output'], num_rows: <obfuscated> }) }) ``` ### Expected behavior The expected result is below: ``` DatasetDict({ train: Dataset({ features: ['input', 'output', 'tags'], num_rows: <obfuscated> }) test: Dataset({ features: ['input', 'output', 'tags'], num_rows: <obfuscated> }) full: Dataset({ features: ['input', 'output', 'tags'], num_rows: <obfuscated> }) }) ``` My workaround is as follows: ```python dsinfo = datasets.get_dataset_config_info( path=data_files, config_name=data_config, token=hf_token, ) allfeatures = dsinfo.features.copy() if "tags" not in allfeatures: allfeatures["tags"] = [{"key": Value(dtype="string", id=None), "value": Value(dtype="string", id=None)}] dataset = datasets.load_dataset( path=data_files, name=data_config, features=allfeatures, token=hf_token, ) ``` Interestingly enough (and perhaps a related bug?), if I don't add the `tags` to `allfeatures` above (i.e. only loading `input` and `output`), it throws an error when executing `load_dataset`: ``` ValueError: Couldn't cast tags: list<element: struct<key: string, value: string>> child 0, element: struct<key: string, value: string> child 0, key: string child 1, value: string input: <obfuscated> output: <obfuscated> -- schema metadata -- huggingface: '{"info": {"features": {"tags": [{"key": {"dtype": "string",' + 532 to {'input': <obfuscated>, 'output': <obfuscated> because column names don't match ``` Traceback for this: ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/bt/github/core/.venv/lib/python3.11/site-packages/datasets/load.py", line 2152, in load_dataset builder_instance.download_and_prepare( File "/Users/bt/github/core/.venv/lib/python3.11/site-packages/datasets/builder.py", line 948, in download_and_prepare self._download_and_prepare( File "/Users/bt/github/core/.venv/lib/python3.11/site-packages/datasets/builder.py", line 1043, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/Users/bt/github/core/.venv/lib/python3.11/site-packages/datasets/builder.py", line 1805, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/Users/bt/github/core/.venv/lib/python3.11/site-packages/datasets/builder.py", line 1950, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.builder.DatasetGenerationError: An error occurred while generating the dataset ``` ### Environment info - `datasets` version: 2.15.0 - Platform: macOS-14.0-arm64-arm-64bit - Python version: 3.11.5 - `huggingface_hub` version: 0.19.4 - PyArrow version: 14.0.1 - Pandas version: 2.1.4 - `fsspec` version: 2023.10.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6522/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6522/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6521
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6521/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6521/comments
https://api.github.com/repos/huggingface/datasets/issues/6521/events
https://github.com/huggingface/datasets/issues/6521
2,052,229,538
I_kwDODunzps56Uomi
6,521
The order of the splits is not preserved
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
1
2023-12-21T11:17:27
2023-12-22T11:36:15
2023-12-22T11:36:15
MEMBER
null
null
null
We had a regression and the order of the splits is not preserved. They are alphabetically sorted, instead of preserving original "train", "validation", "test" order. Check: In branch "main" ```python In [9]: dataset = load_dataset("adversarial_qa", '"adversarialQA") In [10]: dataset Out[10]: DatasetDict({ test: Dataset({ features: ['id', 'title', 'context', 'question', 'answers', 'metadata'], num_rows: 3000 }) train: Dataset({ features: ['id', 'title', 'context', 'question', 'answers', 'metadata'], num_rows: 30000 }) validation: Dataset({ features: ['id', 'title', 'context', 'question', 'answers', 'metadata'], num_rows: 3000 }) }) ``` Before (2.15.0) it was: ```python DatasetDict({ train: Dataset({ features: ['id', 'title', 'context', 'question', 'answers', 'metadata'], num_rows: 30000 }) validation: Dataset({ features: ['id', 'title', 'context', 'question', 'answers', 'metadata'], num_rows: 3000 }) test: Dataset({ features: ['id', 'title', 'context', 'question', 'answers', 'metadata'], num_rows: 3000 }) }) ``` See issues: - https://huggingface.co/datasets/adversarial_qa/discussions/3 - https://huggingface.co/datasets/beans/discussions/4 This is a regression because it was previously fixed. See: - #6196 - #5728
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6521/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6521/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6517
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6517/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6517/comments
https://api.github.com/repos/huggingface/datasets/issues/6517/events
https://github.com/huggingface/datasets/issues/6517
2,050,121,588
I_kwDODunzps56Ml90
6,517
Bug get_metadata_patterns arg error
{ "login": "d710055071", "id": 12895488, "node_id": "MDQ6VXNlcjEyODk1NDg4", "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "gravatar_id": "", "url": "https://api.github.com/users/d710055071", "html_url": "https://github.com/d710055071", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url": "https://api.github.com/users/d710055071/gists{/gist_id}", "starred_url": "https://api.github.com/users/d710055071/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/d710055071/subscriptions", "organizations_url": "https://api.github.com/users/d710055071/orgs", "repos_url": "https://api.github.com/users/d710055071/repos", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "received_events_url": "https://api.github.com/users/d710055071/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-20T08:56:44
2023-12-22T00:24:23
2023-12-22T00:24:23
CONTRIBUTOR
null
null
null
https://github.com/huggingface/datasets/blob/3f149204a2a5948287adcade5e90707aa5207a92/src/datasets/load.py#L1240C1-L1240C69 metadata_patterns = get_metadata_patterns(base_path, download_config=self.download_config)
{ "login": "d710055071", "id": 12895488, "node_id": "MDQ6VXNlcjEyODk1NDg4", "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "gravatar_id": "", "url": "https://api.github.com/users/d710055071", "html_url": "https://github.com/d710055071", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url": "https://api.github.com/users/d710055071/gists{/gist_id}", "starred_url": "https://api.github.com/users/d710055071/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/d710055071/subscriptions", "organizations_url": "https://api.github.com/users/d710055071/orgs", "repos_url": "https://api.github.com/users/d710055071/repos", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "received_events_url": "https://api.github.com/users/d710055071/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6517/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6517/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6515
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6515/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6515/comments
https://api.github.com/repos/huggingface/datasets/issues/6515/events
https://github.com/huggingface/datasets/issues/6515
2,049,724,251
I_kwDODunzps56LE9b
6,515
Why call http_head() when fsspec_head() succeeds
{ "login": "d710055071", "id": 12895488, "node_id": "MDQ6VXNlcjEyODk1NDg4", "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "gravatar_id": "", "url": "https://api.github.com/users/d710055071", "html_url": "https://github.com/d710055071", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url": "https://api.github.com/users/d710055071/gists{/gist_id}", "starred_url": "https://api.github.com/users/d710055071/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/d710055071/subscriptions", "organizations_url": "https://api.github.com/users/d710055071/orgs", "repos_url": "https://api.github.com/users/d710055071/repos", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "received_events_url": "https://api.github.com/users/d710055071/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-20T02:25:51
2023-12-26T05:35:46
2023-12-26T05:35:46
CONTRIBUTOR
null
null
null
https://github.com/huggingface/datasets/blob/a91582de288d98e94bcb5ab634ca1cfeeff544c5/src/datasets/utils/file_utils.py#L510C1-L523C14
{ "login": "d710055071", "id": 12895488, "node_id": "MDQ6VXNlcjEyODk1NDg4", "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "gravatar_id": "", "url": "https://api.github.com/users/d710055071", "html_url": "https://github.com/d710055071", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url": "https://api.github.com/users/d710055071/gists{/gist_id}", "starred_url": "https://api.github.com/users/d710055071/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/d710055071/subscriptions", "organizations_url": "https://api.github.com/users/d710055071/orgs", "repos_url": "https://api.github.com/users/d710055071/repos", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "received_events_url": "https://api.github.com/users/d710055071/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6515/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6515/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6513
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6513/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6513/comments
https://api.github.com/repos/huggingface/datasets/issues/6513/events
https://github.com/huggingface/datasets/issues/6513
2,048,869,151
I_kwDODunzps56H0Mf
6,513
Support huggingface-hub 0.20.0
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-19T15:15:46
2023-12-20T08:44:45
2023-12-20T08:44:45
MEMBER
null
null
null
CI to test the support of `huggingface-hub` 0.20.0: https://github.com/huggingface/datasets/compare/main...ci-test-huggingface-hub-v0.20.0.rc1 We need to merge: - #6510 - #6512 - #6516
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6513/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6513/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6507
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6507/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6507/comments
https://api.github.com/repos/huggingface/datasets/issues/6507/events
https://github.com/huggingface/datasets/issues/6507
2,045,152,928
I_kwDODunzps555o6g
6,507
where is glue_metric.py> @Frankie123421 what was the resolution to this?
{ "login": "Mcccccc1024", "id": 119146162, "node_id": "U_kgDOBxoGsg", "avatar_url": "https://avatars.githubusercontent.com/u/119146162?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Mcccccc1024", "html_url": "https://github.com/Mcccccc1024", "followers_url": "https://api.github.com/users/Mcccccc1024/followers", "following_url": "https://api.github.com/users/Mcccccc1024/following{/other_user}", "gists_url": "https://api.github.com/users/Mcccccc1024/gists{/gist_id}", "starred_url": "https://api.github.com/users/Mcccccc1024/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Mcccccc1024/subscriptions", "organizations_url": "https://api.github.com/users/Mcccccc1024/orgs", "repos_url": "https://api.github.com/users/Mcccccc1024/repos", "events_url": "https://api.github.com/users/Mcccccc1024/events{/privacy}", "received_events_url": "https://api.github.com/users/Mcccccc1024/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-17T09:58:25
2023-12-18T11:42:49
2023-12-18T11:42:49
NONE
null
null
null
> @Frankie123421 what was the resolution to this? use glue_metric.py instead of glue.py in load_metric _Originally posted by @Frankie123421 in https://github.com/huggingface/datasets/issues/2117#issuecomment-905093763_
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6507/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6507/timeline
null
not_planned
false
https://api.github.com/repos/huggingface/datasets/issues/6506
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6506/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6506/comments
https://api.github.com/repos/huggingface/datasets/issues/6506/events
https://github.com/huggingface/datasets/issues/6506
2,044,975,038
I_kwDODunzps5549e-
6,506
Incorrect test set labels for RTE and CoLA datasets via load_dataset
{ "login": "emreonal11", "id": 73316684, "node_id": "MDQ6VXNlcjczMzE2Njg0", "avatar_url": "https://avatars.githubusercontent.com/u/73316684?v=4", "gravatar_id": "", "url": "https://api.github.com/users/emreonal11", "html_url": "https://github.com/emreonal11", "followers_url": "https://api.github.com/users/emreonal11/followers", "following_url": "https://api.github.com/users/emreonal11/following{/other_user}", "gists_url": "https://api.github.com/users/emreonal11/gists{/gist_id}", "starred_url": "https://api.github.com/users/emreonal11/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emreonal11/subscriptions", "organizations_url": "https://api.github.com/users/emreonal11/orgs", "repos_url": "https://api.github.com/users/emreonal11/repos", "events_url": "https://api.github.com/users/emreonal11/events{/privacy}", "received_events_url": "https://api.github.com/users/emreonal11/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-12-16T22:06:08
2023-12-21T09:57:57
2023-12-21T09:57:57
NONE
null
null
null
### Describe the bug The test set labels for the RTE and CoLA datasets when loading via datasets load_dataset are all -1. Edit: It appears this is also the case for every other dataset except for MRPC (stsb, sst2, qqp, mnli (both matched and mismatched), qnli, wnli, ax). Is this intended behavior to safeguard the test set for evaluation purposes? ### Steps to reproduce the bug !pip install datasets from datasets import load_dataset rte_data = load_dataset('glue', 'rte') cola_data = load_dataset('glue', 'cola') print(rte_data['test'][0:30]['label']) print(cola_data['test'][0:30]['label']) Output: [-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1] [-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1] The non-label test data seems to be fine: e.g. rte_data['test'][1] is: {'sentence1': "Authorities in Brazil say that more than 200 people are being held hostage in a prison in the country's remote, Amazonian-jungle state of Rondonia.", 'sentence2': 'Authorities in Brazil hold 200 people as hostage.', 'label': -1, 'idx': 1} Training and validation data are also fine: e.g. rte_data['train][0] is: {'sentence1': 'No Weapons of Mass Destruction Found in Iraq Yet.', 'sentence2': 'Weapons of Mass Destruction Found in Iraq.', 'label': 1, 'idx': 0} ### Expected behavior Expected the labels to be binary 0/1 values; Got all -1s instead ### Environment info - `datasets` version: 2.15.0 - Platform: Linux-6.1.58+-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.19.4 - PyArrow version: 10.0.1 - Pandas version: 1.5.3 - `fsspec` version: 2023.6.0
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6506/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6506/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6505
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6505/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6505/comments
https://api.github.com/repos/huggingface/datasets/issues/6505/events
https://github.com/huggingface/datasets/issues/6505
2,044,721,288
I_kwDODunzps553_iI
6,505
Got stuck when I trying to load a dataset
{ "login": "yirenpingsheng", "id": 18232551, "node_id": "MDQ6VXNlcjE4MjMyNTUx", "avatar_url": "https://avatars.githubusercontent.com/u/18232551?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yirenpingsheng", "html_url": "https://github.com/yirenpingsheng", "followers_url": "https://api.github.com/users/yirenpingsheng/followers", "following_url": "https://api.github.com/users/yirenpingsheng/following{/other_user}", "gists_url": "https://api.github.com/users/yirenpingsheng/gists{/gist_id}", "starred_url": "https://api.github.com/users/yirenpingsheng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yirenpingsheng/subscriptions", "organizations_url": "https://api.github.com/users/yirenpingsheng/orgs", "repos_url": "https://api.github.com/users/yirenpingsheng/repos", "events_url": "https://api.github.com/users/yirenpingsheng/events{/privacy}", "received_events_url": "https://api.github.com/users/yirenpingsheng/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
6
2023-12-16T11:51:07
2024-05-10T05:01:52
null
NONE
null
null
null
### Describe the bug Hello, everyone. I met a problem when I am trying to load a data file using load_dataset method on a Debian 10 system. The data file is not very large, only 1.63MB with 600 records. Here is my code: from datasets import load_dataset dataset = load_dataset('json', data_files='mypath/oaast_rm_zh.json') I waited it for 20 minutes. It still no response. I cannot using Ctrl+C to cancel the command. I have to use Ctrl+Z to kill it. I also try it with a txt file, it still no response in a long time. I can load the same file successfully using my laptop (windows 10, python 3.8.5, datasets==2.14.5). I can also make it on another computer (Ubuntu 20.04.5 LTS, python 3.10.13, datasets 2.14.7). It only takes me 1-2 miniutes. Could you give me some suggestions? Thank you. ### Steps to reproduce the bug from datasets import load_dataset dataset = load_dataset('json', data_files='mypath/oaast_rm_zh.json') ### Expected behavior I hope it can load the file successfully. ### Environment info OS: Debian GNU/Linux 10 Python: Python 3.10.13 Pip list: Package Version ------------------------- ------------ accelerate 0.25.0 addict 2.4.0 aiofiles 23.2.1 aiohttp 3.9.1 aiosignal 1.3.1 aliyun-python-sdk-core 2.14.0 aliyun-python-sdk-kms 2.16.2 altair 5.2.0 annotated-types 0.6.0 anyio 3.7.1 async-timeout 4.0.3 attrs 23.1.0 certifi 2023.11.17 cffi 1.16.0 charset-normalizer 3.3.2 click 8.1.7 contourpy 1.2.0 crcmod 1.7 cryptography 41.0.7 cycler 0.12.1 datasets 2.14.7 dill 0.3.7 docstring-parser 0.15 einops 0.7.0 exceptiongroup 1.2.0 fastapi 0.105.0 ffmpy 0.3.1 filelock 3.13.1 fonttools 4.46.0 frozenlist 1.4.1 fsspec 2023.10.0 gast 0.5.4 gradio 3.50.2 gradio_client 0.6.1 h11 0.14.0 httpcore 1.0.2 httpx 0.25.2 huggingface-hub 0.19.4 idna 3.6 importlib-metadata 7.0.0 importlib-resources 6.1.1 jieba 0.42.1 Jinja2 3.1.2 jmespath 0.10.0 joblib 1.3.2 jsonschema 4.20.0 jsonschema-specifications 2023.11.2 kiwisolver 1.4.5 markdown-it-py 3.0.0 MarkupSafe 2.1.3 matplotlib 3.8.2 mdurl 0.1.2 modelscope 1.10.0 mpmath 1.3.0 multidict 6.0.4 multiprocess 0.70.15 networkx 3.2.1 nltk 3.8.1 numpy 1.26.2 nvidia-cublas-cu12 12.1.3.1 nvidia-cuda-cupti-cu12 12.1.105 nvidia-cuda-nvrtc-cu12 12.1.105 nvidia-cuda-runtime-cu12 12.1.105 nvidia-cudnn-cu12 8.9.2.26 nvidia-cufft-cu12 11.0.2.54 nvidia-curand-cu12 10.3.2.106 nvidia-cusolver-cu12 11.4.5.107 nvidia-cusparse-cu12 12.1.0.106 nvidia-nccl-cu12 2.18.1 nvidia-nvjitlink-cu12 12.3.101 nvidia-nvtx-cu12 12.1.105 orjson 3.9.10 oss2 2.18.3 packaging 23.2 pandas 2.1.4 peft 0.7.1 Pillow 10.1.0 pip 23.3.1 platformdirs 4.1.0 protobuf 4.25.1 psutil 5.9.6 pyarrow 14.0.1 pyarrow-hotfix 0.6 pycparser 2.21 pycryptodome 3.19.0 pydantic 2.5.2 pydantic_core 2.14.5 pydub 0.25.1 Pygments 2.17.2 pyparsing 3.1.1 python-dateutil 2.8.2 python-multipart 0.0.6 pytz 2023.3.post1 PyYAML 6.0.1 referencing 0.32.0 regex 2023.10.3 requests 2.31.0 rich 13.7.0 rouge-chinese 1.0.3 rpds-py 0.13.2 safetensors 0.4.1 scipy 1.11.4 semantic-version 2.10.0 sentencepiece 0.1.99 setuptools 68.2.2 shtab 1.6.5 simplejson 3.19.2 six 1.16.0 sniffio 1.3.0 sortedcontainers 2.4.0 sse-starlette 1.8.2 starlette 0.27.0 sympy 1.12 tiktoken 0.5.2 tokenizers 0.15.0 tomli 2.0.1 toolz 0.12.0 torch 2.1.2 tqdm 4.66.1 transformers 4.36.1 triton 2.1.0 trl 0.7.4 typing_extensions 4.9.0 tyro 0.6.0 tzdata 2023.3 urllib3 2.1.0 uvicorn 0.24.0.post1 websockets 11.0.3 wheel 0.41.2 xxhash 3.4.1 yapf 0.40.2 yarl 1.9.4 zipp 3.17.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6505/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6505/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6504
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6504/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6504/comments
https://api.github.com/repos/huggingface/datasets/issues/6504/events
https://github.com/huggingface/datasets/issues/6504
2,044,541,154
I_kwDODunzps553Tji
6,504
Error Pushing to Hub
{ "login": "Jiayi-Pan", "id": 55055083, "node_id": "MDQ6VXNlcjU1MDU1MDgz", "avatar_url": "https://avatars.githubusercontent.com/u/55055083?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Jiayi-Pan", "html_url": "https://github.com/Jiayi-Pan", "followers_url": "https://api.github.com/users/Jiayi-Pan/followers", "following_url": "https://api.github.com/users/Jiayi-Pan/following{/other_user}", "gists_url": "https://api.github.com/users/Jiayi-Pan/gists{/gist_id}", "starred_url": "https://api.github.com/users/Jiayi-Pan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Jiayi-Pan/subscriptions", "organizations_url": "https://api.github.com/users/Jiayi-Pan/orgs", "repos_url": "https://api.github.com/users/Jiayi-Pan/repos", "events_url": "https://api.github.com/users/Jiayi-Pan/events{/privacy}", "received_events_url": "https://api.github.com/users/Jiayi-Pan/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-16T01:05:22
2023-12-16T06:20:53
2023-12-16T06:20:53
NONE
null
null
null
### Describe the bug Error when trying to push a dataset in a special format to hub ### Steps to reproduce the bug ``` import datasets from datasets import Dataset dataset_dict = { "filename": ["apple", "banana"], "token": [[[1,2],[3,4]],[[1,2],[3,4]]], "label": [0, 1], } dataset = Dataset.from_dict(dataset_dict) dataset = dataset.cast_column("token", datasets.features.features.Array2D(shape=(2, 2),dtype="int16")) dataset.push_to_hub("SequenceModel/imagenet_val_256") ``` Error: ``` ... ConstructorError: could not determine a constructor for the tag 'tag:yaml.org,2002:python/tuple' in "<unicode string>", line 8, column 16: shape: !!python/tuple ^ ``` ### Expected behavior Dataset being pushed to hub ### Environment info - `datasets` version: 2.15.0 - Platform: Linux-5.19.0-1022-gcp-x86_64-with-glibc2.35 - Python version: 3.11.5 - `huggingface_hub` version: 0.19.4 - PyArrow version: 14.0.1 - Pandas version: 2.1.4 - `fsspec` version: 2023.10.0
{ "login": "Jiayi-Pan", "id": 55055083, "node_id": "MDQ6VXNlcjU1MDU1MDgz", "avatar_url": "https://avatars.githubusercontent.com/u/55055083?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Jiayi-Pan", "html_url": "https://github.com/Jiayi-Pan", "followers_url": "https://api.github.com/users/Jiayi-Pan/followers", "following_url": "https://api.github.com/users/Jiayi-Pan/following{/other_user}", "gists_url": "https://api.github.com/users/Jiayi-Pan/gists{/gist_id}", "starred_url": "https://api.github.com/users/Jiayi-Pan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Jiayi-Pan/subscriptions", "organizations_url": "https://api.github.com/users/Jiayi-Pan/orgs", "repos_url": "https://api.github.com/users/Jiayi-Pan/repos", "events_url": "https://api.github.com/users/Jiayi-Pan/events{/privacy}", "received_events_url": "https://api.github.com/users/Jiayi-Pan/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6504/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6504/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6501
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6501/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6501/comments
https://api.github.com/repos/huggingface/datasets/issues/6501/events
https://github.com/huggingface/datasets/issues/6501
2,043,377,240
I_kwDODunzps55y3ZY
6,501
OverflowError: value too large to convert to int32_t
{ "login": "zhangfan-algo", "id": 47747764, "node_id": "MDQ6VXNlcjQ3NzQ3NzY0", "avatar_url": "https://avatars.githubusercontent.com/u/47747764?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zhangfan-algo", "html_url": "https://github.com/zhangfan-algo", "followers_url": "https://api.github.com/users/zhangfan-algo/followers", "following_url": "https://api.github.com/users/zhangfan-algo/following{/other_user}", "gists_url": "https://api.github.com/users/zhangfan-algo/gists{/gist_id}", "starred_url": "https://api.github.com/users/zhangfan-algo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zhangfan-algo/subscriptions", "organizations_url": "https://api.github.com/users/zhangfan-algo/orgs", "repos_url": "https://api.github.com/users/zhangfan-algo/repos", "events_url": "https://api.github.com/users/zhangfan-algo/events{/privacy}", "received_events_url": "https://api.github.com/users/zhangfan-algo/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2023-12-15T10:10:21
2023-12-15T10:10:21
null
NONE
null
null
null
### Describe the bug ![image](https://github.com/huggingface/datasets/assets/47747764/f58044fb-ddda-48b6-ba68-7bbfef781630) ### Steps to reproduce the bug just loading datasets ### Expected behavior how can I fix it ### Environment info pip install /mnt/cluster/zhangfan/study_info/LLaMA-Factory/peft-0.6.0-py3-none-any.whl pip install huggingface_hub-0.19.4-py3-none-any.whl tokenizers-0.15.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl transformers-4.36.1-py3-none-any.whl pyarrow_hotfix-0.6-py3-none-any.whl datasets-2.15.0-py3-none-any.whl tyro-0.5.18-py3-none-any.whl trl-0.7.4-py3-none-any.whl done
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6501/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6501/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6497
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6497/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6497/comments
https://api.github.com/repos/huggingface/datasets/issues/6497/events
https://github.com/huggingface/datasets/issues/6497
2,041,994,274
I_kwDODunzps55tlwi
6,497
Support setting a default config name in push_to_hub
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false } ]
null
0
2023-12-14T15:59:03
2023-12-18T11:50:04
2023-12-18T11:50:04
MEMBER
null
null
null
In order to convert script-datasets to no-script datasets, we need to support setting a default config name for those scripts that set one.
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "repos_url": "https://api.github.com/users/albertvillanova/repos", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6497/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6497/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6496
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6496/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6496/comments
https://api.github.com/repos/huggingface/datasets/issues/6496/events
https://github.com/huggingface/datasets/issues/6496
2,041,589,386
I_kwDODunzps55sC6K
6,496
Error when writing a dataset to HF Hub: A commit has happened since. Please refresh and try again.
{ "login": "GeorgesLorre", "id": 35808396, "node_id": "MDQ6VXNlcjM1ODA4Mzk2", "avatar_url": "https://avatars.githubusercontent.com/u/35808396?v=4", "gravatar_id": "", "url": "https://api.github.com/users/GeorgesLorre", "html_url": "https://github.com/GeorgesLorre", "followers_url": "https://api.github.com/users/GeorgesLorre/followers", "following_url": "https://api.github.com/users/GeorgesLorre/following{/other_user}", "gists_url": "https://api.github.com/users/GeorgesLorre/gists{/gist_id}", "starred_url": "https://api.github.com/users/GeorgesLorre/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/GeorgesLorre/subscriptions", "organizations_url": "https://api.github.com/users/GeorgesLorre/orgs", "repos_url": "https://api.github.com/users/GeorgesLorre/repos", "events_url": "https://api.github.com/users/GeorgesLorre/events{/privacy}", "received_events_url": "https://api.github.com/users/GeorgesLorre/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
1
2023-12-14T11:24:54
2023-12-14T12:22:21
null
NONE
null
null
null
**Describe the bug** Getting a `412 Client Error: Precondition Failed` when trying to write a dataset to the HF hub. ``` huggingface_hub.utils._errors.HfHubHTTPError: 412 Client Error: Precondition Failed for url: https://huggingface.co/api/datasets/GLorr/test-dask/commit/main (Request ID: Root=1-657ae26f-3bd92bf861bb254b2cc0826c;50a09ab7-9347-406a-ba49-69f98abee9cc) A commit has happened since. Please refresh and try again. ``` **Steps to reproduce the bug** This is a minimal reproducer: ``` import dask.dataframe as dd import pandas as pd import random import os import huggingface_hub import datasets huggingface_hub.login(token=os.getenv("HF_TOKEN")) data = {"number": [random.randint(0,10) for _ in range(1000)]} df = pd.DataFrame.from_dict(data) dataframe = dd.from_pandas(df, npartitions=1) dataframe = dataframe.repartition(npartitions=3) schema = datasets.Features({"number": datasets.Value("int64")}).arrow_schema repo_id = "GLorr/test-dask" repo_path = f"hf://datasets/{repo_id}" huggingface_hub.create_repo(repo_id=repo_id, repo_type="dataset", exist_ok=True) dd.to_parquet(dataframe, path=f"{repo_path}/data", schema=schema) ``` **Expected behavior** Would expect to write to the hub without any problem. **Environment info** ``` datasets==2.15.0 huggingface-hub==0.19.4 ```
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6496/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6496/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6494
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6494/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6494/comments
https://api.github.com/repos/huggingface/datasets/issues/6494/events
https://github.com/huggingface/datasets/issues/6494
2,039,684,839
I_kwDODunzps55kx7n
6,494
Image Data loaded Twice
{ "login": "ArcaneLex", "id": 28867010, "node_id": "MDQ6VXNlcjI4ODY3MDEw", "avatar_url": "https://avatars.githubusercontent.com/u/28867010?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ArcaneLex", "html_url": "https://github.com/ArcaneLex", "followers_url": "https://api.github.com/users/ArcaneLex/followers", "following_url": "https://api.github.com/users/ArcaneLex/following{/other_user}", "gists_url": "https://api.github.com/users/ArcaneLex/gists{/gist_id}", "starred_url": "https://api.github.com/users/ArcaneLex/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArcaneLex/subscriptions", "organizations_url": "https://api.github.com/users/ArcaneLex/orgs", "repos_url": "https://api.github.com/users/ArcaneLex/repos", "events_url": "https://api.github.com/users/ArcaneLex/events{/privacy}", "received_events_url": "https://api.github.com/users/ArcaneLex/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2023-12-13T13:11:42
2023-12-13T13:11:42
null
NONE
null
null
null
### Describe the bug ![1702472610561](https://github.com/huggingface/datasets/assets/28867010/4b7ef5e7-32c3-4b73-84cb-5de059caa0b6) When I learn from https://huggingface.co/docs/datasets/image_load and try to load image data from a folder. I noticed that the image was read twice in the returned data. As you can see in the attached image, there are only four images in the train folder, but reading brings up eight images ### Steps to reproduce the bug from datasets import Dataset, load_dataset dataset = load_dataset("imagefolder", data_dir="data/", drop_labels=False) # print(dataset["train"][0]["image"] == dataset["train"][1]["image"]) print(dataset) print(dataset["train"]["image"]) print(len(dataset["train"]["image"])) ### Expected behavior DatasetDict({ train: Dataset({ features: ['image', 'label'], num_rows: 8 }) }) [<PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=2877x2129 at 0x1BD1D1CA8B0>, <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=2877x2129 at 0x1BD1D2452E0>, <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=4208x3120 at 0x1BD1D245310>, <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=4208x3120 at 0x1BD1D2453A0>, <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=2877x2129 at 0x1BD1D245460>, <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=2877x2129 at 0x1BD1D245430>, <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=4208x3120 at 0x1BD1D2454F0>, <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=4208x3120 at 0x1BD1D245550>] 8 ### Environment info - `datasets` version: 2.14.5 - Platform: Windows-10-10.0.22621-SP0 - Python version: 3.9.17 - Huggingface_hub version: 0.19.4 - PyArrow version: 13.0.0 - Pandas version: 2.0.3
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6494/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6494/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6495
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6495/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6495/comments
https://api.github.com/repos/huggingface/datasets/issues/6495/events
https://github.com/huggingface/datasets/issues/6495
2,039,708,529
I_kwDODunzps55k3tx
6,495
Newline characters don't behave as expected when calling dataset.info
{ "login": "gerald-wrona", "id": 32300890, "node_id": "MDQ6VXNlcjMyMzAwODkw", "avatar_url": "https://avatars.githubusercontent.com/u/32300890?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gerald-wrona", "html_url": "https://github.com/gerald-wrona", "followers_url": "https://api.github.com/users/gerald-wrona/followers", "following_url": "https://api.github.com/users/gerald-wrona/following{/other_user}", "gists_url": "https://api.github.com/users/gerald-wrona/gists{/gist_id}", "starred_url": "https://api.github.com/users/gerald-wrona/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gerald-wrona/subscriptions", "organizations_url": "https://api.github.com/users/gerald-wrona/orgs", "repos_url": "https://api.github.com/users/gerald-wrona/repos", "events_url": "https://api.github.com/users/gerald-wrona/events{/privacy}", "received_events_url": "https://api.github.com/users/gerald-wrona/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
0
2023-12-12T23:07:51
2023-12-13T13:24:22
null
NONE
null
null
null
### System Info - `transformers` version: 4.32.1 - Platform: Windows-10-10.0.19045-SP0 - Python version: 3.11.5 - Huggingface_hub version: 0.15.1 - Safetensors version: 0.3.2 - Accelerate version: not installed - Accelerate config: not found - PyTorch version (GPU?): 2.1.1+cpu (False) - Tensorflow version (GPU?): 2.15.0 (False) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using GPU in script?: no - Using distributed or parallel set-up in script?: no ### Who can help? @marios ### Information - [X] The official example scripts - [ ] My own modified scripts ### Tasks - [X] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [ ] My own task or dataset (give details below) ### Reproduction [Source](https://huggingface.co/docs/datasets/v2.2.1/en/access) ``` from datasets import load_dataset dataset = load_dataset('glue', 'mrpc', split='train') dataset.info ``` DatasetInfo(description='GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n', citation='@inproceedings{dolan2005automatically,\n title={Automatically constructing a corpus of sentential paraphrases},\n author={Dolan, William B and Brockett, Chris},\n booktitle={Proceedings of the Third International Workshop on Paraphrasing (IWP2005)},\n year={2005}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n', homepage='https://www.microsoft.com/en-us/download/details.aspx?id=52398', license='', features={'sentence1': Value(dtype='string', id=None), 'sentence2': Value(dtype='string', id=None), 'label': ClassLabel(names=['not_equivalent', 'equivalent'], id=None), 'idx': Value(dtype='int32', id=None)}, post_processed=None, supervised_keys=None, task_templates=None, builder_name='glue', dataset_name=None, config_name='mrpc', version=1.0.0, splits={'train': SplitInfo(name='train', num_bytes=943843, num_examples=3668, shard_lengths=None, dataset_name='glue'), 'validation': SplitInfo(name='validation', num_bytes=105879, num_examples=408, shard_lengths=None, dataset_name='glue'), 'test': SplitInfo(name='test', num_bytes=442410, num_examples=1725, shard_lengths=None, dataset_name='glue')}, download_checksums={'https://dl.fbaipublicfiles.com/glue/data/mrpc_dev_ids.tsv': {'num_bytes': 6222, 'checksum': None}, 'https://dl.fbaipublicfiles.com/senteval/senteval_data/msr_paraphrase_train.txt': {'num_bytes': 1047044, 'checksum': None}, 'https://dl.fbaipublicfiles.com/senteval/senteval_data/msr_paraphrase_test.txt': {'num_bytes': 441275, 'checksum': None}}, download_size=1494541, post_processing_size=None, dataset_size=1492132, size_in_bytes=2986673) ### Expected behavior ``` from datasets import load_dataset dataset = load_dataset('glue', 'mrpc', split='train') dataset.info ``` DatasetInfo( description='GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n', citation='@inproceedings{dolan2005automatically,\n title={Automatically constructing a corpus of sentential paraphrases},\n author={Dolan, William B and Brockett, Chris},\n booktitle={Proceedings of the Third International Workshop on Paraphrasing (IWP2005)},\n year={2005}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n', homepage='https://www.microsoft.com/en-us/download/details.aspx?id=52398', license='', features={'sentence1': Value(dtype='string', id=None), 'sentence2': Value(dtype='string', id=None), 'label': ClassLabel(num_classes=2, names=['not_equivalent', 'equivalent'], names_file=None, id=None), 'idx': Value(dtype='int32', id=None)}, post_processed=None, supervised_keys=None, builder_name='glue', config_name='mrpc', version=1.0.0, splits={'train': SplitInfo(name='train', num_bytes=943851, num_examples=3668, dataset_name='glue'), 'validation': SplitInfo(name='validation', num_bytes=105887, num_examples=408, dataset_name='glue'), 'test': SplitInfo(name='test', num_bytes=442418, num_examples=1725, dataset_name='glue')}, download_checksums={'https://dl.fbaipublicfiles.com/glue/data/mrpc_dev_ids.tsv': {'num_bytes': 6222, 'checksum': '971d7767d81b997fd9060ade0ec23c4fc31cbb226a55d1bd4a1bac474eb81dc7'}, 'https://dl.fbaipublicfiles.com/senteval/senteval_data/msr_paraphrase_train.txt': {'num_bytes': 1047044, 'checksum': '60a9b09084528f0673eedee2b69cb941920f0b8cd0eeccefc464a98768457f89'}, 'https://dl.fbaipublicfiles.com/senteval/senteval_data/msr_paraphrase_test.txt': {'num_bytes': 441275, 'checksum': 'a04e271090879aaba6423d65b94950c089298587d9c084bf9cd7439bd785f784'}}, download_size=1494541, post_processing_size=None, dataset_size=1492156, size_in_bytes=2986697 )
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6495/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6495/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6490
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6490/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6490/comments
https://api.github.com/repos/huggingface/datasets/issues/6490/events
https://github.com/huggingface/datasets/issues/6490
2,037,204,892
I_kwDODunzps55bUec
6,490
`load_dataset(...,save_infos=True)` not working without loading script
{ "login": "morganveyret", "id": 114978051, "node_id": "U_kgDOBtptAw", "avatar_url": "https://avatars.githubusercontent.com/u/114978051?v=4", "gravatar_id": "", "url": "https://api.github.com/users/morganveyret", "html_url": "https://github.com/morganveyret", "followers_url": "https://api.github.com/users/morganveyret/followers", "following_url": "https://api.github.com/users/morganveyret/following{/other_user}", "gists_url": "https://api.github.com/users/morganveyret/gists{/gist_id}", "starred_url": "https://api.github.com/users/morganveyret/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/morganveyret/subscriptions", "organizations_url": "https://api.github.com/users/morganveyret/orgs", "repos_url": "https://api.github.com/users/morganveyret/repos", "events_url": "https://api.github.com/users/morganveyret/events{/privacy}", "received_events_url": "https://api.github.com/users/morganveyret/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
1
2023-12-12T08:09:18
2023-12-12T08:36:22
null
NONE
null
null
null
### Describe the bug It seems that saving a dataset infos back into the card file is not working for datasets without a loading script. After tracking the problem a bit it looks like saving the infos uses `Builder.get_imported_module_dir()` as its destination directory. Internally this is a call to `inspect.getfile()` but since the actual builder class used is dynamically created (cf. `datasets.load.configure_builder_class`) this method actually return te path to the parent builder class (e.g. `datasets.packaged_modules.json.JSON`). ### Steps to reproduce the bug 1. Have a local dataset without any loading script 2. Make sure there are no dataset infos in the README.md 3. Load with `save_infos=True` 4. No change in the dataset README.md 5. A new README.md file is created in the directory of the parent builder class (e.g. for json in `.../site-packages/datasets/packaged_modules/json/README.md`) ### Expected behavior The dataset README.md should be updated and no file should be created in the python environment. ### Environment info - `datasets` version: 2.15.0 - Platform: Linux-6.2.0-37-generic-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.19.4 - PyArrow version: 14.0.1 - Pandas version: 2.1.3 - `fsspec` version: 2023.6.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6490/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6490/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6489
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6489/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6489/comments
https://api.github.com/repos/huggingface/datasets/issues/6489/events
https://github.com/huggingface/datasets/issues/6489
2,036,743,777
I_kwDODunzps55Zj5h
6,489
load_dataset imageflder for aws s3 path
{ "login": "segalinc", "id": 9353106, "node_id": "MDQ6VXNlcjkzNTMxMDY=", "avatar_url": "https://avatars.githubusercontent.com/u/9353106?v=4", "gravatar_id": "", "url": "https://api.github.com/users/segalinc", "html_url": "https://github.com/segalinc", "followers_url": "https://api.github.com/users/segalinc/followers", "following_url": "https://api.github.com/users/segalinc/following{/other_user}", "gists_url": "https://api.github.com/users/segalinc/gists{/gist_id}", "starred_url": "https://api.github.com/users/segalinc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/segalinc/subscriptions", "organizations_url": "https://api.github.com/users/segalinc/orgs", "repos_url": "https://api.github.com/users/segalinc/repos", "events_url": "https://api.github.com/users/segalinc/events{/privacy}", "received_events_url": "https://api.github.com/users/segalinc/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
0
2023-12-12T00:08:43
2023-12-12T00:09:27
null
NONE
null
null
null
### Feature request I would like to load a dataset from S3 using the imagefolder option something like `dataset = datasets.load_dataset('imagefolder', data_dir='s3://.../lsun/train/bedroom', fs=S3FileSystem(), streaming=True) ` ### Motivation no need of data_files ### Your contribution no experience with this
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6489/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6489/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6488
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6488/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6488/comments
https://api.github.com/repos/huggingface/datasets/issues/6488/events
https://github.com/huggingface/datasets/issues/6488
2,035,899,898
I_kwDODunzps55WV36
6,488
429 Client Error
{ "login": "sasaadi", "id": 7882383, "node_id": "MDQ6VXNlcjc4ODIzODM=", "avatar_url": "https://avatars.githubusercontent.com/u/7882383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sasaadi", "html_url": "https://github.com/sasaadi", "followers_url": "https://api.github.com/users/sasaadi/followers", "following_url": "https://api.github.com/users/sasaadi/following{/other_user}", "gists_url": "https://api.github.com/users/sasaadi/gists{/gist_id}", "starred_url": "https://api.github.com/users/sasaadi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sasaadi/subscriptions", "organizations_url": "https://api.github.com/users/sasaadi/orgs", "repos_url": "https://api.github.com/users/sasaadi/repos", "events_url": "https://api.github.com/users/sasaadi/events{/privacy}", "received_events_url": "https://api.github.com/users/sasaadi/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
2
2023-12-11T15:06:01
2024-06-20T05:55:45
null
NONE
null
null
null
Hello, I was downloading the following dataset and after 20% of data was downloaded, I started getting error 429. It is not resolved since a few days. How should I resolve it? Thanks Dataset: https://huggingface.co/datasets/cerebras/SlimPajama-627B Error: `requests.exceptions.HTTPError: 429 Client Error: Too Many Requests for url: https://huggingface.co/datasets/cerebras/SlimPajama-627B/resolve/2d0accdd58c5d5511943ca1f5ff0e3eb5e293543/train/chunk1/example_train_3300.jsonl.zst`
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6488/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6488/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6485
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6485/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6485/comments
https://api.github.com/repos/huggingface/datasets/issues/6485/events
https://github.com/huggingface/datasets/issues/6485
2,035,141,884
I_kwDODunzps55Tcz8
6,485
FileNotFoundError: [Errno 2] No such file or directory: 'nul'
{ "login": "amanyara", "id": 73683903, "node_id": "MDQ6VXNlcjczNjgzOTAz", "avatar_url": "https://avatars.githubusercontent.com/u/73683903?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amanyara", "html_url": "https://github.com/amanyara", "followers_url": "https://api.github.com/users/amanyara/followers", "following_url": "https://api.github.com/users/amanyara/following{/other_user}", "gists_url": "https://api.github.com/users/amanyara/gists{/gist_id}", "starred_url": "https://api.github.com/users/amanyara/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amanyara/subscriptions", "organizations_url": "https://api.github.com/users/amanyara/orgs", "repos_url": "https://api.github.com/users/amanyara/repos", "events_url": "https://api.github.com/users/amanyara/events{/privacy}", "received_events_url": "https://api.github.com/users/amanyara/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-12-11T08:52:13
2023-12-14T08:09:08
2023-12-14T08:09:08
NONE
null
null
null
### Describe the bug it seems that sth wrong with my terrible "bug body" life, When i run this code, "import datasets" i meet this error FileNotFoundError: [Errno 2] No such file or directory: 'nul' ![image](https://github.com/huggingface/datasets/assets/73683903/3973c120-ebb1-42b7-bede-b9de053e861d) ![image](https://github.com/huggingface/datasets/assets/73683903/0496adff-a7a7-4dcb-929e-ec11ede71f04) ### Steps to reproduce the bug 1.import datasets ### Expected behavior i just run a single line code and stuct in this bug ### Environment info OS: Windows10 Datasets==2.15.0 python=3.10
{ "login": "amanyara", "id": 73683903, "node_id": "MDQ6VXNlcjczNjgzOTAz", "avatar_url": "https://avatars.githubusercontent.com/u/73683903?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amanyara", "html_url": "https://github.com/amanyara", "followers_url": "https://api.github.com/users/amanyara/followers", "following_url": "https://api.github.com/users/amanyara/following{/other_user}", "gists_url": "https://api.github.com/users/amanyara/gists{/gist_id}", "starred_url": "https://api.github.com/users/amanyara/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amanyara/subscriptions", "organizations_url": "https://api.github.com/users/amanyara/orgs", "repos_url": "https://api.github.com/users/amanyara/repos", "events_url": "https://api.github.com/users/amanyara/events{/privacy}", "received_events_url": "https://api.github.com/users/amanyara/received_events", "type": "User", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/6485/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/6485/timeline
null
completed
false