Dataset Viewer
Auto-converted to Parquet Duplicate
url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
3.28B
node_id
stringlengths
18
32
number
int64
1
7.71k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
listlengths
0
30
created_at
timestamp[us, tz=UTC]date
2020-04-14 10:18:02
2025-07-30 11:34:53
updated_at
timestamp[us, tz=UTC]date
2020-04-27 16:04:17
2025-07-31 05:22:35
closed_at
timestamp[us, tz=UTC]date
2020-04-14 12:01:40
2025-07-30 14:22:21
author_association
stringclasses
4 values
type
null
active_lock_reason
null
sub_issues_summary
dict
body
stringlengths
0
228k
closed_by
dict
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
4 values
draft
float64
0
1
pull_request
dict
created_at_dt
timestamp[us, tz=UTC]date
2020-04-14 10:18:02
2025-07-30 11:34:53
closed_at_dt
timestamp[us, tz=UTC]date
2020-04-14 12:01:40
2025-07-30 14:22:21
time_to_close
duration[us]
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/7700
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7700/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7700/comments
https://api.github.com/repos/huggingface/datasets/issues/7700/events
https://github.com/huggingface/datasets/issues/7700
3,263,922,255
I_kwDODunzps7Ci4BP
7,700
[doc] map.num_proc needs clarification
{ "avatar_url": "https://avatars.githubusercontent.com/u/196988264?v=4", "events_url": "https://api.github.com/users/sfc-gh-sbekman/events{/privacy}", "followers_url": "https://api.github.com/users/sfc-gh-sbekman/followers", "following_url": "https://api.github.com/users/sfc-gh-sbekman/following{/other_user}", ...
[]
open
false
null
[]
null
[]
2025-07-25T17:35:09
2025-07-25T17:39:36
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
https://huggingface.co/docs/datasets/v4.0.0/en/package_reference/main_classes#datasets.Dataset.map.num_proc ``` num_proc (int, optional, defaults to None) — Max number of processes when generating cache. Already cached shards are loaded sequentially. ``` for batch: ``` num_proc (int, optional, defaults to None): The n...
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7700/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7700/timeline
null
null
null
null
2025-07-25T17:35:09
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7706
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7706/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7706/comments
https://api.github.com/repos/huggingface/datasets/issues/7706/events
https://github.com/huggingface/datasets/pull/7706
3,271,129,240
PR_kwDODunzps6hC5uD
7,706
Reimplemented partial split download support (revival of #6832)
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gis...
[]
open
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": " Mario’s Patch (in PR #6832):\r\n```\r\ndef _make_split_generators_kwargs(self, prepare_split_kwargs):\r\n # Pass `pipeline` into `_split_generators()` from `prepare_split_kwargs` if\r\n # it's in the call signature of `_split_generators()`.\r\n # Thi...
2025-07-28T19:40:40
2025-07-29T09:25:12
null
CONTRIBUTOR
null
null
null
(revival of #6832) https://github.com/huggingface/datasets/pull/7648#issuecomment-3084050130 Close https://github.com/huggingface/datasets/issues/4101, and more --- ### PR under work!!!!
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7706/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7706/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7706.diff", "html_url": "https://github.com/huggingface/datasets/pull/7706", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7706.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7706" }
2025-07-28T19:40:40
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7709
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7709/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7709/comments
https://api.github.com/repos/huggingface/datasets/issues/7709/events
https://github.com/huggingface/datasets/issues/7709
3,276,677,990
I_kwDODunzps7DTiNm
7,709
Release 4.0.0 breaks usage patterns of with_format
{ "avatar_url": "https://avatars.githubusercontent.com/u/9154515?v=4", "events_url": "https://api.github.com/users/wittenator/events{/privacy}", "followers_url": "https://api.github.com/users/wittenator/followers", "following_url": "https://api.github.com/users/wittenator/following{/other_user}", "gists_url":...
[]
open
false
null
[]
null
[ { "author_association": "MEMBER", "body": "This is a breaking change with 4.0 which introduced `Column` objects. To get the numpy array from a `Column` you can `col[i]`, `col[i:j]` or even `col[:]` if you want the full column as a numpy array:\n\n```python\nfrom datasets import load_dataset\ndataset = load_...
2025-07-30T11:34:53
2025-07-30T15:41:59
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug Previously it was possible to access a whole column that was e.g. in numpy format via `with_format` by indexing the column. Now this possibility seems to be gone with the new Column() class. As far as I see, this makes working on a whole column (in-memory) more complex, i.e. normalizing an in-memo...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7709/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7709/timeline
null
null
null
null
2025-07-30T11:34:53
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7708
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7708/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7708/comments
https://api.github.com/repos/huggingface/datasets/issues/7708/events
https://github.com/huggingface/datasets/pull/7708
3,273,614,584
PR_kwDODunzps6hLVip
7,708
Concurrent push_to_hub
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
open
false
null
[]
null
[ { "author_association": "NONE", "body": "The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Fdatasets%2Fpr_7708). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "created_at": "2025-07-29T13:17:07Z", ...
2025-07-29T13:14:30
2025-07-30T15:55:00
null
MEMBER
null
null
null
Retry the step that (download + update + upload) the README.md using `create_commit(..., parent_commit=...)` if there was a commit in the meantime. This should enable concurrent `push_to_hub()` since it won't overwrite the README.md metadata anymore. DO NOT MERGE FOR NOW since it seems there is one bug that prevents...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7708/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7708/timeline
null
null
1
{ "diff_url": "https://github.com/huggingface/datasets/pull/7708.diff", "html_url": "https://github.com/huggingface/datasets/pull/7708", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7708.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7708" }
2025-07-29T13:14:30
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7705
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7705/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7705/comments
https://api.github.com/repos/huggingface/datasets/issues/7705/events
https://github.com/huggingface/datasets/issues/7705
3,269,070,499
I_kwDODunzps7C2g6j
7,705
Can Not read installed dataset in dataset.load(.)
{ "avatar_url": "https://avatars.githubusercontent.com/u/52521165?v=4", "events_url": "https://api.github.com/users/HuangChiEn/events{/privacy}", "followers_url": "https://api.github.com/users/HuangChiEn/followers", "following_url": "https://api.github.com/users/HuangChiEn/following{/other_user}", "gists_url"...
[]
open
false
null
[]
null
[ { "author_association": "MEMBER", "body": "You can download the dataset locally using [huggingface_hub.snapshot_download](https://huggingface.co/docs/huggingface_hub/v0.34.3/en/package_reference/file_download#huggingface_hub.snapshot_download) and then do\n\n```python\ndataset = load_dataset(local_directory...
2025-07-28T09:43:54
2025-07-30T15:44:26
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, folks, I'm newbie in huggingface dataset api. As title, i'm facing the issue that the dataset.load api can not connect to the installed dataset. code snippet : <img width="572" height="253" alt="Image" src="https://github.com/user-attachments/assets/10f48aaf-d6ca-4239-b1cf-145d74f125d1" /> data path : "/xxx/jose...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7705/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7705/timeline
null
null
null
null
2025-07-28T09:43:54
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7704
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7704/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7704/comments
https://api.github.com/repos/huggingface/datasets/issues/7704/events
https://github.com/huggingface/datasets/pull/7704
3,265,730,177
PR_kwDODunzps6gwtb8
7,704
Fix map() example in datasets documentation: define tokenizer before use
{ "avatar_url": "https://avatars.githubusercontent.com/u/183703408?v=4", "events_url": "https://api.github.com/users/Sanjaykumar030/events{/privacy}", "followers_url": "https://api.github.com/users/Sanjaykumar030/followers", "following_url": "https://api.github.com/users/Sanjaykumar030/following{/other_user}", ...
[]
open
false
null
[]
null
[ { "author_association": "NONE", "body": "Hi @lhoestq, just a gentle follow-up on this doc fix PR (#7704). Let me know if any changes are needed — happy to update.\r\nHope this improvement helps users run the example without confusion!", "created_at": "2025-08-01T13:48:35Z", "html_url": "https://gith...
2025-07-26T14:18:17
2025-07-26T14:18:17
null
NONE
null
null
null
## Problem The current datasets.Dataset.map() example in the documentation demonstrates batched processing using a tokenizer object without defining or importing it. This causes a NameError when users copy and run the example as-is, breaking the expected seamless experience. ## Correction This PR fixes the issue b...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7704/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7704/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7704.diff", "html_url": "https://github.com/huggingface/datasets/pull/7704", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7704.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7704" }
2025-07-26T14:18:17
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7701
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7701/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7701/comments
https://api.github.com/repos/huggingface/datasets/issues/7701/events
https://github.com/huggingface/datasets/pull/7701
3,265,236,296
PR_kwDODunzps6gvJ83
7,701
Update fsspec max version to current release 2025.7.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/5445560?v=4", "events_url": "https://api.github.com/users/rootAvish/events{/privacy}", "followers_url": "https://api.github.com/users/rootAvish/followers", "following_url": "https://api.github.com/users/rootAvish/following{/other_user}", "gists_url": "h...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "@lhoestq I ran the test suite locally and while some tests were failing those failures are present on the main branch too. Could you please review and trigger the CI?", "created_at": "2025-07-26T08:02:37Z", "html_url": "https://github.com/huggingface/d...
2025-07-26T06:47:59
2025-07-28T11:58:11
2025-07-28T11:58:11
CONTRIBUTOR
null
null
null
Diffusers currently asks for a max fsspec version of `2025.3.0`. This change updates it to the current latest version. This change is mainly required to resolve conflicts with other packages in an environment. In my particular case, `aider-chat` which is a part of my environment installs `2025.5.1` which is incompatibl...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7701/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7701/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7701.diff", "html_url": "https://github.com/huggingface/datasets/pull/7701", "merged_at": "2025-07-28T11:58:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/7701.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
2025-07-26T06:47:59
2025-07-28T11:58:11
2 days, 5:10:12
true
https://api.github.com/repos/huggingface/datasets/issues/7703
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7703/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7703/comments
https://api.github.com/repos/huggingface/datasets/issues/7703/events
https://github.com/huggingface/datasets/issues/7703
3,265,648,942
I_kwDODunzps7Cpdku
7,703
[Docs] map() example uses undefined `tokenizer` — causes NameError
{ "avatar_url": "https://avatars.githubusercontent.com/u/183703408?v=4", "events_url": "https://api.github.com/users/Sanjaykumar030/events{/privacy}", "followers_url": "https://api.github.com/users/Sanjaykumar030/followers", "following_url": "https://api.github.com/users/Sanjaykumar030/following{/other_user}", ...
[]
open
false
null
[]
null
[ { "author_association": "NONE", "body": "I've submitted PR #7704 which adds documentation to clarify the behavior of `map()` when returning `None`.", "created_at": "2025-07-27T05:39:18Z", "html_url": "https://github.com/huggingface/datasets/issues/7703#issuecomment-3124002704", "id": 3124002704,...
2025-07-26T13:35:11
2025-07-27T09:44:35
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
## Description The current documentation example for `datasets.Dataset.map()` demonstrates batched processing but uses a `tokenizer` object without defining or importing it. This causes an error every time it's copied. Here is the problematic line: ```python # process a batch of examples >>> ds = ds.map(lambda examp...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7703/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7703/timeline
null
null
null
null
2025-07-26T13:35:11
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7707
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7707/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7707/comments
https://api.github.com/repos/huggingface/datasets/issues/7707/events
https://github.com/huggingface/datasets/issues/7707
3,271,867,998
I_kwDODunzps7DBL5e
7,707
load_dataset() in 4.0.0 failed when decoding audio
{ "avatar_url": "https://avatars.githubusercontent.com/u/107918818?v=4", "events_url": "https://api.github.com/users/jiqing-feng/events{/privacy}", "followers_url": "https://api.github.com/users/jiqing-feng/followers", "following_url": "https://api.github.com/users/jiqing-feng/following{/other_user}", "gists_...
[]
open
false
null
[]
null
[ { "author_association": "NONE", "body": "Hi @lhoestq . Would you please have a look at it? I use the official NV Docker ([NV official docker image](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/pytorch): `nvcr.io/nvidia/pytorch:25.06-py3`) on A100 and encountered this issue, but I don't know how to ...
2025-07-29T03:25:03
2025-07-31T03:01:09
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug Cannot decode audio data. ### Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("hf-internal-testing/librispeech_asr_demo", "clean", split="validation") print(dataset[0]["audio"]["array"]) ``` 1st round run, got ``` File "/usr/local/lib/python3.1...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7707/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7707/timeline
null
null
null
null
2025-07-29T03:25:03
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7702
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7702/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7702/comments
https://api.github.com/repos/huggingface/datasets/issues/7702/events
https://github.com/huggingface/datasets/pull/7702
3,265,328,549
PR_kwDODunzps6gvdYC
7,702
num_proc=0 behave like None and clarify num_proc documentation in .map()
{ "avatar_url": "https://avatars.githubusercontent.com/u/84439872?v=4", "events_url": "https://api.github.com/users/tanuj-rai/events{/privacy}", "followers_url": "https://api.github.com/users/tanuj-rai/followers", "following_url": "https://api.github.com/users/tanuj-rai/following{/other_user}", "gists_url": "...
[]
open
false
null
[]
null
[ { "author_association": "MEMBER", "body": "I think we can support num_proc=0 and make it equivalent to `None` to make it simpler", "created_at": "2025-07-30T15:46:03Z", "html_url": "https://github.com/huggingface/datasets/pull/7702#issuecomment-3136895555", "id": 3136895555, "issue_url": "ht...
2025-07-26T08:19:39
2025-07-31T05:22:35
null
NONE
null
null
null
Fixes issue #7700 This PR makes num_proc=0 behave like None in Dataset.map(), disabling multiprocessing. It improves UX by aligning with DataLoader(num_workers=0) behavior. The num_proc docstring is also updated to clearly explain valid values and behavior. @SunMarc
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7702/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7702/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7702.diff", "html_url": "https://github.com/huggingface/datasets/pull/7702", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7702.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7702" }
2025-07-26T08:19:39
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7697
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7697/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7697/comments
https://api.github.com/repos/huggingface/datasets/issues/7697/events
https://github.com/huggingface/datasets/issues/7697
3,254,526,399
I_kwDODunzps7B_CG_
7,697
-
{ "avatar_url": "https://avatars.githubusercontent.com/u/44517413?v=4", "events_url": "https://api.github.com/users/kakamond/events{/privacy}", "followers_url": "https://api.github.com/users/kakamond/followers", "following_url": "https://api.github.com/users/kakamond/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[]
2025-07-23T01:30:32
2025-07-25T15:21:39
2025-07-25T15:21:39
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
-
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7697/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7697/timeline
null
completed
null
null
2025-07-23T01:30:32
2025-07-25T15:21:39
2 days, 13:51:07
false
https://api.github.com/repos/huggingface/datasets/issues/7698
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7698/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7698/comments
https://api.github.com/repos/huggingface/datasets/issues/7698/events
https://github.com/huggingface/datasets/issues/7698
3,255,350,916
I_kwDODunzps7CCLaE
7,698
NotImplementedError when using streaming=True in Google Colab environment
{ "avatar_url": "https://avatars.githubusercontent.com/u/100470741?v=4", "events_url": "https://api.github.com/users/Aniket17200/events{/privacy}", "followers_url": "https://api.github.com/users/Aniket17200/followers", "following_url": "https://api.github.com/users/Aniket17200/following{/other_user}", "gists_...
[]
open
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Hi, @Aniket17200, try upgrading datasets using '!pip install -U datasets'. I hope this will resolve your issue.", "created_at": "2025-07-23T13:46:33Z", "html_url": "https://github.com/huggingface/datasets/issues/7698#issuecomment-3108643682", "id":...
2025-07-23T08:04:53
2025-07-23T15:06:23
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug When attempting to load a large dataset (like tiiuae/falcon-refinedweb or allenai/c4) using streaming=True in a standard Google Colab notebook, the process fails with a NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet. This issue persists even after...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7698/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7698/timeline
null
null
null
null
2025-07-23T08:04:53
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7691
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7691/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7691/comments
https://api.github.com/repos/huggingface/datasets/issues/7691/events
https://github.com/huggingface/datasets/issues/7691
3,245,547,170
I_kwDODunzps7Bcx6i
7,691
Large WebDataset: pyarrow.lib.ArrowCapacityError on load() even with streaming
{ "avatar_url": "https://avatars.githubusercontent.com/u/122366389?v=4", "events_url": "https://api.github.com/users/cleong110/events{/privacy}", "followers_url": "https://api.github.com/users/cleong110/followers", "following_url": "https://api.github.com/users/cleong110/following{/other_user}", "gists_url": ...
[]
open
false
null
[]
null
[ { "author_association": "NONE", "body": "It seems the error occurs right here, as it tries to infer the Features: https://github.com/huggingface/datasets/blob/main/src/datasets/packaged_modules/webdataset/webdataset.py#L78-L90", "created_at": "2025-07-19T18:44:34Z", "html_url": "https://github.com/h...
2025-07-19T18:40:27
2025-07-25T08:51:10
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug I am creating a large WebDataset-format dataset for sign language processing research, and a number of the videos are over 2GB. The instant I hit one of the shards with one of those videos, I get a ArrowCapacityError, even with streaming. I made a config for the dataset that specifically inclu...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7691/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7691/timeline
null
null
null
null
2025-07-19T18:40:27
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7699
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7699/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7699/comments
https://api.github.com/repos/huggingface/datasets/issues/7699/events
https://github.com/huggingface/datasets/issues/7699
3,261,053,171
I_kwDODunzps7CX7jz
7,699
Broken link in documentation for "Create a video dataset"
{ "avatar_url": "https://avatars.githubusercontent.com/u/122366389?v=4", "events_url": "https://api.github.com/users/cleong110/events{/privacy}", "followers_url": "https://api.github.com/users/cleong110/followers", "following_url": "https://api.github.com/users/cleong110/following{/other_user}", "gists_url": ...
[]
open
false
null
[]
null
[ { "author_association": "MEMBER", "body": "The URL is ok but it seems the webdataset website is down. There seems to be a related issue here: https://github.com/webdataset/webdataset/issues/155\n\nFeel free to ask the authors there for an update. Otherwise happy to witch the link to the mirror shared in tha...
2025-07-24T19:46:28
2025-07-25T15:27:47
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
The link to "the [WebDataset documentation](https://webdataset.github.io/webdataset)." is broken. https://huggingface.co/docs/datasets/main/en/video_dataset#webdataset <img width="2048" height="264" alt="Image" src="https://github.com/user-attachments/assets/975dd10c-aad8-42fc-9fbc-de0e2747a326" />
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7699/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7699/timeline
null
null
null
null
2025-07-24T19:46:28
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7692
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7692/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7692/comments
https://api.github.com/repos/huggingface/datasets/issues/7692/events
https://github.com/huggingface/datasets/issues/7692
3,246,268,635
I_kwDODunzps7BfiDb
7,692
xopen: invalid start byte for streaming dataset with trust_remote_code=True
{ "avatar_url": "https://avatars.githubusercontent.com/u/5188731?v=4", "events_url": "https://api.github.com/users/sedol1339/events{/privacy}", "followers_url": "https://api.github.com/users/sedol1339/followers", "following_url": "https://api.github.com/users/sedol1339/following{/other_user}", "gists_url": "h...
[]
open
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hi ! it would be cool to convert this dataset to Parquet. This will make it work for `datasets>=4.0`, enable the Dataset Viewer and make it more reliable to load/stream (currently it uses a loading script in python and those are known for having issues sometimes)\n...
2025-07-20T11:08:20
2025-07-25T14:38:54
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug I am trying to load YODAS2 dataset with datasets==3.6.0 ``` from datasets import load_dataset next(iter(load_dataset('espnet/yodas2', name='ru000', split='train', streaming=True, trust_remote_code=True))) ``` And get `UnicodeDecodeError: 'utf-8' codec can't decode byte 0xa8 in position 1: invalid ...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7692/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7692/timeline
null
null
null
null
2025-07-20T11:08:20
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7694
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7694/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7694/comments
https://api.github.com/repos/huggingface/datasets/issues/7694/events
https://github.com/huggingface/datasets/issues/7694
3,247,600,408
I_kwDODunzps7BknMY
7,694
Dataset.to_json consumes excessive memory, appears to not be a streaming operation
{ "avatar_url": "https://avatars.githubusercontent.com/u/49603999?v=4", "events_url": "https://api.github.com/users/ycq0125/events{/privacy}", "followers_url": "https://api.github.com/users/ycq0125/followers", "following_url": "https://api.github.com/users/ycq0125/following{/other_user}", "gists_url": "https:...
[]
open
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hi ! to_json is memory efficient and writes the data by batch:\n\nhttps://github.com/huggingface/datasets/blob/d9861d86be222884dabbd534a2db770c70c9b558/src/datasets/io/json.py#L153-L159\n\nWhat memory are you mesuring ? If you are mesuring RSS, it is likely that it...
2025-07-21T07:51:25
2025-07-25T14:42:21
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug When exporting a Dataset object to a JSON Lines file using the .to_json(lines=True) method, the process consumes a very large amount of memory. The memory usage is proportional to the size of the entire Dataset object being saved, rather than being a low, constant memory operation. This behavior ...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7694/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7694/timeline
null
null
null
null
2025-07-21T07:51:25
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7695
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7695/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7695/comments
https://api.github.com/repos/huggingface/datasets/issues/7695/events
https://github.com/huggingface/datasets/pull/7695
3,251,904,843
PR_kwDODunzps6gB7jS
7,695
Support downloading specific splits in load_dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gis...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "I’ve completed the following steps to continue the partial split download support (from PR #6832):\r\n\r\nI did changes on top of what has been done by mario. Here are some of those changes: \r\n- Restored support for writing multiple split shards:\r\n\r\n- In...
2025-07-22T09:33:54
2025-07-28T17:33:30
2025-07-28T17:15:45
CONTRIBUTOR
null
null
null
This PR builds on #6832 by @mariosasko. May close - #4101, #2538 Discussion - https://github.com/huggingface/datasets/pull/7648#issuecomment-3084050130 --- ### Note - This PR is under work and frequent changes will be pushed.
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gis...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7695/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7695/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7695.diff", "html_url": "https://github.com/huggingface/datasets/pull/7695", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7695.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7695" }
2025-07-22T09:33:54
2025-07-28T17:15:45
6 days, 7:41:51
true
https://api.github.com/repos/huggingface/datasets/issues/7690
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7690/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7690/comments
https://api.github.com/repos/huggingface/datasets/issues/7690/events
https://github.com/huggingface/datasets/pull/7690
3,244,380,691
PR_kwDODunzps6fozag
7,690
HDF5 support
{ "avatar_url": "https://avatars.githubusercontent.com/u/17013474?v=4", "events_url": "https://api.github.com/users/klamike/events{/privacy}", "followers_url": "https://api.github.com/users/klamike/followers", "following_url": "https://api.github.com/users/klamike/following{/other_user}", "gists_url": "https:...
[]
open
false
null
[]
null
[ { "author_association": "NONE", "body": "A few to-dos which I think can be left for future PRs (which I am happy to do/help with -- just this one is already huge 😄 ):\r\n- [Enum types](https://docs.h5py.org/en/stable/special.html#enumerated-types)\r\n- HDF5 [io](https://github.com/huggingface/datasets/tree...
2025-07-18T21:09:41
2025-07-28T21:32:12
null
NONE
null
null
null
This PR adds support for tabular HDF5 file(s) by converting each row to an Arrow table. It supports columns with the usual dtypes including up to 5-dimensional arrays as well as support for complex/compound types by splitting them into several columns. All datasets within the HDF5 file should have rows on the first dim...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7690/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7690/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7690.diff", "html_url": "https://github.com/huggingface/datasets/pull/7690", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7690.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7690" }
2025-07-18T21:09:41
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7693
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7693/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7693/comments
https://api.github.com/repos/huggingface/datasets/issues/7693/events
https://github.com/huggingface/datasets/issues/7693
3,246,369,678
I_kwDODunzps7Bf6uO
7,693
Dataset scripts are no longer supported, but found superb.py
{ "avatar_url": "https://avatars.githubusercontent.com/u/114297534?v=4", "events_url": "https://api.github.com/users/edwinzajac/events{/privacy}", "followers_url": "https://api.github.com/users/edwinzajac/followers", "following_url": "https://api.github.com/users/edwinzajac/following{/other_user}", "gists_url...
[]
open
false
null
[]
null
[ { "author_association": "NONE", "body": "I got a pretty similar issue when I try to load bigbio/neurotrial_ner dataset. \n`Dataset scripts are no longer supported, but found neurotrial_ner.py`", "created_at": "2025-07-21T14:10:07Z", "html_url": "https://github.com/huggingface/datasets/issues/7693#is...
2025-07-20T13:48:06
2025-07-30T15:01:03
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug Hello, I'm trying to follow the [Hugging Face Pipelines tutorial](https://huggingface.co/docs/transformers/main_classes/pipelines) but the tutorial seems to work only on old datasets versions. I then get the error : ``` -------------------------------------------------------------------------- ...
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7693/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7693/timeline
null
null
null
null
2025-07-20T13:48:06
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7696
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7696/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7696/comments
https://api.github.com/repos/huggingface/datasets/issues/7696/events
https://github.com/huggingface/datasets/issues/7696
3,253,433,350
I_kwDODunzps7B63QG
7,696
load_dataset() in 4.0.0 returns different audio samples compared to earlier versions breaking reproducibility
{ "avatar_url": "https://avatars.githubusercontent.com/u/25346345?v=4", "events_url": "https://api.github.com/users/Manalelaidouni/events{/privacy}", "followers_url": "https://api.github.com/users/Manalelaidouni/followers", "following_url": "https://api.github.com/users/Manalelaidouni/following{/other_user}", ...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hi ! This is because `datasets` now uses the FFmpeg-based library `torchcodec` instead of the libsndfile-based library `soundfile` to decode audio data. Those two have different decoding implementations", "created_at": "2025-07-25T14:27:36Z", "html_url": "h...
2025-07-22T17:02:17
2025-07-30T14:22:21
2025-07-30T14:22:21
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug In datasets 4.0.0 release, `load_dataset()` returns different audio samples compared to earlier versions, this breaks integration tests that depend on consistent sample data across different environments (first and second envs specified below). ### Steps to reproduce the bug ```python from dat...
{ "avatar_url": "https://avatars.githubusercontent.com/u/25346345?v=4", "events_url": "https://api.github.com/users/Manalelaidouni/events{/privacy}", "followers_url": "https://api.github.com/users/Manalelaidouni/followers", "following_url": "https://api.github.com/users/Manalelaidouni/following{/other_user}", ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7696/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7696/timeline
null
completed
null
null
2025-07-22T17:02:17
2025-07-30T14:22:21
7 days, 21:20:04
false
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
7