Dataset Viewer
Auto-converted to Parquet Duplicate
url
string
repository_url
string
labels_url
string
comments_url
string
events_url
string
html_url
string
id
int64
node_id
string
number
int64
title
string
user
dict
labels
list
state
string
locked
bool
assignee
dict
assignees
list
milestone
dict
comments
list
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
string
type
null
active_lock_reason
null
sub_issues_summary
dict
issue_dependencies_summary
dict
body
string
closed_by
dict
reactions
dict
timeline_url
string
performed_via_github_app
null
state_reason
string
draft
bool
pull_request
dict
is_pull_request
bool
https://api.github.com/repos/huggingface/datasets/issues/7813
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7813/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7813/comments
https://api.github.com/repos/huggingface/datasets/issues/7813/events
https://github.com/huggingface/datasets/issues/7813
3,503,446,288
I_kwDODunzps7Q0lkQ
7,813
Caching does not work when using python3.14
{ "login": "intexcor", "id": 142020129, "node_id": "U_kgDOCHcOIQ", "avatar_url": "https://avatars.githubusercontent.com/u/142020129?v=4", "gravatar_id": "", "url": "https://api.github.com/users/intexcor", "html_url": "https://github.com/intexcor", "followers_url": "https://api.github.com/users/intexcor/followers", "following_url": "https://api.github.com/users/intexcor/following{/other_user}", "gists_url": "https://api.github.com/users/intexcor/gists{/gist_id}", "starred_url": "https://api.github.com/users/intexcor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/intexcor/subscriptions", "organizations_url": "https://api.github.com/users/intexcor/orgs", "repos_url": "https://api.github.com/users/intexcor/repos", "events_url": "https://api.github.com/users/intexcor/events{/privacy}", "received_events_url": "https://api.github.com/users/intexcor/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "https://github.com/uqfoundation/dill/issues/725" ]
2025-10-10T15:36:46
2025-10-10T15:47:32
null
NONE
null
null
{ "total": 0, "completed": 0, "percent_completed": 0 }
{ "blocked_by": 0, "total_blocked_by": 0, "blocking": 0, "total_blocking": 0 }
### Describe the bug Traceback (most recent call last): File "/workspace/ctn.py", line 8, in <module> ds = load_dataset(f"naver-clova-ix/synthdog-{lang}") # или "synthdog-zh" для китайского File "/workspace/.venv/lib/python3.14/site-packages/datasets/load.py", line 1397, in load_dataset builder_instance = load_dataset_builder( path=path, ...<10 lines>... **config_kwargs, ) File "/workspace/.venv/lib/python3.14/site-packages/datasets/load.py", line 1185, in load_dataset_builder builder_instance._use_legacy_cache_dir_if_possible(dataset_module) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/builder.py", line 612, in _use_legacy_cache_dir_if_possible self._check_legacy_cache2(dataset_module) or self._check_legacy_cache() or None ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/builder.py", line 485, in _check_legacy_cache2 config_id = self.config.name + "-" + Hasher.hash({"data_files": self.config.data_files}) ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/fingerprint.py", line 188, in hash return cls.hash_bytes(dumps(value)) ~~~~~^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 120, in dumps dump(obj, file) ~~~~^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 114, in dump Pickler(file, recurse=True).dump(obj) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/dill/_dill.py", line 428, in dump StockPickler.dump(self, obj) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^ File "/usr/lib/python3.14/pickle.py", line 498, in dump self.save(obj) ~~~~~~~~~^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 70, in save dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/dill/_dill.py", line 422, in save StockPickler.save(self, obj, save_persistent_id) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.14/pickle.py", line 572, in save f(self, obj) # Call unbound method with explicit self ~^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/dill/_dill.py", line 1262, in save_module_dict StockPickler.save_dict(pickler, obj) ~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^ File "/usr/lib/python3.14/pickle.py", line 1064, in save_dict self._batch_setitems(obj.items(), obj) ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^ TypeError: Pickler._batch_setitems() takes 2 positional arguments but 3 were given ### Steps to reproduce the bug ds_train = ds["train"].map(lambda x: {**x, "lang": lang}) ### Expected behavior Fixed bugs ### Environment info - `datasets` version: 4.2.0 - Platform: Linux-6.8.0-85-generic-x86_64-with-glibc2.39 - Python version: 3.14.0 - `huggingface_hub` version: 0.35.3 - PyArrow version: 21.0.0 - Pandas version: 2.3.3 - `fsspec` version: 2025.9.0
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7813/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7813/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7812
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7812/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7812/comments
https://api.github.com/repos/huggingface/datasets/issues/7812/events
https://github.com/huggingface/datasets/pull/7812
3,500,901,422
PR_kwDODunzps6s_New
7,812
docs: document_dataset PDFs & OCR
{ "login": "ethanknights", "id": 34215814, "node_id": "MDQ6VXNlcjM0MjE1ODE0", "avatar_url": "https://avatars.githubusercontent.com/u/34215814?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ethanknights", "html_url": "https://github.com/ethanknights", "followers_url": "https://api.github.com/users/ethanknights/followers", "following_url": "https://api.github.com/users/ethanknights/following{/other_user}", "gists_url": "https://api.github.com/users/ethanknights/gists{/gist_id}", "starred_url": "https://api.github.com/users/ethanknights/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ethanknights/subscriptions", "organizations_url": "https://api.github.com/users/ethanknights/orgs", "repos_url": "https://api.github.com/users/ethanknights/repos", "events_url": "https://api.github.com/users/ethanknights/events{/privacy}", "received_events_url": "https://api.github.com/users/ethanknights/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[]
2025-10-09T23:31:41
2025-10-09T23:31:41
null
NONE
null
null
null
null
Use acronyms consistently across document_dataset docs.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7812/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7812/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7812", "html_url": "https://github.com/huggingface/datasets/pull/7812", "diff_url": "https://github.com/huggingface/datasets/pull/7812.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7812.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/7811
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7811/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7811/comments
https://api.github.com/repos/huggingface/datasets/issues/7811/events
https://github.com/huggingface/datasets/issues/7811
3,500,741,658
I_kwDODunzps7QqRQa
7,811
SIGSEGV when Python exits due to near null deref
{ "login": "iankronquist", "id": 5192353, "node_id": "MDQ6VXNlcjUxOTIzNTM=", "avatar_url": "https://avatars.githubusercontent.com/u/5192353?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iankronquist", "html_url": "https://github.com/iankronquist", "followers_url": "https://api.github.com/users/iankronquist/followers", "following_url": "https://api.github.com/users/iankronquist/following{/other_user}", "gists_url": "https://api.github.com/users/iankronquist/gists{/gist_id}", "starred_url": "https://api.github.com/users/iankronquist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iankronquist/subscriptions", "organizations_url": "https://api.github.com/users/iankronquist/orgs", "repos_url": "https://api.github.com/users/iankronquist/repos", "events_url": "https://api.github.com/users/iankronquist/events{/privacy}", "received_events_url": "https://api.github.com/users/iankronquist/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "The issue seems to come from `dill` which is a `datasets` dependency, e.g. this segfaults:\n\n```python\nimport dill\nfrom tqdm import tqdm\nprogress_bar = tqdm(total=(1000), unit='cols', desc='cols ')\nprogress_bar.update(1)\n```\n\n`tqdm` seems to segfault when `dill` is imported. I only found this about segfaul...
2025-10-09T22:00:11
2025-10-10T22:09:24
null
NONE
null
null
{ "total": 0, "completed": 0, "percent_completed": 0 }
{ "blocked_by": 0, "total_blocked_by": 0, "blocking": 0, "total_blocking": 0 }
### Describe the bug When I run the following python script using datasets I get a segfault. ```python from datasets import load_dataset from tqdm import tqdm progress_bar = tqdm(total=(1000), unit='cols', desc='cols ') progress_bar.update(1) ``` ``` % lldb -- python3 crashmin.py (lldb) target create "python3" Current executable set to '/Users/ian/bug/venv/bin/python3' (arm64). (lldb) settings set -- target.run-args "crashmin.py" (lldb) r Process 8095 launched: '/Users/ian/bug/venv/bin/python3' (arm64) Process 8095 stopped * thread #2, stop reason = exec frame #0: 0x0000000100014b30 dyld`_dyld_start dyld`_dyld_start: -> 0x100014b30 <+0>: mov x0, sp 0x100014b34 <+4>: and sp, x0, #0xfffffffffffffff0 0x100014b38 <+8>: mov x29, #0x0 ; =0 Target 0: (Python) stopped. (lldb) c Process 8095 resuming cols : 0% 0/1000 [00:00<?, ?cols/s]Process 8095 stopped * thread #2, queue = 'com.apple.main-thread', stop reason = EXC_BAD_ACCESS (code=1, address=0x10) frame #0: 0x0000000101783454 _datetime.cpython-313-darwin.so`delta_new + 188 _datetime.cpython-313-darwin.so`delta_new: -> 0x101783454 <+188>: ldr x3, [x20, #0x10] 0x101783458 <+192>: adrp x0, 10 0x10178345c <+196>: add x0, x0, #0x6fc ; "seconds" Target 0: (Python) stopped. (lldb) bt * thread #2, queue = 'com.apple.main-thread', stop reason = EXC_BAD_ACCESS (code=1, address=0x10) * frame #0: 0x0000000101783454 _datetime.cpython-313-darwin.so`delta_new + 188 frame #1: 0x0000000100704b60 Python`type_call + 96 frame #2: 0x000000010067ba34 Python`_PyObject_MakeTpCall + 120 frame #3: 0x00000001007aae3c Python`_PyEval_EvalFrameDefault + 30236 frame #4: 0x000000010067c900 Python`PyObject_CallOneArg + 112 frame #5: 0x000000010070f0a0 Python`slot_tp_finalize + 116 frame #6: 0x000000010070c3b4 Python`subtype_dealloc + 788 frame #7: 0x00000001006c378c Python`insertdict + 756 frame #8: 0x00000001006db2b0 Python`_PyModule_ClearDict + 660 frame #9: 0x000000010080a9a8 Python`finalize_modules + 1772 frame #10: 0x0000000100809a44 Python`_Py_Finalize + 264 frame #11: 0x0000000100837630 Python`Py_RunMain + 252 frame #12: 0x0000000100837ef8 Python`pymain_main + 304 frame #13: 0x0000000100837f98 Python`Py_BytesMain + 40 frame #14: 0x000000019cfcc274 dyld`start + 2840 (lldb) register read x20 x20 = 0x0000000000000000 (lldb) ``` ### Steps to reproduce the bug Run the script above, and observe the segfault. ### Expected behavior No segfault ### Environment info ``` % pip freeze datasets | grep -i datasets datasets==4.2.0 (venv) 0 ~/bug 14:58:06 % pip freeze tqdm | grep -i tqdm tqdm==4.67.1 (venv) 0 ~/bug 14:58:16 % python --version Python 3.13.7 ```
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7811/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7811/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7810
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7810/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7810/comments
https://api.github.com/repos/huggingface/datasets/issues/7810/events
https://github.com/huggingface/datasets/pull/7810
3,499,855,569
PR_kwDODunzps6s7wHa
7,810
fix conda deps
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7810). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T16:32:04
2025-10-09T16:35:15
2025-10-09T16:35:14
MEMBER
null
null
null
null
null
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7810/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7810/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7810", "html_url": "https://github.com/huggingface/datasets/pull/7810", "diff_url": "https://github.com/huggingface/datasets/pull/7810.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7810.patch", "merged_at": "2025-10-09T16:35:14" }
true
https://api.github.com/repos/huggingface/datasets/issues/7809
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7809/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7809/comments
https://api.github.com/repos/huggingface/datasets/issues/7809/events
https://github.com/huggingface/datasets/pull/7809
3,499,811,179
PR_kwDODunzps6s7mwb
7,809
Set dev version
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7809). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T16:19:19
2025-10-09T16:22:12
2025-10-09T16:19:31
MEMBER
null
null
null
null
null
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7809/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7809/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7809", "html_url": "https://github.com/huggingface/datasets/pull/7809", "diff_url": "https://github.com/huggingface/datasets/pull/7809.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7809.patch", "merged_at": "2025-10-09T16:19:31" }
true
https://api.github.com/repos/huggingface/datasets/issues/7808
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7808/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7808/comments
https://api.github.com/repos/huggingface/datasets/issues/7808/events
https://github.com/huggingface/datasets/pull/7808
3,499,779,993
PR_kwDODunzps6s7gBq
7,808
release: 4.2.0
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7808). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T16:10:53
2025-10-09T16:21:01
2025-10-09T16:11:08
MEMBER
null
null
null
null
null
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7808/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7808/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7808", "html_url": "https://github.com/huggingface/datasets/pull/7808", "diff_url": "https://github.com/huggingface/datasets/pull/7808.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7808.patch", "merged_at": "2025-10-09T16:11:08" }
true
https://api.github.com/repos/huggingface/datasets/issues/7807
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7807/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7807/comments
https://api.github.com/repos/huggingface/datasets/issues/7807/events
https://github.com/huggingface/datasets/pull/7807
3,499,765,725
PR_kwDODunzps6s7c_U
7,807
typo
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7807). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T16:06:47
2025-10-09T16:16:31
2025-10-09T16:06:58
MEMBER
null
null
null
null
add an s to be consistent with pandas' on_bad_lines
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7807/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7807/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7807", "html_url": "https://github.com/huggingface/datasets/pull/7807", "diff_url": "https://github.com/huggingface/datasets/pull/7807.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7807.patch", "merged_at": "2025-10-09T16:06:58" }
true
https://api.github.com/repos/huggingface/datasets/issues/7806
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7806/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7806/comments
https://api.github.com/repos/huggingface/datasets/issues/7806/events
https://github.com/huggingface/datasets/pull/7806
3,499,483,246
PR_kwDODunzps6s6gnr
7,806
Parquet: add `on_bad_file` argument to error/warn/skip bad files
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7806). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T14:41:46
2025-10-09T16:04:35
2025-10-09T16:04:33
MEMBER
null
null
null
null
```python from datasets import load_dataset on_bad_file = "error" # default # on_bad_file = "warn" # warn and skip bad file # on_bad_file = "skip" # skip bad file ds = load_dataset(parquet_dataset_id, on_bad_file=on_bad_file) ```
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7806/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7806/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7806", "html_url": "https://github.com/huggingface/datasets/pull/7806", "diff_url": "https://github.com/huggingface/datasets/pull/7806.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7806.patch", "merged_at": "2025-10-09T16:04:33" }
true
https://api.github.com/repos/huggingface/datasets/issues/7805
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7805/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7805/comments
https://api.github.com/repos/huggingface/datasets/issues/7805/events
https://github.com/huggingface/datasets/pull/7805
3,499,286,947
PR_kwDODunzps6s52Ew
7,805
Less api calls when resolving data_files
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7805). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T13:53:06
2025-10-09T14:01:57
2025-10-09T14:01:56
MEMBER
null
null
null
null
There ~10 unnecessary `/api/datasets/user/dataset/revision`calls due to multithreading in data files resolution. I disabled multithreading, which was actually not useful anymore since `HfFileSystem` has been using caching for a while now.
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7805/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7805/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7805", "html_url": "https://github.com/huggingface/datasets/pull/7805", "diff_url": "https://github.com/huggingface/datasets/pull/7805.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7805.patch", "merged_at": "2025-10-09T14:01:55" }
true
https://api.github.com/repos/huggingface/datasets/issues/7804
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7804/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7804/comments
https://api.github.com/repos/huggingface/datasets/issues/7804/events
https://github.com/huggingface/datasets/issues/7804
3,498,534,596
I_kwDODunzps7Qh2bE
7,804
Support scientific data formats
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "Please add the support for `Zarr`! That's what we use in the Bioimaging community. It is crucial, because raw upload of a *single* bio image can take _terrabytes in memory_!\n\nThe python library would be `bioio` or `zarr`:\n- [ ] Zarr: `bioio` or `zarr`\n\nSee a [Zarr example](https://ome.github.io/ome-ngff-valid...
2025-10-09T10:18:24
2025-10-10T11:26:23
null
MEMBER
null
null
{ "total": 0, "completed": 0, "percent_completed": 0 }
{ "blocked_by": 0, "total_blocked_by": 0, "blocking": 0, "total_blocking": 0 }
List of formats and libraries we can use to load the data in `datasets`: - [ ] DICOMs: pydicom - [ ] NIfTIs: nibabel - [ ] WFDB: wfdb cc @zaRizk7 for viz Feel free to comment / suggest other formats and libs you'd like to see or to share your interest in one of the mentioned format
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7804/reactions", "total_count": 6, "+1": 1, "-1": 0, "laugh": 0, "hooray": 2, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7804/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7803
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7803/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7803/comments
https://api.github.com/repos/huggingface/datasets/issues/7803/events
https://github.com/huggingface/datasets/pull/7803
3,498,395,879
PR_kwDODunzps6s2zyO
7,803
More Parquet streaming docs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7803). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T09:39:11
2025-10-09T10:01:46
2025-10-09T10:01:43
MEMBER
null
null
null
null
null
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7803/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7803/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7803", "html_url": "https://github.com/huggingface/datasets/pull/7803", "diff_url": "https://github.com/huggingface/datasets/pull/7803.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7803.patch", "merged_at": "2025-10-09T10:01:43" }
true
https://api.github.com/repos/huggingface/datasets/issues/7802
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7802/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7802/comments
https://api.github.com/repos/huggingface/datasets/issues/7802/events
https://github.com/huggingface/datasets/issues/7802
3,497,454,119
I_kwDODunzps7Qduon
7,802
[Docs] Missing documentation for `Dataset.from_dict`
{ "login": "aaronshenhao", "id": 69421545, "node_id": "MDQ6VXNlcjY5NDIxNTQ1", "avatar_url": "https://avatars.githubusercontent.com/u/69421545?v=4", "gravatar_id": "", "url": "https://api.github.com/users/aaronshenhao", "html_url": "https://github.com/aaronshenhao", "followers_url": "https://api.github.com/users/aaronshenhao/followers", "following_url": "https://api.github.com/users/aaronshenhao/following{/other_user}", "gists_url": "https://api.github.com/users/aaronshenhao/gists{/gist_id}", "starred_url": "https://api.github.com/users/aaronshenhao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aaronshenhao/subscriptions", "organizations_url": "https://api.github.com/users/aaronshenhao/orgs", "repos_url": "https://api.github.com/users/aaronshenhao/repos", "events_url": "https://api.github.com/users/aaronshenhao/events{/privacy}", "received_events_url": "https://api.github.com/users/aaronshenhao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[]
2025-10-09T02:54:41
2025-10-09T02:57:42
null
NONE
null
null
{ "total": 0, "completed": 0, "percent_completed": 0 }
{ "blocked_by": 0, "total_blocked_by": 0, "blocking": 0, "total_blocking": 0 }
Documentation link: https://huggingface.co/docs/datasets/en/package_reference/main_classes Link to method (docstring present): https://github.com/huggingface/datasets/blob/6f2502c5a026caa89839713f6f7c8b958e5e83eb/src/datasets/arrow_dataset.py#L1029 The docstring is present for the function, but seems missing from the official documentation for the `Dataset` class on HuggingFace. The method in question: ```python @classmethod def from_dict( cls, mapping: dict, features: Optional[Features] = None, info: Optional[DatasetInfo] = None, split: Optional[NamedSplit] = None, ) -> "Dataset": """ Convert `dict` to a `pyarrow.Table` to create a [`Dataset`]. Important: a dataset created with from_dict() lives in memory and therefore doesn't have an associated cache directory. This may change in the future, but in the meantime if you want to reduce memory usage you should write it back on disk and reload using e.g. save_to_disk / load_from_disk. Args: mapping (`Mapping`): Mapping of strings to Arrays or Python lists. features ([`Features`], *optional*): Dataset features. info (`DatasetInfo`, *optional*): Dataset information, like description, citation, etc. split (`NamedSplit`, *optional*): Name of the dataset split. Returns: [`Dataset`] """ ```
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7802/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7802/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7801
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7801/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7801/comments
https://api.github.com/repos/huggingface/datasets/issues/7801/events
https://github.com/huggingface/datasets/pull/7801
3,496,388,063
PR_kwDODunzps6swITn
7,801
Add parquet scan options and docs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7801). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-08T18:04:52
2025-10-09T07:55:58
2025-10-09T07:55:56
MEMBER
null
null
null
null
I added scan options, useful to control buffering and caching when streaming and docs, including how to select a subset of columns and apply filters
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7801/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7801/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7801", "html_url": "https://github.com/huggingface/datasets/pull/7801", "diff_url": "https://github.com/huggingface/datasets/pull/7801.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7801.patch", "merged_at": "2025-10-09T07:55:56" }
true
https://api.github.com/repos/huggingface/datasets/issues/7800
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7800/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7800/comments
https://api.github.com/repos/huggingface/datasets/issues/7800/events
https://github.com/huggingface/datasets/pull/7800
3,494,747,495
PR_kwDODunzps6sqkmT
7,800
Fix polars cast column image
{ "login": "CloseChoice", "id": 31857876, "node_id": "MDQ6VXNlcjMxODU3ODc2", "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "gravatar_id": "", "url": "https://api.github.com/users/CloseChoice", "html_url": "https://github.com/CloseChoice", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "repos_url": "https://api.github.com/users/CloseChoice/repos", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "The Image() type is set to have a storage of `string` for \"path\" and not `large_string`. Therefore while your change does work to do the conversion, it can create issues in other places. For example I'm pretty sure you wouldn't be able to concatenate the resulting dataset with a dataset with Image() using `strin...
2025-10-08T10:01:18
2025-10-12T06:14:08
null
NONE
null
null
null
null
Fixes #7765 The problem here is that polars uses pyarrow large_string for images, while pandas and others just use the string type. This PR solves that and adds a test. ```python import polars as pl from datasets import Dataset import pandas as pd import pyarrow as pa from pathlib import Path shared_datadir = Path("tests/features/data") image_path = str(shared_datadir / "test_image_rgb.jpg") # Load via polars df_polars = pl.DataFrame({"image_path": [image_path]}) dataset_polars = Dataset.from_polars(df_polars) print("Polars DF is large string:", pa.types.is_large_string(df_polars.to_arrow().schema[0].type)) print("Polars DF is string:", pa.types.is_string(df_polars.to_arrow().schema[0].type)) # Load via pandas df_pandas = pd.DataFrame({"image_path": [image_path]}) dataset_pandas = Dataset.from_pandas(df_pandas) arrow_table_pd = pa.Table.from_pandas(df_pandas) print("Pandas DF is large string", pa.types.is_large_string(arrow_table_pd.schema[0].type)) print("Pandas DF is string", pa.types.is_string(arrow_table_pd.schema[0].type)) ``` Outputs: ```bash Polars DF is large string: True Polars DF is string: False Pandas DF is large string False Pandas DF is string True ```
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7800/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7800/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7800", "html_url": "https://github.com/huggingface/datasets/pull/7800", "diff_url": "https://github.com/huggingface/datasets/pull/7800.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7800.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/7799
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7799/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7799/comments
https://api.github.com/repos/huggingface/datasets/issues/7799/events
https://github.com/huggingface/datasets/pull/7799
3,487,791,741
PR_kwDODunzps6sTJKA
7,799
Define CI future
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7799). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-06T15:15:45
2025-10-07T14:30:21
2025-10-07T14:30:19
MEMBER
null
null
null
null
this should fix the CI which currently uses transformers on 3.9 while it's now unsupported
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7799/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7799/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7799", "html_url": "https://github.com/huggingface/datasets/pull/7799", "diff_url": "https://github.com/huggingface/datasets/pull/7799.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7799.patch", "merged_at": "2025-10-07T14:30:19" }
true
https://api.github.com/repos/huggingface/datasets/issues/7798
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7798/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7798/comments
https://api.github.com/repos/huggingface/datasets/issues/7798/events
https://github.com/huggingface/datasets/issues/7798
3,484,470,782
I_kwDODunzps7PsM3-
7,798
Audio dataset is not decoding on 4.1.1
{ "login": "thewh1teagle", "id": 61390950, "node_id": "MDQ6VXNlcjYxMzkwOTUw", "avatar_url": "https://avatars.githubusercontent.com/u/61390950?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thewh1teagle", "html_url": "https://github.com/thewh1teagle", "followers_url": "https://api.github.com/users/thewh1teagle/followers", "following_url": "https://api.github.com/users/thewh1teagle/following{/other_user}", "gists_url": "https://api.github.com/users/thewh1teagle/gists{/gist_id}", "starred_url": "https://api.github.com/users/thewh1teagle/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thewh1teagle/subscriptions", "organizations_url": "https://api.github.com/users/thewh1teagle/orgs", "repos_url": "https://api.github.com/users/thewh1teagle/repos", "events_url": "https://api.github.com/users/thewh1teagle/events{/privacy}", "received_events_url": "https://api.github.com/users/thewh1teagle/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "Previously (datasets<=3.6.0), audio columns were decoded automatically when accessing a row. Now, for performance reasons, audio decoding is lazy by default: you just see the file path unless you explicitly cast the column to Audio.\n\nHere’s the fix (following the current [datasets audio docs](https://huggingface...
2025-10-05T06:37:50
2025-10-06T14:07:55
null
NONE
null
null
{ "total": 0, "completed": 0, "percent_completed": 0 }
{ "blocked_by": 0, "total_blocked_by": 0, "blocking": 0, "total_blocking": 0 }
### Describe the bug The audio column remain as non-decoded objects even when accessing them. ```python dataset = load_dataset("MrDragonFox/Elise", split = "train") dataset[0] # see that it doesn't show 'array' etc... ``` Works fine with `datasets==3.6.0` Followed the docs in - https://huggingface.co/docs/datasets/en/audio_load ### Steps to reproduce the bug ```python dataset = load_dataset("MrDragonFox/Elise", split = "train") dataset[0] # see that it doesn't show 'array' etc... ``` ### Expected behavior It should decode when accessing the elemenet ### Environment info 4.1.1 ubuntu 22.04 Related - https://github.com/huggingface/datasets/issues/7707
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7798/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7798/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7797
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7797/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7797/comments
https://api.github.com/repos/huggingface/datasets/issues/7797/events
https://github.com/huggingface/datasets/pull/7797
3,473,011,621
PR_kwDODunzps6rhtf_
7,797
Datasets: Add WMT21 & WMT22 loaders (basic TSV loaders, sample data, tests)
{ "login": "tanisha-samant", "id": 164366940, "node_id": "U_kgDOCcwKXA", "avatar_url": "https://avatars.githubusercontent.com/u/164366940?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tanisha-samant", "html_url": "https://github.com/tanisha-samant", "followers_url": "https://api.github.com/users/tanisha-samant/followers", "following_url": "https://api.github.com/users/tanisha-samant/following{/other_user}", "gists_url": "https://api.github.com/users/tanisha-samant/gists{/gist_id}", "starred_url": "https://api.github.com/users/tanisha-samant/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tanisha-samant/subscriptions", "organizations_url": "https://api.github.com/users/tanisha-samant/orgs", "repos_url": "https://api.github.com/users/tanisha-samant/repos", "events_url": "https://api.github.com/users/tanisha-samant/events{/privacy}", "received_events_url": "https://api.github.com/users/tanisha-samant/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "closing since datasets should be added on https://huggingface.co directly" ]
2025-10-01T10:46:01
2025-10-10T15:33:25
2025-10-10T15:33:25
NONE
null
null
null
null
- Implemented TSV-based dataset loaders: - WMT21Dataset (local_datasets/wmt21/wmt21_dataset.py) - WMT22Dataset (local_datasets/wmt22/wmt22_dataset.py) These classes load source-target pairs from .tsv files for train, validation, and test splits. - Created sample dummy data for both datasets: - dummy_data/train.tsv, dummy_data/validation.tsv, dummy_data/test.tsv - Includes a few realistic example lines to allow CI and local tests to pass without downloading full datasets. - Added automated tests for robust validation: - tests/test_wmt21.py and tests/test_wmt22.py - Checks that all splits load correctly, empty lines are ignored, and the number of examples matches the number of lines in the .tsv files. - Edge cases handled: empty lines, malformed lines, extra tabs. - Added README.md files for both datasets: - Provides dataset structure, usage instructions, and placeholders for citation & license information. - Ensures that other developers and reviewers can understand dataset usage immediately. - Ensured easy local testing: - Load datasets programmatically using WMT21Dataset / WMT22Dataset. - Verified train/validation/test splits are correctly returned as Python dictionaries of Dataset objects. -Provides initial support for WMT21 and WMT22 NLP/translation experiments. -Allows contributors and reviewers to test dataset loading locally or in CI without downloading large datasets. -Serves as a template to extend to other WMT datasets in the future. Testing instructions: ``` # Activate your environment pytest tests/test_wmt21.py -v pytest tests/test_wmt22.py -v ``` ``` from local_datasets.wmt21.wmt21_dataset import WMT21Dataset from local_datasets.wmt22.wmt22_dataset import WMT22Dataset # WMT21 wmt21 = WMT21Dataset("local_datasets/wmt21/dummy_data") ds21 = wmt21.load() print(ds21["train"][0]) print(ds21["validation"][0]) print(ds21["test"][0]) # WMT22 wmt22 = WMT22Dataset("local_datasets/wmt22/dummy_data") ds22 = wmt22.load() print(ds22["train"][0]) print(ds22["validation"][0]) print(ds22["test"][0]) ```
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7797/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7797/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7797", "html_url": "https://github.com/huggingface/datasets/pull/7797", "diff_url": "https://github.com/huggingface/datasets/pull/7797.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7797.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/7796
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7796/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7796/comments
https://api.github.com/repos/huggingface/datasets/issues/7796/events
https://github.com/huggingface/datasets/pull/7796
3,470,616,799
PR_kwDODunzps6rZjrW
7,796
Docs: fix typo, improve readability, add code comments
{ "login": "tanisha-samant", "id": 164366940, "node_id": "U_kgDOCcwKXA", "avatar_url": "https://avatars.githubusercontent.com/u/164366940?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tanisha-samant", "html_url": "https://github.com/tanisha-samant", "followers_url": "https://api.github.com/users/tanisha-samant/followers", "following_url": "https://api.github.com/users/tanisha-samant/following{/other_user}", "gists_url": "https://api.github.com/users/tanisha-samant/gists{/gist_id}", "starred_url": "https://api.github.com/users/tanisha-samant/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tanisha-samant/subscriptions", "organizations_url": "https://api.github.com/users/tanisha-samant/orgs", "repos_url": "https://api.github.com/users/tanisha-samant/repos", "events_url": "https://api.github.com/users/tanisha-samant/events{/privacy}", "received_events_url": "https://api.github.com/users/tanisha-samant/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[]
2025-09-30T18:34:16
2025-10-10T18:44:12
null
NONE
null
null
null
null
What I did: - Fixed a small typo in README to improve clarity - Fixed repeated word "frameworks frameworks" - Split long paragraphs into shorter sentences for readability - Added # Example comments before code blocks for clarity Why: - Helps new users avoid confusion How I tested: - Checked locally in Markdown preview
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7796/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7796/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7796", "html_url": "https://github.com/huggingface/datasets/pull/7796", "diff_url": "https://github.com/huggingface/datasets/pull/7796.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7796.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/7795
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7795/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7795/comments
https://api.github.com/repos/huggingface/datasets/issues/7795/events
https://github.com/huggingface/datasets/pull/7795
3,463,990,654
PR_kwDODunzps6rDEce
7,795
Add pyarrow's binary view to features
{ "login": "delta003", "id": 6834061, "node_id": "MDQ6VXNlcjY4MzQwNjE=", "avatar_url": "https://avatars.githubusercontent.com/u/6834061?v=4", "gravatar_id": "", "url": "https://api.github.com/users/delta003", "html_url": "https://github.com/delta003", "followers_url": "https://api.github.com/users/delta003/followers", "following_url": "https://api.github.com/users/delta003/following{/other_user}", "gists_url": "https://api.github.com/users/delta003/gists{/gist_id}", "starred_url": "https://api.github.com/users/delta003/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/delta003/subscriptions", "organizations_url": "https://api.github.com/users/delta003/orgs", "repos_url": "https://api.github.com/users/delta003/repos", "events_url": "https://api.github.com/users/delta003/events{/privacy}", "received_events_url": "https://api.github.com/users/delta003/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "@lhoestq 🙏 ", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7795). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-09-29T09:12:55
2025-10-10T16:04:21
2025-10-10T16:04:21
CONTRIBUTOR
null
null
null
null
Basically https://github.com/huggingface/datasets/pull/7718 just for binary view instead of string view
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7795/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 2, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7795/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7795", "html_url": "https://github.com/huggingface/datasets/pull/7795", "diff_url": "https://github.com/huggingface/datasets/pull/7795.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7795.patch", "merged_at": "2025-10-10T16:04:21" }
true
https://api.github.com/repos/huggingface/datasets/issues/7794
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7794/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7794/comments
https://api.github.com/repos/huggingface/datasets/issues/7794/events
https://github.com/huggingface/datasets/pull/7794
3,460,793,966
PR_kwDODunzps6q4XyU
7,794
Fix nested data conversions error in parquet loading (fixes #7793)
{ "login": "Aishwarya0811", "id": 41635755, "node_id": "MDQ6VXNlcjQxNjM1NzU1", "avatar_url": "https://avatars.githubusercontent.com/u/41635755?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Aishwarya0811", "html_url": "https://github.com/Aishwarya0811", "followers_url": "https://api.github.com/users/Aishwarya0811/followers", "following_url": "https://api.github.com/users/Aishwarya0811/following{/other_user}", "gists_url": "https://api.github.com/users/Aishwarya0811/gists{/gist_id}", "starred_url": "https://api.github.com/users/Aishwarya0811/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Aishwarya0811/subscriptions", "organizations_url": "https://api.github.com/users/Aishwarya0811/orgs", "repos_url": "https://api.github.com/users/Aishwarya0811/repos", "events_url": "https://api.github.com/users/Aishwarya0811/events{/privacy}", "received_events_url": "https://api.github.com/users/Aishwarya0811/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "Unfortunately, I'm running into this error:\r\n```\r\n~/scratch » uv run python test_hf.py \r\nResolving data file...
2025-09-27T22:04:13
2025-10-01T16:56:20
null
NONE
null
null
null
null
Fixes #7793 ## Problem Loading datasets with deeply nested structures (like `metr-evals/malt-public`) fails with: ArrowNotImplementedError: Nested data conversions not implemented for chunked array outputs This occurs when parquet files contain nested data (lists, structs, maps) that exceed PyArrow's 16MB chunk limit. ## Root Cause PyArrow's C++ implementation explicitly rejects nested data conversions when data is split across multiple chunks. The limitation exists in the `WrapIntoListArray` function where repetition levels cannot be reconstructed across chunk boundaries. ## Solution - **Fallback mechanism**: Catches the specific PyArrow error and switches to non-chunked reading - **Selective optimization**: Only combines chunks for problematic nested columns to minimize memory impact - **Manual batching**: Maintains batching behavior even in fallback mode - **Backward compatibility**: Zero impact on existing datasets ## Implementation Details - Added `_is_nested_type()` helper to detect nested PyArrow types - Added `_handle_nested_chunked_conversion()` for selective chunk combining - Modified `_generate_tables()` to catch and handle the specific error - Preserves all existing error handling and logging ## Testing - [x] No regressions: Normal parquet datasets continue working - [x] Code follows existing patterns in the datasets codebase - [x] tested by original reporter (gated dataset access needed) **Note**: This fix is based on thorough research of PyArrow limitations and similar issues in the ecosystem. While we cannot test with the original dataset due to access restrictions, the implementation follows established patterns for handling this PyArrow limitation. ## Request for Testing @neevparikh Could you please test this fix with your original failing dataset? The implementation should resolve the nested data conversion error you encountered.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7794/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7794/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7794", "html_url": "https://github.com/huggingface/datasets/pull/7794", "diff_url": "https://github.com/huggingface/datasets/pull/7794.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7794.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/7793
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7793/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7793/comments
https://api.github.com/repos/huggingface/datasets/issues/7793/events
https://github.com/huggingface/datasets/issues/7793
3,459,496,971
I_kwDODunzps7OM7wL
7,793
Cannot load dataset, fails with nested data conversions not implemented for chunked array outputs
{ "login": "neevparikh", "id": 41182432, "node_id": "MDQ6VXNlcjQxMTgyNDMy", "avatar_url": "https://avatars.githubusercontent.com/u/41182432?v=4", "gravatar_id": "", "url": "https://api.github.com/users/neevparikh", "html_url": "https://github.com/neevparikh", "followers_url": "https://api.github.com/users/neevparikh/followers", "following_url": "https://api.github.com/users/neevparikh/following{/other_user}", "gists_url": "https://api.github.com/users/neevparikh/gists{/gist_id}", "starred_url": "https://api.github.com/users/neevparikh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neevparikh/subscriptions", "organizations_url": "https://api.github.com/users/neevparikh/orgs", "repos_url": "https://api.github.com/users/neevparikh/repos", "events_url": "https://api.github.com/users/neevparikh/events{/privacy}", "received_events_url": "https://api.github.com/users/neevparikh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "Hey @neevparikh,\nThanks for reporting this! I can reproduce the issue and have identified the root cause.\nProblem: The metr-evals/malt-public dataset contains deeply nested conversation data that exceeds PyArrow's 16MB chunk limit. When PyArrow tries to read it in chunks, it hits a fundamental limitation: \"Nest...
2025-09-27T01:03:12
2025-09-27T21:35:31
null
NONE
null
null
{ "total": 0, "completed": 0, "percent_completed": 0 }
{ "blocked_by": 0, "total_blocked_by": 0, "blocking": 0, "total_blocking": 0 }
### Describe the bug Hi! When I load this dataset, it fails with a pyarrow error. I'm using datasets 4.1.1, though I also see this with datasets 4.1.2 To reproduce: ``` import datasets ds = datasets.load_dataset(path="metr-evals/malt-public", name="irrelevant_detail") ``` Error: ``` Traceback (most recent call last): File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 1815, in _prepare_split_single for _, table in generator: ^^^^^^^^^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/packaged_modules/parquet/parquet.py", line 93, in _generate_tables for batch_idx, record_batch in enumerate( ~~~~~~~~~^ parquet_fragment.to_batches( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ...<5 lines>... ) ^ ): ^ File "pyarrow/_dataset.pyx", line 3904, in _iterator File "pyarrow/_dataset.pyx", line 3494, in pyarrow._dataset.TaggedRecordBatchIterator.__next__ File "pyarrow/error.pxi", line 155, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Nested data conversions not implemented for chunked array outputs The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Users/neev/scratch/test_hf.py", line 3, in <module> ds = datasets.load_dataset(path="metr-evals/malt-public", name="irrelevant_detail") File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/load.py", line 1412, in load_dataset builder_instance.download_and_prepare( ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^ download_config=download_config, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ...<3 lines>... storage_options=storage_options, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 894, in download_and_prepare self._download_and_prepare( ~~~~~~~~~~~~~~~~~~~~~~~~~~^ dl_manager=dl_manager, ^^^^^^^^^^^^^^^^^^^^^^ ...<2 lines>... **download_and_prepare_kwargs, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 970, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 1702, in _prepare_split for job_id, done, content in self._prepare_split_single( ~~~~~~~~~~~~~~~~~~~~~~~~~~^ gen_kwargs=gen_kwargs, job_id=job_id, **_prepare_split_args ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ): ^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 1858, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset ``` ### Steps to reproduce the bug To reproduce: ``` import datasets ds = datasets.load_dataset(path="metr-evals/malt-public", name="irrelevant_detail") ``` ### Expected behavior The dataset loads ### Environment info Datasets: 4.1.1 Python: 3.13 Platform: Macos
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7793/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7793/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7792
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7792/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7792/comments
https://api.github.com/repos/huggingface/datasets/issues/7792/events
https://github.com/huggingface/datasets/issues/7792
3,456,802,210
I_kwDODunzps7OCp2i
7,792
Concatenate IterableDataset instances and distribute underlying shards in a RoundRobin manner
{ "login": "LTMeyer", "id": 13559010, "node_id": "MDQ6VXNlcjEzNTU5MDEw", "avatar_url": "https://avatars.githubusercontent.com/u/13559010?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LTMeyer", "html_url": "https://github.com/LTMeyer", "followers_url": "https://api.github.com/users/LTMeyer/followers", "following_url": "https://api.github.com/users/LTMeyer/following{/other_user}", "gists_url": "https://api.github.com/users/LTMeyer/gists{/gist_id}", "starred_url": "https://api.github.com/users/LTMeyer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LTMeyer/subscriptions", "organizations_url": "https://api.github.com/users/LTMeyer/orgs", "repos_url": "https://api.github.com/users/LTMeyer/repos", "events_url": "https://api.github.com/users/LTMeyer/events{/privacy}", "received_events_url": "https://api.github.com/users/LTMeyer/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "# With `datasets.Dataset`\n\nHere is an small script that shows the distribution differences of samples between `interleave_datasets`, `concatenate_datasets` and `concatenate_datasets` + shuffling.\n\n```python\nimport datasets as hf_datasets\n\ndef gen(dataset: int, n_samples: int):\n for i in range(n_samples)...
2025-09-26T10:05:19
2025-10-06T15:24:19
null
NONE
null
null
{ "total": 0, "completed": 0, "percent_completed": 0 }
{ "blocked_by": 0, "total_blocked_by": 0, "blocking": 0, "total_blocking": 0 }
### Feature request I would like to be able to concatenate multiple `IterableDataset` with possibly different features. I would like to then be able to stream the results in parallel (both using DDP and multiple workers in the pytorch DataLoader). I want the merge of datasets to be well balanced between the different processes. ### Motivation I want to train a model on a combination of datasets, which I can convert to a single representation. This applies to converting different datasets items to the same Python class, as using a tokenizer on multiple modalities. Assuming that my original datasets are not necessarily well balanced as they may have different size and thus different number of shards, I would like the merged dataset to be distributed evenly over the multiple processes. I don't mind if it's not perfectly balanced, and as result, some workers of the torch DataLoader do nothing, as long as the DDP is properly handled causing no deadlock. ### What I've tried I've tried the two functions already provided in datasets, namely `interleave_datasets` and `concatenate_datasets`. - Interleave seems to be the best approach of what I'm trying to do. However, it doesn't suit my purpose because as I understand it, it stops as soon as one of the dataset source is exhausted, or repeat the smallest source items until the largest is exhausted. I would like something in-between, similarly to what [roundrobin does](https://more-itertools.readthedocs.io/en/stable/api.html#more_itertools.roundrobin). - Concatenate does not mix the data enough and one dataset may be overrepresented in some early batches. Let's consider we have 3 datasets composed of different number of shards as follow [[s0_0, s0_1], [s1_0], [s2_0, s2_1, s2_3]], where s denotes the underlying shard, the first index the dataset and the second the shard number. If we request 3 shards in the `shard_data_source` we should obtain the following: index 0 gets s0_0 s2_0 index 1 gets s0_1 s2_1 index 2 gets s1_0 s2_3 I started implementing the following, but I'm afraid my sharding logic is incorrect. ```python from copy import deepcopy from itertools import chain, islice import datasets import numpy as np from datasets import IterableDataset from datasets.iterable_dataset import _BaseExamplesIterable from more_itertools import roundrobin class MixMultiSourcesExampleIterable(_BaseExamplesIterable): def __init__(self, ex_iterables: list[_BaseExamplesIterable]): super().__init__() self.ex_iterables = ex_iterables def _init_state_dict(self) -> dict: self._state_dict = { "ex_iterables": [ex_iterable._init_state_dict() for ex_iterable in self.ex_iterables], "type": self.__class__.__name__, } return self._state_dict @property def num_shards(self) -> int: return sum(ex_iterable.num_shards for ex_iterable in self.ex_iterables) def __iter__(self): yield from roundrobin(*self.ex_iterables) def shuffle_data_sources(self, generator: np.random.Generator) -> "MixMultiSourcesExampleIterable": """Shuffle the list of examples iterable, as well as each underlying examples iterable.""" rng = deepcopy(generator) ex_iterables = list(self.ex_iterables) rng.shuffle(ex_iterables) ex_iterables = [ex_iterable.shuffle_data_sources(generator) for ex_iterable in ex_iterables] return MixMultiSourcesExampleIterable(ex_iterables) def shard_data_sources(self, num_shards: int, index: int, contiguous=True) -> "MixMultiSourceExampleIterable": """Shard the underlying iterables in a roundrobin manner. Let's consider we have our iterables as [[s0_0, s0_1], [s1_0], [s2_0, s2_1, s2_3]], and we request 3 shards. index 0 gets s0_0 s2_0 index 1 gets s0_1 s2_1 index 2 gets s1_0 s2_3 """ return MixMultiSourcesExampleIterable( list( islice( # flatten all underlying iterables chain.from_iterable([ex_iterable.shard_data_sources(1, 0) for ex_iterable in self.ex_iterables]), # offset the starting point by the index index, # take over the full list, so exhaust the iterators None, # step by the number of shards requested num_shards, ) ) ) def mix_dataset(iterable_datasets: list[datasets.IterableDataset]) -> IterableDataset: ex_iterable = MixMultiSourcesExampleIterable([ds._ex_iterable for ds in iterable_datasets]) return IterableDataset( ex_iterable, distributed=iterable_datasets[0]._distributed, formatting=iterable_datasets[0]._formatting ) ``` ### Questions - Am I missing something? Is there a way to use `interleave_datasets` or `concatenate_datasets` to fit my purpose? - Would it be the right approach to spread the maximum number of underlying shards across my different processes? ### Your contribution As much as I can.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7792/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7792/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7791
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7791/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7791/comments
https://api.github.com/repos/huggingface/datasets/issues/7791/events
https://github.com/huggingface/datasets/pull/7791
3,454,046,306
PR_kwDODunzps6qh_2W
7,791
fix: add `num_proc` argument to `Dataset.to_sql`
{ "login": "EricSaikali", "id": 100021446, "node_id": "U_kgDOBfY0xg", "avatar_url": "https://avatars.githubusercontent.com/u/100021446?v=4", "gravatar_id": "", "url": "https://api.github.com/users/EricSaikali", "html_url": "https://github.com/EricSaikali", "followers_url": "https://api.github.com/users/EricSaikali/followers", "following_url": "https://api.github.com/users/EricSaikali/following{/other_user}", "gists_url": "https://api.github.com/users/EricSaikali/gists{/gist_id}", "starred_url": "https://api.github.com/users/EricSaikali/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EricSaikali/subscriptions", "organizations_url": "https://api.github.com/users/EricSaikali/orgs", "repos_url": "https://api.github.com/users/EricSaikali/repos", "events_url": "https://api.github.com/users/EricSaikali/events{/privacy}", "received_events_url": "https://api.github.com/users/EricSaikali/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "Hi ! could you also write a test to make sure this works fine ?\r\n\r\n(in case there needs to be a special logic to handle the concurrent writes to the database)" ]
2025-09-25T15:02:46
2025-10-06T10:42:23
null
NONE
null
null
null
null
**Task Done:** - Resolve issue #7788 : Add the missing argument mapping in Dataset.to_sql (`src/datasets/arrow_dataset.py`)
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7791/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7791/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7791", "html_url": "https://github.com/huggingface/datasets/pull/7791", "diff_url": "https://github.com/huggingface/datasets/pull/7791.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7791.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/7790
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7790/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7790/comments
https://api.github.com/repos/huggingface/datasets/issues/7790/events
https://github.com/huggingface/datasets/pull/7790
3,453,679,876
PR_kwDODunzps6qgvjv
7,790
update tips in docs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7790). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "cc @mishig25" ]
2025-09-25T13:36:02
2025-09-25T13:39:28
2025-09-25T13:39:22
MEMBER
null
null
null
null
null
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7790/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7790/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7790", "html_url": "https://github.com/huggingface/datasets/pull/7790", "diff_url": "https://github.com/huggingface/datasets/pull/7790.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7790.patch", "merged_at": "2025-09-25T13:39:22" }
true
https://api.github.com/repos/huggingface/datasets/issues/7789
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7789/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7789/comments
https://api.github.com/repos/huggingface/datasets/issues/7789/events
https://github.com/huggingface/datasets/pull/7789
3,453,273,059
PR_kwDODunzps6qfZUc
7,789
fix link for rotten_tomatoes dataset
{ "login": "0xmohit", "id": 8176079, "node_id": "MDQ6VXNlcjgxNzYwNzk=", "avatar_url": "https://avatars.githubusercontent.com/u/8176079?v=4", "gravatar_id": "", "url": "https://api.github.com/users/0xmohit", "html_url": "https://github.com/0xmohit", "followers_url": "https://api.github.com/users/0xmohit/followers", "following_url": "https://api.github.com/users/0xmohit/following{/other_user}", "gists_url": "https://api.github.com/users/0xmohit/gists{/gist_id}", "starred_url": "https://api.github.com/users/0xmohit/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/0xmohit/subscriptions", "organizations_url": "https://api.github.com/users/0xmohit/orgs", "repos_url": "https://api.github.com/users/0xmohit/repos", "events_url": "https://api.github.com/users/0xmohit/events{/privacy}", "received_events_url": "https://api.github.com/users/0xmohit/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[]
2025-09-25T11:51:36
2025-09-25T11:51:36
null
NONE
null
null
null
null
The current link leads to a 404 page.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7789/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7789/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7789", "html_url": "https://github.com/huggingface/datasets/pull/7789", "diff_url": "https://github.com/huggingface/datasets/pull/7789.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7789.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/7788
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7788/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7788/comments
https://api.github.com/repos/huggingface/datasets/issues/7788/events
https://github.com/huggingface/datasets/issues/7788
3,450,913,796
I_kwDODunzps7NsMQE
7,788
`Dataset.to_sql` doesn't utilize `num_proc`
{ "login": "tcsmaster", "id": 30357072, "node_id": "MDQ6VXNlcjMwMzU3MDcy", "avatar_url": "https://avatars.githubusercontent.com/u/30357072?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tcsmaster", "html_url": "https://github.com/tcsmaster", "followers_url": "https://api.github.com/users/tcsmaster/followers", "following_url": "https://api.github.com/users/tcsmaster/following{/other_user}", "gists_url": "https://api.github.com/users/tcsmaster/gists{/gist_id}", "starred_url": "https://api.github.com/users/tcsmaster/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tcsmaster/subscriptions", "organizations_url": "https://api.github.com/users/tcsmaster/orgs", "repos_url": "https://api.github.com/users/tcsmaster/repos", "events_url": "https://api.github.com/users/tcsmaster/events{/privacy}", "received_events_url": "https://api.github.com/users/tcsmaster/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[]
2025-09-24T20:34:47
2025-09-24T20:35:01
null
NONE
null
null
{ "total": 0, "completed": 0, "percent_completed": 0 }
{ "blocked_by": 0, "total_blocked_by": 0, "blocking": 0, "total_blocking": 0 }
The underlying `SqlDatasetWriter` has `num_proc` as an available argument [here](https://github.com/huggingface/datasets/blob/5dc1a179783dff868b0547c8486268cfaea1ea1f/src/datasets/io/sql.py#L63) , but `Dataset.to_sql()` does not accept it, therefore it is always using one process for the SQL conversion.
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7788/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7788/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7787
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7787/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7787/comments
https://api.github.com/repos/huggingface/datasets/issues/7787/events
https://github.com/huggingface/datasets/pull/7787
3,450,858,674
PR_kwDODunzps6qXRo-
7,787
feat: avoid some copies in torch formatter
{ "login": "drbh", "id": 9896130, "node_id": "MDQ6VXNlcjk4OTYxMzA=", "avatar_url": "https://avatars.githubusercontent.com/u/9896130?v=4", "gravatar_id": "", "url": "https://api.github.com/users/drbh", "html_url": "https://github.com/drbh", "followers_url": "https://api.github.com/users/drbh/followers", "following_url": "https://api.github.com/users/drbh/following{/other_user}", "gists_url": "https://api.github.com/users/drbh/gists{/gist_id}", "starred_url": "https://api.github.com/users/drbh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/drbh/subscriptions", "organizations_url": "https://api.github.com/users/drbh/orgs", "repos_url": "https://api.github.com/users/drbh/repos", "events_url": "https://api.github.com/users/drbh/events{/privacy}", "received_events_url": "https://api.github.com/users/drbh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7787). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "can you re-read your PR please ?" ]
2025-09-24T20:19:44
2025-09-26T15:04:25
2025-09-26T15:04:23
CONTRIBUTOR
null
null
null
null
## perf: reduce copies in TorchFormatter This PR make changes the torch formatter to avoid unnecessary copies and casts when converting decoded batches to tensors. Because many arrays are already in a torch-friendly memory layout and dtype, we can do zero‑copy conversions (`torch.from_numpy`) and only fall back to `as_tensor` when a dtype/device change is required. We also consolidate lists of same‑shape tensors with a cheap `stack` only when safe. Why it helps - Avoids extra materialization and dtype churn during batched map and indexing. - Preserves API and outputs; only changes internal conversion logic. Small benchmark script (based on https://github.com/huggingface/datasets/issues/6104) ```python import time from datasets import load_dataset def main(): dataset = load_dataset("NightMachinery/hf_datasets_bug1") dataset = dataset["train"] if "train" in dataset else dataset t0 = time.time() dataset.set_format(type="torch") # identity map with small batches dataset = dataset.map(lambda x: x, batched=True, batch_size=20) # force materialization data = dataset[:300] print(len(data.keys())) t1 = time.time() print(f"Duration: {t1 - t0:.2f} s") if __name__ == "__main__": main() ``` Without changes ```bash uv run bench.py ``` ```bash # 303 # Duration: 7.26 s ``` With changes ```bash uv run bench.py ``` ```bash # 303 # Duration: 4.43 s ``` # Updated reproduction scripts Below are some simple test cases using `main` and this `refactor-torch-formatter` branch. I've included the two scripts and output when running on a local machine. ```python # /// script # requires-python = ">=3.10" # dependencies = [ # "torch", # "datasets", # "pillow", # ] # # [tool.uv.sources] # datasets = { git = "https://github.com/huggingface/datasets.git" } # /// import time import random import numpy as np from PIL import Image from datasets import Dataset, load_dataset import torch def create_mock_images_dataset(num_samples=5000): """Create a deterministic mock dataset with PIL images.""" random.seed(42) np.random.seed(42) images = [] labels = [] for i in range(num_samples): # Create deterministic RGB image width, height = 64, 64 rgb_array = np.random.randint(0, 256, (height, width, 3), dtype=np.uint8) image = Image.fromarray(rgb_array) images.append(image) labels.append(i % 10) # 10 classes return Dataset.from_dict({"image": images, "label": labels}) def create_mock_text_dataset(num_samples=5000): """Create a deterministic mock dataset with text.""" random.seed(42) words = ["apple", "banana", "cherry", "date", "elderberry", "fig", "grape", "honeydew"] texts = [] labels = [] for i in range(num_samples): # Create deterministic text text_length = 5 + (i % 20) # 5-24 words text = " ".join(random.choices(words, k=text_length)) texts.append(text) labels.append(i % 3) # 3 classes return Dataset.from_dict({"text": texts, "label": labels}) def create_mock_ints_dataset(num_samples=5000): """Create a deterministic mock dataset with integers.""" random.seed(42) data = [] labels = [] for i in range(num_samples): # Create deterministic integer arrays arr = [random.randint(0, 1000) for _ in range(50)] # 50 integers each data.append(arr) labels.append(i % 5) # 5 classes return Dataset.from_dict({"data": data, "label": labels}) def create_mock_floats_dataset(num_samples=5000): """Create a deterministic mock dataset with floats.""" random.seed(42) data = [] labels = [] for i in range(num_samples): # Create deterministic float arrays arr = [random.uniform(0.0, 100.0) for _ in range(30)] # 30 floats each data.append(arr) labels.append(i % 4) # 4 classes return Dataset.from_dict({"data": data, "label": labels}) def benchmark_dataset(name, dataset, num_samples=1000): """Benchmark dataset access speed.""" print(f"\n=== {name} Dataset Benchmark ===") t0 = time.time() dataset.set_format(type="torch") # identity map with small batches dataset = dataset.map(lambda x: x, batched=True, batch_size=20) # force materialization data = dataset[:num_samples] print(f"Keys: {list(data.keys())}") print(f"Sample count: {len(data[list(data.keys())[0]])}") t1 = time.time() print(f"Duration: {t1 - t0:.2f} s") print(f"Speed: {num_samples / (t1 - t0):.1f} samples/s") def main(): # PIL Images benchmark images_dataset = create_mock_images_dataset() benchmark_dataset("PIL Images", images_dataset) # Text benchmark text_dataset = create_mock_text_dataset() benchmark_dataset("Text", text_dataset) # Integers benchmark ints_dataset = create_mock_ints_dataset() benchmark_dataset("Integers", ints_dataset) # Floats benchmark floats_dataset = create_mock_floats_dataset() benchmark_dataset("Floats", floats_dataset) if __name__ == "__main__": main() ``` output ```bash uv run --refresh example1.py ``` ```text === PIL Images Dataset Benchmark === Map: 0%| | 0/5000 [00:00<?, ? examples/s]/Users/drbh/.cache/uv/environments-v2/example1-2aca1a30e84bdead/lib/python3.10/site-packages/datasets/features/image.py:352: UserWarning: Downcasting array dtype int64 to uint8 to be compatible with 'Pillow' warnings.warn(f"Downcasting array dtype {dtype} to {dest_dtype} to be compatible with 'Pillow'") Map: 100%|█████████████████████████████████████████████| 5000/5000 [00:01<00:00, 3669.15 examples/s] Keys: ['image', 'label'] Sample count: 1000 Duration: 2.14 s Speed: 466.5 samples/s === Text Dataset Benchmark === Map: 100%|███████████████████████████████████████████| 5000/5000 [00:00<00:00, 141327.04 examples/s] Keys: ['text', 'label'] Sample count: 1000 Duration: 0.04 s Speed: 27004.3 samples/s === Integers Dataset Benchmark === Map: 100%|███████████████████████████████████████████| 5000/5000 [00:00<00:00, 112904.90 examples/s] Keys: ['data', 'label'] Sample count: 1000 Duration: 0.05 s Speed: 21680.6 samples/s === Floats Dataset Benchmark === Map: 100%|███████████████████████████████████████████| 5000/5000 [00:00<00:00, 104084.25 examples/s] Keys: ['data', 'label'] Sample count: 1000 Duration: 0.05 s Speed: 20215.1 samples/s ``` and this branch specifically ```python # /// script # requires-python = ">=3.10" # dependencies = [ # "torch", # "datasets", # "pillow", # ] # # [tool.uv.sources] # datasets = { git = "https://github.com/huggingface/datasets.git", rev = "refactor-torch-formatter" } # /// import time import random import numpy as np from PIL import Image from datasets import Dataset, load_dataset import torch def create_mock_images_dataset(num_samples=5000): """Create a deterministic mock dataset with PIL images.""" random.seed(42) np.random.seed(42) images = [] labels = [] for i in range(num_samples): # Create deterministic RGB image width, height = 64, 64 rgb_array = np.random.randint(0, 256, (height, width, 3), dtype=np.uint8) image = Image.fromarray(rgb_array) images.append(image) labels.append(i % 10) # 10 classes return Dataset.from_dict({"image": images, "label": labels}) def create_mock_text_dataset(num_samples=5000): """Create a deterministic mock dataset with text.""" random.seed(42) words = [ "apple", "banana", "cherry", "date", "elderberry", "fig", "grape", "honeydew", ] texts = [] labels = [] for i in range(num_samples): # Create deterministic text text_length = 5 + (i % 20) # 5-24 words text = " ".join(random.choices(words, k=text_length)) texts.append(text) labels.append(i % 3) # 3 classes return Dataset.from_dict({"text": texts, "label": labels}) def create_mock_ints_dataset(num_samples=5000): """Create a deterministic mock dataset with integers.""" random.seed(42) data = [] labels = [] for i in range(num_samples): # Create deterministic integer arrays arr = [random.randint(0, 1000) for _ in range(50)] # 50 integers each data.append(arr) labels.append(i % 5) # 5 classes return Dataset.from_dict({"data": data, "label": labels}) def create_mock_floats_dataset(num_samples=5000): """Create a deterministic mock dataset with floats.""" random.seed(42) data = [] labels = [] for i in range(num_samples): # Create deterministic float arrays arr = [random.uniform(0.0, 100.0) for _ in range(30)] # 30 floats each data.append(arr) labels.append(i % 4) # 4 classes return Dataset.from_dict({"data": data, "label": labels}) def benchmark_dataset(name, dataset, num_samples=1000): """Benchmark dataset access speed.""" print(f"\n=== {name} Dataset Benchmark ===") t0 = time.time() dataset.set_format(type="torch") # identity map with small batches dataset = dataset.map(lambda x: x, batched=True, batch_size=20) # force materialization data = dataset[:num_samples] print(f"Keys: {list(data.keys())}") print(f"Sample count: {len(data[list(data.keys())[0]])}") t1 = time.time() print(f"Duration: {t1 - t0:.2f} s") print(f"Speed: {num_samples / (t1 - t0):.1f} samples/s") def main(): # PIL Images benchmark images_dataset = create_mock_images_dataset() benchmark_dataset("PIL Images", images_dataset) # Text benchmark text_dataset = create_mock_text_dataset() benchmark_dataset("Text", text_dataset) # Integers benchmark ints_dataset = create_mock_ints_dataset() benchmark_dataset("Integers", ints_dataset) # Floats benchmark floats_dataset = create_mock_floats_dataset() benchmark_dataset("Floats", floats_dataset) if __name__ == "__main__": main() ``` ```bash uv run --refresh example2.py ``` ```text Updated https://github.com/huggingface/datasets.git (2cb64d1b6503afb49d822b20979760efe4519d03) Built datasets @ git+https://github.com/huggingface/datasets.git@2cb64d1b6503afb49d822b20979760efe Uninstalled 1 package in 20ms Installed 1 package in 5ms === PIL Images Dataset Benchmark === Map: 0%| | 0/5000 [00:00<?, ? examples/s]/Users/drbh/.cache/uv/environments-v2/example2-d4af608668b706ec/lib/python3.10/site-packages/datasets/features/image.py:352: UserWarning: Downcasting array dtype int64 to uint8 to be compatible with 'Pillow' warnings.warn(f"Downcasting array dtype {dtype} to {dest_dtype} to be compatible with 'Pillow'") Map: 100%|█████████████████████████████████████████████| 5000/5000 [00:01<00:00, 3645.14 examples/s] Keys: ['image', 'label'] Sample count: 1000 Duration: 2.04 s Speed: 491.2 samples/s === Text Dataset Benchmark === Map: 100%|████████████████████████████████████████████████████| 5000/5000 [00:00<00:00, 169877.28 examples/s] Keys: ['text', 'label'] Sample count: 1000 Duration: 0.03 s Speed: 32236.1 samples/s === Integers Dataset Benchmark === Map: 100%|████████████████████████████████████████████████████| 5000/5000 [00:00<00:00, 131940.33 examples/s] Keys: ['data', 'label'] Sample count: 1000 Duration: 0.04 s Speed: 25493.3 samples/s === Floats Dataset Benchmark === Map: 100%|████████████████████████████████████████████████████| 5000/5000 [00:00<00:00, 120621.64 examples/s] Keys: ['data', 'label'] Sample count: 1000 Duration: 0.04 s Speed: 23370.6 samples/s ```
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7787/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7787/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7787", "html_url": "https://github.com/huggingface/datasets/pull/7787", "diff_url": "https://github.com/huggingface/datasets/pull/7787.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7787.patch", "merged_at": "2025-09-26T15:04:23" }
true
https://api.github.com/repos/huggingface/datasets/issues/7786
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7786/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7786/comments
https://api.github.com/repos/huggingface/datasets/issues/7786/events
https://github.com/huggingface/datasets/pull/7786
3,448,506,148
PR_kwDODunzps6qPTgs
7,786
Sample without replacement option when interleaving datasets
{ "login": "radulescupetru", "id": 26553095, "node_id": "MDQ6VXNlcjI2NTUzMDk1", "avatar_url": "https://avatars.githubusercontent.com/u/26553095?v=4", "gravatar_id": "", "url": "https://api.github.com/users/radulescupetru", "html_url": "https://github.com/radulescupetru", "followers_url": "https://api.github.com/users/radulescupetru/followers", "following_url": "https://api.github.com/users/radulescupetru/following{/other_user}", "gists_url": "https://api.github.com/users/radulescupetru/gists{/gist_id}", "starred_url": "https://api.github.com/users/radulescupetru/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/radulescupetru/subscriptions", "organizations_url": "https://api.github.com/users/radulescupetru/orgs", "repos_url": "https://api.github.com/users/radulescupetru/repos", "events_url": "https://api.github.com/users/radulescupetru/events{/privacy}", "received_events_url": "https://api.github.com/users/radulescupetru/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "@lhoestq Continuing on the idea from https://github.com/huggingface/datasets/issues/217 \r\nThis doesn't add a new stopping criteria, but a new argument to interleave_datasets method. Let me know what you think and if you see a better way of doing this I'm open to suggestions.", "Great ! this is a cool additions...
2025-09-24T09:18:14
2025-10-07T14:50:16
2025-10-07T14:50:16
CONTRIBUTOR
null
null
null
null
Right now, `interleave_datasets` function with probabilities will sample with replacement. The PR adds the ability to sample without replacement. ``` import datasets # Create datasets of different sizes to test exhaustion data_a = [{"value": i, "source": "A"} for i in range(5)] data_b = [{"value": i, "source": "B"} for i in range(10, 15)] ds_a = datasets.Dataset.from_list(data_a).to_iterable_dataset() ds_b = datasets.Dataset.from_list(data_b).to_iterable_dataset() # Interleave with probabilities ds_interleaved = datasets.interleave_datasets( [ds_a, ds_b], probabilities=[0.6, 0.4], seed=42, stopping_strategy="all_exhausted", sample_with_replacement=True, ) for i, example in enumerate(ds_interleaved): print(f"Sample:{i}: value:{example['value']:02d} source:{example['source']}") ``` In this example, `sample_with_replacement=True` and it prints: ``` Sample:0: value:10 source:B Sample:1: value:00 source:A Sample:2: value:11 source:B Sample:3: value:12 source:B Sample:4: value:01 source:A Sample:5: value:13 source:B Sample:6: value:14 source:B Sample:7: value:10 source:B Sample:8: value:02 source:A Sample:9: value:03 source:A Sample:10: value:04 source:A ``` Note that sample with value:10 source: B is sampled twice (Sample:0 and Sample:7) Re-running with `sample_with_replacement=False` in prints: ``` Sample:0: value:10 source:B Sample:1: value:00 source:A Sample:2: value:11 source:B Sample:3: value:12 source:B Sample:4: value:01 source:A Sample:5: value:13 source:B Sample:6: value:14 source:B Sample:7: value:02 source:A Sample:8: value:03 source:A Sample:9: value:04 source:A ``` Note that we don't see any repeated items.
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7786/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7786/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7786", "html_url": "https://github.com/huggingface/datasets/pull/7786", "diff_url": "https://github.com/huggingface/datasets/pull/7786.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7786.patch", "merged_at": "2025-10-07T14:50:15" }
true
https://api.github.com/repos/huggingface/datasets/issues/7785
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7785/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7785/comments
https://api.github.com/repos/huggingface/datasets/issues/7785/events
https://github.com/huggingface/datasets/pull/7785
3,439,897,018
PR_kwDODunzps6pyTM_
7,785
Fix Audio docstring by removing unsupported mono argument
{ "login": "tanuj-rai", "id": 84439872, "node_id": "MDQ6VXNlcjg0NDM5ODcy", "avatar_url": "https://avatars.githubusercontent.com/u/84439872?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tanuj-rai", "html_url": "https://github.com/tanuj-rai", "followers_url": "https://api.github.com/users/tanuj-rai/followers", "following_url": "https://api.github.com/users/tanuj-rai/following{/other_user}", "gists_url": "https://api.github.com/users/tanuj-rai/gists{/gist_id}", "starred_url": "https://api.github.com/users/tanuj-rai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tanuj-rai/subscriptions", "organizations_url": "https://api.github.com/users/tanuj-rai/orgs", "repos_url": "https://api.github.com/users/tanuj-rai/repos", "events_url": "https://api.github.com/users/tanuj-rai/events{/privacy}", "received_events_url": "https://api.github.com/users/tanuj-rai/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
[ "I think we can keep the arg and add the missing torch.mean() in the Audio.decode_example method", "> I think we can keep the arg and add the missing torch.mean() in the Audio.decode_example method\r\n\r\nThank you @lhoestq. I will add torch.mean()." ]
2025-09-22T09:06:52
2025-09-23T09:57:37
null
CONTRIBUTOR
null
null
null
null
This PR fixes issue #7745. Who can review: @lhoestq
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7785/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7785/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7785", "html_url": "https://github.com/huggingface/datasets/pull/7785", "diff_url": "https://github.com/huggingface/datasets/pull/7785.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7785.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/7783
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7783/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7783/comments
https://api.github.com/repos/huggingface/datasets/issues/7783/events
https://github.com/huggingface/datasets/pull/7783
3,430,715,779
PR_kwDODunzps6pT7pg
7,783
Support huggingface_hub v0.x and v1.x
{ "login": "Wauplin", "id": 11801849, "node_id": "MDQ6VXNlcjExODAxODQ5", "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Wauplin", "html_url": "https://github.com/Wauplin", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "repos_url": "https://api.github.com/users/Wauplin/repos", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7783). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "@lhoestq could you have a look at this PR please? It makes `datasets` compatible with t...
2025-09-18T14:45:20
2025-10-01T13:56:05
2025-10-01T13:56:03
CONTRIBUTOR
null
null
null
null
Related to https://github.com/huggingface/huggingface_hub/issues/3340. This PR adapts `datasets` to be compatible with both huggingface_hub v0.x and v1.x. In practice nothing else should change (I've checked the codebase). The `HfHubHTTPError` is a base error defined in `huggingface_hub` that inherits from `requests.HTTPError` in v0.x and will inherit from `httpx.HTTPError` in v1.x. It has been introduced ~2 years ago so it's fine to use it right now (i.e. no need to wait for v1.x release or bump minimal version). Most of the changes have been around the test suite to make sure that tests are passing with both `requests` and `httpx` backends. Mid-term it would be good to completely remove the `requests` dependency from `datasets` but that's an orthogonal topic.
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7783/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7783/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7783", "html_url": "https://github.com/huggingface/datasets/pull/7783", "diff_url": "https://github.com/huggingface/datasets/pull/7783.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7783.patch", "merged_at": "2025-10-01T13:56:03" }
true
End of preview. Expand in Data Studio

Dataset Card for Dataset Name

This dataset card aims to be a base template for new datasets. It has been generated using this raw template.

Dataset Details

Dataset Description

  • Curated by: [More Information Needed]
  • Funded by [optional]: [More Information Needed]
  • Shared by [optional]: [More Information Needed]
  • Language(s) (NLP): [More Information Needed]
  • License: [More Information Needed]

Dataset Sources [optional]

  • Repository: [More Information Needed]
  • Paper [optional]: [More Information Needed]
  • Demo [optional]: [More Information Needed]

Uses

Direct Use

[More Information Needed]

Out-of-Scope Use

[More Information Needed]

Dataset Structure

[More Information Needed]

Dataset Creation

Curation Rationale

[More Information Needed]

Source Data

Data Collection and Processing

[More Information Needed]

Who are the source data producers?

[More Information Needed]

Annotations [optional]

Annotation process

[More Information Needed]

Who are the annotators?

[More Information Needed]

Personal and Sensitive Information

[More Information Needed]

Bias, Risks, and Limitations

[More Information Needed]

Recommendations

Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.

Citation [optional]

BibTeX:

[More Information Needed]

APA:

[More Information Needed]

Glossary [optional]

[More Information Needed]

More Information [optional]

[More Information Needed]

Dataset Card Authors [optional]

[More Information Needed]

Dataset Card Contact

[More Information Needed]

Downloads last month
6