id int64 599M 3.48B | number int64 1 7.8k | title stringlengths 1 290 | state stringclasses 2
values | comments listlengths 0 30 | created_at timestamp[s]date 2020-04-14 10:18:02 2025-10-05 06:37:50 | updated_at timestamp[s]date 2020-04-27 16:04:17 2025-10-05 10:32:43 | closed_at timestamp[s]date 2020-04-14 12:01:40 2025-10-01 13:56:03 β | body stringlengths 0 228k β | user stringlengths 3 26 | html_url stringlengths 46 51 | pull_request dict | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,438,367,678 | 5,209 | Implement ability to define splits in metadata section of dataset card | closed | [
"@merveenoyan Do you want different files to be splits or configurations?\r\n\r\nFrom [what you specified in `Readme.md`](https://huggingface.co/datasets/inria-soda/tabular-benchmark/commit/fb4575853772c62a20203bdd6cc0202f5db4ce4e) I hypothesize that you want to have 4 **configs** corresponding to directories: `\"c... | 2022-11-07T13:27:16 | 2023-07-21T14:36:02 | 2023-07-21T14:36:01 | ### Feature request
If you go here: https://huggingface.co/datasets/inria-soda/tabular-benchmark/tree/main you will see bunch of folders that has various CSV files. Iβd like dataset viewer to show these files instead of only one dataset like it currently does. (and also people to be able to load them as splits inste... | merveenoyan | https://github.com/huggingface/datasets/issues/5209 | null | false |
1,438,035,707 | 5,208 | Refactor CI hub fixtures to use monkeypatch instead of patch | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-07T09:25:05 | 2022-11-08T06:51:20 | 2022-11-08T06:49:17 | Minor refactoring of CI to use `pytest` `monkeypatch` instead of `unittest` `patch`. | albertvillanova | https://github.com/huggingface/datasets/pull/5208 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5208",
"html_url": "https://github.com/huggingface/datasets/pull/5208",
"diff_url": "https://github.com/huggingface/datasets/pull/5208.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5208.patch",
"merged_at": "2022-11-08T06:49... | true |
1,437,858,506 | 5,207 | Connection error of the HuggingFace's dataset Hub due to SSLError with proxy | open | [
"Hi ! It looks like an issue with your python environment, can you make sure you're able to run GET requests to https://huggingface.co using `requests` in python ?",
"Thanks for your reply. Does this mean that I have to use the `do_dataset `function and the `requests `function to download the dataset from the com... | 2022-11-07T06:56:23 | 2025-03-08T09:04:10 | null | ### Describe the bug
It's weird. I could not normally connect the dataset Hub of HuggingFace due to a SSLError in my office.
Even when I try to connect using my company's proxy address (e.g., http_proxy and https_proxy),
I'm getting the SSLError issue. What should I do to download the datanet stored in Hugg... | leemgs | https://github.com/huggingface/datasets/issues/5207 | null | false |
1,437,223,894 | 5,206 | Use logging instead of printing to console | closed | [
"Actually upon closer inspection, it is documented in the code that this behavior is intentional, so I'll close this."
] | 2022-11-05T23:48:02 | 2022-11-06T00:06:00 | 2022-11-06T00:05:59 | ### Describe the bug
Some logs ([here](https://github.com/huggingface/datasets/blob/4a6e1fe2735505efc7e3a3dbd3e1835da0702575/src/datasets/builder.py#L778), [here](https://github.com/huggingface/datasets/blob/4a6e1fe2735505efc7e3a3dbd3e1835da0702575/src/datasets/builder.py#L786), and [here](https://github.com/huggingfa... | bilelomrani1 | https://github.com/huggingface/datasets/issues/5206 | null | false |
1,437,221,987 | 5,205 | Add missing `DownloadConfig.use_auth_token` value | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-05T23:36:36 | 2022-11-08T08:13:00 | 2022-11-07T16:20:24 | This PR solves https://github.com/huggingface/datasets/issues/5204
Now the `token` is propagated so that `DownloadConfig.use_auth_token` value is set before trying to download private files from existing datasets in the Hub. | alvarobartt | https://github.com/huggingface/datasets/pull/5205 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5205",
"html_url": "https://github.com/huggingface/datasets/pull/5205",
"diff_url": "https://github.com/huggingface/datasets/pull/5205.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5205.patch",
"merged_at": "2022-11-07T16:20... | true |
1,437,221,259 | 5,204 | `push_to_hub` not propagating `token` through `DownloadConfig` | closed | [
"#self-assign",
"@lhoestq can you close this issue as part of the recent #5205 merge? Thanks π€ ",
"Thank you :)"
] | 2022-11-05T23:32:20 | 2022-11-08T10:12:09 | 2022-11-08T10:12:08 | ### Describe the bug
When trying to upload a new π€ Dataset to the Hub via Python, and providing the `token` as a parameter to the `Dataset.push_to_hub` function, it just works for the first time, assuming that the dataset didn't exist before.
But when trying to run `Dataset.push_to_hub` again over the same dataset... | alvarobartt | https://github.com/huggingface/datasets/issues/5204 | null | false |
1,436,710,518 | 5,203 | Update canonical links to Hub links | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-04T22:50:50 | 2022-11-07T18:43:05 | 2022-11-07T18:40:19 | This PR updates some of the canonical dataset links to their corresponding links on the Hub; closes #5200. | stevhliu | https://github.com/huggingface/datasets/pull/5203 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5203",
"html_url": "https://github.com/huggingface/datasets/pull/5203",
"diff_url": "https://github.com/huggingface/datasets/pull/5203.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5203.patch",
"merged_at": "2022-11-07T18:40... | true |
1,435,886,090 | 5,202 | CI fails after bulk edit of canonical datasets | closed | [
"Fixed by: https://huggingface.co/datasets/paws/discussions/1"
] | 2022-11-04T10:51:20 | 2023-02-16T09:11:10 | 2023-02-16T09:11:10 | ```
______ test_get_dataset_config_info[paws-labeled_final-expected_splits2] _______
[gw0] linux -- Python 3.7.15 /opt/hostedtoolcache/Python/3.7.15/x64/bin/python
path = 'paws', config_name = 'labeled_final'
expected_splits = ['train', 'test', 'validation']
@pytest.mark.parametrize(
"path, config... | albertvillanova | https://github.com/huggingface/datasets/issues/5202 | null | false |
1,435,881,554 | 5,201 | Do not sort splits in dataset info | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"It would be coherent with https://github.com/huggingface/datasets-server/issues/614#issuecomment-1290534153",
"I think we started working on this issue nearly at the same time... :sweat_smile: \r\n- CI was fixed with this: https://... | 2022-11-04T10:47:21 | 2022-11-04T14:47:37 | 2022-11-04T14:45:09 | I suggest not to sort splits by their names in dataset_info in README so that they are displayed in the order specified in the loading script. Otherwise `test` split is displayed first, see this repo: https://huggingface.co/datasets/paws
What do you think?
But I added sorting in tests to fix CI (for the same datase... | polinaeterna | https://github.com/huggingface/datasets/pull/5201 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5201",
"html_url": "https://github.com/huggingface/datasets/pull/5201",
"diff_url": "https://github.com/huggingface/datasets/pull/5201.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5201.patch",
"merged_at": "2022-11-04T14:45... | true |
1,435,831,559 | 5,200 | Some links to canonical datasets in the docs are outdated | closed | [
"Thanks for catching this, I can go through the docs and replace the links to their corresponding datasets on the Hub!"
] | 2022-11-04T10:06:21 | 2022-11-07T18:40:20 | 2022-11-07T18:40:20 | As we don't have canonical datasets in the github repo anymore, some old links to them doesn't work. I don't know how many of them are there, I found link to SuperGlue here: https://huggingface.co/docs/datasets/dataset_script#multiple-configurations, probably there are more of them. These links should be replaced by li... | polinaeterna | https://github.com/huggingface/datasets/issues/5200 | null | false |
1,434,818,836 | 5,199 | Deprecate dummy data generation command | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-03T15:05:54 | 2022-11-04T14:01:50 | 2022-11-04T13:59:47 | Deprecate the `dummy_data` CLI command. | mariosasko | https://github.com/huggingface/datasets/pull/5199 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5199",
"html_url": "https://github.com/huggingface/datasets/pull/5199",
"diff_url": "https://github.com/huggingface/datasets/pull/5199.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5199.patch",
"merged_at": "2022-11-04T13:59... | true |
1,434,699,165 | 5,198 | Add note about the name of a dataset script | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-03T13:51:32 | 2022-11-04T12:47:59 | 2022-11-04T12:46:01 | Add note that a dataset script should has the same name as a repo/dir, a bit related to this issue https://github.com/huggingface/datasets/issues/5193
also fixed two minor issues in audio docs (broken links) | polinaeterna | https://github.com/huggingface/datasets/pull/5198 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5198",
"html_url": "https://github.com/huggingface/datasets/pull/5198",
"diff_url": "https://github.com/huggingface/datasets/pull/5198.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5198.patch",
"merged_at": "2022-11-04T12:46... | true |
1,434,676,150 | 5,197 | [zstd] Use max window log size | open | [
"@albertvillanova Please take a review.",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5197). All of your documentation changes will be reflected on that endpoint."
] | 2022-11-03T13:35:58 | 2022-11-03T13:45:19 | null | ZstdDecompressor has a parameter `max_window_size` to limit max memory usage when decompressing zstd files. The default `max_window_size` is not enough when files are compressed by `zstd --ultra` flags.
Change `max_window_size` to the zstd's max window size. NOTE, the `zstd.WINDOWLOG_MAX` is the log_2 value of the m... | reyoung | https://github.com/huggingface/datasets/pull/5197 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5197",
"html_url": "https://github.com/huggingface/datasets/pull/5197",
"diff_url": "https://github.com/huggingface/datasets/pull/5197.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5197.patch",
"merged_at": null
} | true |
1,434,401,646 | 5,196 | Use hfh hf_hub_url function | closed | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5196). All of your documentation changes will be reflected on that endpoint.",
"@lhoestq I think we should first agree if `datasets` can introduce the breaking change of ignoring `config.HUB_DATASETS_URL`: some users may have o... | 2022-11-03T10:08:09 | 2022-12-06T11:38:17 | 2022-11-09T07:15:12 | Small refactoring to use `hf_hub_url` function from `huggingface_hub`.
This PR also creates the `hub` module that will contain all `huggingface_hub` functionalities relevant to `datasets`.
This is a necessary stage before implementing the use of the `hfh` caching system (which uses its `hf_hub_url` under the hood... | albertvillanova | https://github.com/huggingface/datasets/pull/5196 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5196",
"html_url": "https://github.com/huggingface/datasets/pull/5196",
"diff_url": "https://github.com/huggingface/datasets/pull/5196.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5196.patch",
"merged_at": "2022-11-09T07:15... | true |
1,434,290,689 | 5,195 | [wip testing docs] | closed | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5195). All of your documentation changes will be reflected on that endpoint."
] | 2022-11-03T08:37:34 | 2023-04-04T15:10:37 | 2023-04-04T15:10:33 | null | mishig25 | https://github.com/huggingface/datasets/pull/5195 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5195",
"html_url": "https://github.com/huggingface/datasets/pull/5195",
"diff_url": "https://github.com/huggingface/datasets/pull/5195.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5195.patch",
"merged_at": null
} | true |
1,434,206,951 | 5,194 | Fix docs about dataset_info in YAML | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-03T07:10:23 | 2022-11-03T13:31:27 | 2022-11-03T13:29:21 | This PR fixes some misalignment in the docs after we transferred the dataset_info from `dataset_infos.json` to YAML in the dataset card:
- #4926
Related to:
- #5193 | albertvillanova | https://github.com/huggingface/datasets/pull/5194 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5194",
"html_url": "https://github.com/huggingface/datasets/pull/5194",
"diff_url": "https://github.com/huggingface/datasets/pull/5194.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5194.patch",
"merged_at": "2022-11-03T13:29... | true |
1,433,883,780 | 5,193 | "One or several metadata. were found, but not in the same directory or in a parent directory" | closed | [
"Also unrelated but still: https://huggingface.co/docs/datasets/image_dataset#generate-the-dataset\r\n```If your loading script passed the test, you should now have a dataset_infos.json file in your dataset folder.```\r\nIt's not the case anymore as it's now in the readme.md, it was confusing to me",
"And here is... | 2022-11-02T22:46:25 | 2022-11-03T13:39:16 | 2022-11-03T13:35:44 | ### Describe the bug
When loading my own dataset, on loading it I get an error.
Here is my dataset link: https://huggingface.co/datasets/corentinm7/MyoQuant-SDH-Data
And the error after loading with:
```python
from datasets import load_dataset
load_dataset("corentinm7/MyoQuant-SDH-Data")
```
```python
Downlo... | lambda-science | https://github.com/huggingface/datasets/issues/5193 | null | false |
1,433,199,790 | 5,192 | Drop labels in Image and Audio folders if files are on different levels in directory or if there is only one label | closed | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5192). All of your documentation changes will be reflected on that endpoint.",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5192). All of your documentation changes will be reflected on... | 2022-11-02T14:01:41 | 2022-11-15T16:32:53 | 2022-11-15T16:31:07 | Will close https://github.com/huggingface/datasets/issues/5153
Drop labels by default (`drop_labels=None`) when:
* there are files on different levels of directory hierarchy by checking their path depth
* all files are in the same directory (=only one label was inferred)
First one fixes cases like this:
```
r... | polinaeterna | https://github.com/huggingface/datasets/pull/5192 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5192",
"html_url": "https://github.com/huggingface/datasets/pull/5192",
"diff_url": "https://github.com/huggingface/datasets/pull/5192.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5192.patch",
"merged_at": "2022-11-15T16:31... | true |
1,433,191,658 | 5,191 | Make torch.Tensor and spacy models cacheable | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-02T13:56:18 | 2022-11-02T17:20:48 | 2022-11-02T17:18:42 | Override `Pickler.save` to implement deterministic reduction (lazily registered; inspired by https://github.com/uqfoundation/dill/blob/master/dill/_dill.py#L343) functions for `torch.Tensor` and spaCy models.
Fix https://github.com/huggingface/datasets/issues/5170, fix https://github.com/huggingface/datasets/issues/... | mariosasko | https://github.com/huggingface/datasets/pull/5191 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5191",
"html_url": "https://github.com/huggingface/datasets/pull/5191",
"diff_url": "https://github.com/huggingface/datasets/pull/5191.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5191.patch",
"merged_at": "2022-11-02T17:18... | true |
1,433,014,626 | 5,190 | `path` is `None` when downloading a custom audio dataset from the Hub | closed | [
"Hi! Yes, this is expected behavior - we do this as a security measure to not leak local paths (this info would be useless on other users' machines anyways) and only push audio bytes. \r\n"
] | 2022-11-02T11:51:25 | 2022-11-02T12:55:02 | 2022-11-02T12:55:02 | ### Describe the bug
I've created an [audio dataset](https://huggingface.co/datasets/lewtun/audio-test-push) using the `audiofolder` feature desribed in the [docs](https://huggingface.co/docs/datasets/audio_dataset#audiofolder) and then pushed it to the Hub.
Locally, I can see the `audio.path` feature is of the ... | lewtun | https://github.com/huggingface/datasets/issues/5190 | null | false |
1,432,769,143 | 5,189 | Reduce friction in tabular dataset workflow by eliminating having splits when dataset is loaded | open | [
"I have to admit I'm not a fan of this idea, as this would result in a non-consistent behavior between tabular and non-tabular datasets, which is confusing if done without the context you provided. Instead, we could consider returning a `Dataset` object rather than `DatasetDict` if there is only one split in the ge... | 2022-11-02T09:15:02 | 2022-12-06T12:13:17 | null | ### Feature request
Sorry for cryptic name but I'd like to explain using code itself. When I want to load a specific dataset from a repository (for instance, this: https://huggingface.co/datasets/inria-soda/tabular-benchmark)
```python
from datasets import load_dataset
dataset = load_dataset("inria-soda/tabular-b... | merveenoyan | https://github.com/huggingface/datasets/issues/5189 | null | false |
1,432,477,139 | 5,188 | add: segmentation guide. | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thanks @osanseviero. Am I good to merge? ",
"I would wait for a second approval just in case :) ",
"Sure :) ",
"Merging since the images have been pushed as LFS files ([PR](https://huggingface.co/datasets/huggingface/documentat... | 2022-11-02T04:34:36 | 2022-11-04T18:25:57 | 2022-11-04T18:23:34 | Closes #5181
I have opened a PR on Hub (https://huggingface.co/datasets/huggingface/documentation-images/discussions/5) to include the images in our central Hub repository. Once the PR is merged I will edit the image links.
I have also prepared a [Colab Notebook](https://colab.research.google.com/drive/1BMDCfOT... | sayakpaul | https://github.com/huggingface/datasets/pull/5188 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5188",
"html_url": "https://github.com/huggingface/datasets/pull/5188",
"diff_url": "https://github.com/huggingface/datasets/pull/5188.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5188.patch",
"merged_at": "2022-11-04T18:23... | true |
1,432,375,375 | 5,187 | chore: add notebook links to img cls and obj det. | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"@nateraw I guess the failing test is unrelated. ",
"@sayakpaul Yea failures are unrelated. ",
"Alright. Will wait for @osanseviero's take and then merge. ",
"FYI @stevhliu ",
"@osanseviero @stevhliu @nateraw thank you for yo... | 2022-11-02T02:30:09 | 2022-11-03T01:52:24 | 2022-11-03T01:49:56 | Closes https://github.com/huggingface/datasets/issues/5182 | sayakpaul | https://github.com/huggingface/datasets/pull/5187 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5187",
"html_url": "https://github.com/huggingface/datasets/pull/5187",
"diff_url": "https://github.com/huggingface/datasets/pull/5187.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5187.patch",
"merged_at": "2022-11-03T01:49... | true |
1,432,045,011 | 5,186 | Incorrect error message when Dataset.from_sql fails and sqlalchemy not installed | closed | [
"Hi! The first `Dataset.from_sql` call also outputs the \"ImportError: Using URI string without sqlalchemy installed.\" message, but you also get \"During handling of the above exception another exception occurred: ...\" after which the ValueError is printed. I agree that this behavior makes it easy to miss the ori... | 2022-11-01T20:25:51 | 2022-11-15T18:24:39 | 2022-11-15T18:24:39 | ### Describe the bug
When calling `Dataset.from_sql` (in my case, with sqlite3), it fails with a message ```ValueError: Please pass `features` or at least one example when writing data``` when I don't have `sqlalchemy` installed.
### Steps to reproduce the bug
Make a new sqlite db with `sqlite3` and `pandas` from... | nateraw | https://github.com/huggingface/datasets/issues/5186 | null | false |
1,432,021,611 | 5,185 | Allow passing a subset of output features to Dataset.map | open | [] | 2022-11-01T20:07:20 | 2022-11-01T20:07:34 | null | ### Feature request
Currently, map does one of two things to the features (if I'm not mistaken):
* when you do not pass features, types are assumed to be equal to the input if they can be cast, and inferred otherwise
* when you pass a full specification of features, output features are set to this
However, so... | sanderland | https://github.com/huggingface/datasets/issues/5185 | null | false |
1,431,418,066 | 5,183 | Loading an external dataset in a format similar to conll2003 | closed | [] | 2022-11-01T13:18:29 | 2022-11-02T11:57:50 | 2022-11-02T11:57:50 | I'm trying to load a custom dataset in a Dataset object, it's similar to conll2003 but with 2 columns only (word entity), I used the following script:
features = datasets.Features(
{"tokens": datasets.Sequence(datasets.Value("string")),
"ner_tags": datasets.Sequence(
datasets.featu... | Taghreed7878 | https://github.com/huggingface/datasets/issues/5183 | null | false |
1,431,029,547 | 5,182 | Add notebook / other resource links to the task-specific data loading guides | closed | [
"Yea this would be great! We would need an object detection tutorial notebook too if it doesn't already exist there. ",
"There is one: https://huggingface.co/docs/datasets/object_detection.\r\n\r\nI will start the work. "
] | 2022-11-01T07:57:26 | 2022-11-03T01:49:57 | 2022-11-03T01:49:57 | Does it make sense to include links to notebooks / scripts that show how to use a dataset for training / fine-tuning a model?
For example, here in [https://huggingface.co/docs/datasets/image_classification] we could include a mention of https://github.com/huggingface/notebooks/blob/main/examples/image_classificatio... | sayakpaul | https://github.com/huggingface/datasets/issues/5182 | null | false |
1,431,027,102 | 5,181 | Add a guide for semantic segmentation | closed | [
"Sure this sounds great! Would this be pure torchvision, albumentations, or something else?",
"I am considering `torchvision` and `albumentations`. Also [works with TensorFlow](https://github.com/deep-diver/segformer-tf-transformers/blob/main/notebooks/TFSegFormer_Finetune.ipynb). \r\n\r\nI am assigning the issue... | 2022-11-01T07:54:50 | 2022-11-04T18:23:36 | 2022-11-04T18:23:36 | Currently, we have these guides for object detection and image classification:
* https://huggingface.co/docs/datasets/object_detection
* https://huggingface.co/docs/datasets/image_classification
I am proposing adding a similar guide for semantic segmentation.
I am happy to contribute a PR for it.
Cc: @os... | sayakpaul | https://github.com/huggingface/datasets/issues/5181 | null | false |
1,431,012,438 | 5,180 | An example or recommendations for creating large image datasets? | open | [
"The beam utilities allow to prepare a dataset as parquet in your cloud storage. From my perspective this CLI is not super easy to use, but we've been working on a new python API to prepare a dataset in your cloud storage:\r\n```python\r\nfrom datasets import load_dataset_builder\r\n\r\nbuilder = load_dataset_build... | 2022-11-01T07:38:38 | 2022-11-02T10:17:11 | null | I know that Apache Beam and `datasets` have [some connector utilities](https://huggingface.co/docs/datasets/beam). But it's a little unclear what we mean by "But if you want to run your own Beam pipeline with Dataflow, here is how:". What does that pipeline do?
As a user, I was wondering if we have this support for... | sayakpaul | https://github.com/huggingface/datasets/issues/5180 | null | false |
1,430,826,100 | 5,179 | `map()` fails midway due to format incompatibility | closed | [
"Cc: @lhoestq ",
"You can end up with a list instead of a tensor if all the tensors inside the list can't be stacked together - can you make sure all your inputs are tensors with the same shape ?",
"Is there an easy way to ensure it?",
"You can make sure your `tokenize` function always return tensors of the s... | 2022-11-01T03:57:59 | 2022-11-08T11:35:26 | 2022-11-08T11:35:26 | ### Describe the bug
I am using the `emotion` dataset from Hub for sequence classification. After training the model, I am using it to generate predictions for all the entries present in the `validation` split of the dataset.
```py
def get_test_accuracy(model):
def fn(batch):
inputs = {k:v.to(device... | sayakpaul | https://github.com/huggingface/datasets/issues/5179 | null | false |
1,430,800,810 | 5,178 | Unable to download the Chinese `wikipedia`, the dumpstatus.json not found! | closed | [
"In the dumps page of the wiki (https://dumps.wikimedia.org/zhwiki/), I found the following dumps:\r\n```\r\nIndex of /zhwiki/\r\n[../](https://dumps.wikimedia.org/)\r\n[20220701/](https://dumps.wikimedia.org/zhwiki/20220701/) 21-Aug-2022 01:48 -\r\n[202207... | 2022-11-01T03:17:55 | 2022-11-02T08:27:15 | 2022-11-02T08:24:29 | ### Describe the bug
I tried:
`data = load_dataset('wikipedia', '20220301.zh', beam_runner='DirectRunner')`
and
`data = load_dataset("wikipedia", language="zh", date="20220301", beam_runner='DirectRunner')`
but both got:
`FileNotFoundError: Couldn't find file at https://dumps.wikimedia.org/zhwiki/20220301/dumpsta... | beyondguo | https://github.com/huggingface/datasets/issues/5178 | null | false |
1,430,238,556 | 5,177 | Update create image dataset docs | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-31T17:45:56 | 2022-11-02T17:15:22 | 2022-11-02T17:13:02 | Based on @osanseviero and community feedback, it wasn't super clear how to upload a dataset to the Hub after creating something like an image captioning dataset. This PR adds a brief section on how to upload the dataset with `push_to_hub`. | stevhliu | https://github.com/huggingface/datasets/pull/5177 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5177",
"html_url": "https://github.com/huggingface/datasets/pull/5177",
"diff_url": "https://github.com/huggingface/datasets/pull/5177.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5177.patch",
"merged_at": "2022-11-02T17:13... | true |
1,430,214,539 | 5,176 | prepare dataset for cloud storage doesn't work | closed | [
"It looks like an issue with `gcsfs`, are you able to instantiate a `GCSFileSystem` manually ?",
"closing since it was probably due to gcsfs"
] | 2022-10-31T17:28:57 | 2023-03-28T09:11:46 | 2023-03-28T09:11:45 | ### Describe the bug
Following the [documentation](https://huggingface.co/docs/datasets/filesystems#load-and-save-your-datasets-using-your-cloud-storage-filesystem) and [this PR](https://github.com/huggingface/datasets/pull/4724), I was downloading and storing huggingface dataset to cloud storage.
```
from datasets ... | araonblake | https://github.com/huggingface/datasets/issues/5176 | null | false |
1,428,696,231 | 5,175 | Loading an external NER dataset | closed | [] | 2022-10-30T09:31:55 | 2022-11-01T13:15:49 | 2022-11-01T13:15:49 | I need to use huggingface datasets to load a custom dataset similar to conll2003 but with more entities and each the files contain only two columns: word and ner tag.
I tried this code snnipet that I found here as an answer to a similar issue:
from datasets import Dataset
INPUT_COLUMNS = "ID Text NER".split()
... | Taghreed7878 | https://github.com/huggingface/datasets/issues/5175 | null | false |
1,427,216,416 | 5,174 | Preserve None in list type cast in PyArrow 10 | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-28T12:48:30 | 2022-10-28T13:15:33 | 2022-10-28T13:13:18 | The `ListArray` type in PyArrow 10.0.0 supports the `mask` parameter, which allows us to preserve Nones in nested lists in `cast` instead of replacing them with empty lists.
Fix https://github.com/huggingface/datasets/issues/3676 | mariosasko | https://github.com/huggingface/datasets/pull/5174 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5174",
"html_url": "https://github.com/huggingface/datasets/pull/5174",
"diff_url": "https://github.com/huggingface/datasets/pull/5174.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5174.patch",
"merged_at": "2022-10-28T13:13... | true |
1,425,880,441 | 5,173 | Raise ffmpeg warnings only once | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-27T15:58:33 | 2022-10-28T16:03:05 | 2022-10-28T16:00:51 | Our warnings looks nice now.
`librosa` warning that was raised at each decoding:
```
/usr/local/lib/python3.7/dist-packages/librosa/core/audio.py:165: UserWarning: PySoundFile failed. Trying audioread instead.
warnings.warn("PySoundFile failed. Trying audioread instead.")
```
is suppressed with `filterwarnin... | polinaeterna | https://github.com/huggingface/datasets/pull/5173 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5173",
"html_url": "https://github.com/huggingface/datasets/pull/5173",
"diff_url": "https://github.com/huggingface/datasets/pull/5173.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5173.patch",
"merged_at": "2022-10-28T16:00... | true |
1,425,523,114 | 5,172 | Inconsistency behavior between handling local file protocol and other FS protocols | open | [] | 2022-10-27T12:03:20 | 2024-05-08T19:31:13 | null | ### Describe the bug
These lines us used during load_from_disk:
```
if is_remote_filesystem(fs):
dest_dataset_dict_path = extract_path_from_uri(dataset_dict_path)
else:
fs = fsspec.filesystem("file")
dest_dataset_dict_path = dataset_dict_path
```
If a local FS is given, then it will the URL as th... | leoleoasd | https://github.com/huggingface/datasets/issues/5172 | null | false |
1,425,355,111 | 5,171 | Add PB and TB in convert_file_size_to_int | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-27T09:50:31 | 2022-10-27T12:14:27 | 2022-10-27T12:12:30 | null | lhoestq | https://github.com/huggingface/datasets/pull/5171 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5171",
"html_url": "https://github.com/huggingface/datasets/pull/5171",
"diff_url": "https://github.com/huggingface/datasets/pull/5171.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5171.patch",
"merged_at": "2022-10-27T12:12... | true |
1,425,301,835 | 5,170 | [Caching] Deterministic hashing of torch tensors | closed | [] | 2022-10-27T09:15:15 | 2022-11-02T17:18:43 | 2022-11-02T17:18:43 | Currently this fails
```python
import torch
from datasets.fingerprint import Hasher
t = torch.tensor([1.])
def func(x):
return t + x
hash1 = Hasher.hash(func)
t = torch.tensor([1.])
hash2 = Hasher.hash(func)
assert hash1 == hash2
```
Also as noticed in https://discuss.huggingface.co/t/dataset-ca... | lhoestq | https://github.com/huggingface/datasets/issues/5170 | null | false |
1,425,075,254 | 5,169 | Add "ipykernel" to list of `co_filename`s to remove | closed | [
"I don't know how I could add some tests for this, although jupyter is not among the dependencies so at least that would need to be added. If someone can tell a recommended way I will try to do it!",
"So testing by myself and looking around the jupyter codebase it looks like the `co_filename` of objects created w... | 2022-10-27T05:56:17 | 2022-11-02T15:46:00 | 2022-11-02T15:43:20 | Should resolve #5157 | gpucce | https://github.com/huggingface/datasets/pull/5169 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5169",
"html_url": "https://github.com/huggingface/datasets/pull/5169",
"diff_url": "https://github.com/huggingface/datasets/pull/5169.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5169.patch",
"merged_at": "2022-11-02T15:43... | true |
1,424,368,572 | 5,168 | Fix CI require beam | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I'm merging this PR because it is quite a trivial fix and this is required by:\r\n- #5166"
] | 2022-10-26T16:49:33 | 2022-10-27T09:25:19 | 2022-10-27T09:23:26 | This PR:
- Fixes the CI `require_beam`: before it was requiring PyTorch instead
```python
def require_beam(test_case):
if not config.TORCH_AVAILABLE:
test_case = unittest.skip("test requires PyTorch")(test_case)
return test_case
```
- Fixes a missing `require_beam` in `test_beam_base... | albertvillanova | https://github.com/huggingface/datasets/pull/5168 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5168",
"html_url": "https://github.com/huggingface/datasets/pull/5168",
"diff_url": "https://github.com/huggingface/datasets/pull/5168.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5168.patch",
"merged_at": "2022-10-27T09:23... | true |
1,424,124,477 | 5,167 | Add ffmpeg4 installation instructions in warnings | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"To make it warn only once, feel free to use a global counter in python - and if the warning has already been done, you don't do it again",
"> Added the same formatting for the error message :)\r\n\r\nnice!! thank you! \r\n\r\n> Oh ... | 2022-10-26T14:21:14 | 2022-10-27T09:01:12 | 2022-10-27T08:58:58 | Adds instructions on how to install `ffmpeg=4` on Linux (relevant for Colab users).
Looks pretty ugly because I didn't find a way to check `ffmpeg` version from python (without `subprocess.call()`; `ctypes.util.find_library` doesn't work`), so the warning is raised on each decoding. Any suggestions on how to make it... | polinaeterna | https://github.com/huggingface/datasets/pull/5167 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5167",
"html_url": "https://github.com/huggingface/datasets/pull/5167",
"diff_url": "https://github.com/huggingface/datasets/pull/5167.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5167.patch",
"merged_at": "2022-10-27T08:58... | true |
1,423,629,582 | 5,166 | Support dill 0.3.6 | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I think it hasn't been merged ? https://github.com/uqfoundation/dill/pull/501\r\n\r\nThough I can see that the CI is green because it uses dill 0.3.1.1 - we should probably fix the dill version in both CIs:\r\n- use 0.3.1.1 for the C... | 2022-10-26T08:24:59 | 2022-10-28T05:41:05 | 2022-10-28T05:38:14 | This PR:
- ~~Unpins dill to allow installing dill>=0.3.6~~
- ~~Removes the fix on dill for >=0.3.6 because they implemented a deterministic mode (to be confirmed by @anivegesana)~~
- Pins dill<0.3.7 to allow latest dill 0.3.6
- Implements a fix for dill `save_function` for dill 0.3.6
- Additionally had to implemen... | albertvillanova | https://github.com/huggingface/datasets/pull/5166 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5166",
"html_url": "https://github.com/huggingface/datasets/pull/5166",
"diff_url": "https://github.com/huggingface/datasets/pull/5166.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5166.patch",
"merged_at": "2022-10-28T05:38... | true |
1,423,616,677 | 5,165 | Memory explosion when trying to access 4d tensors in datasets cast to torch or np | open | [] | 2022-10-26T08:14:47 | 2022-10-26T08:14:47 | null | ### Describe the bug
When trying to access an item by index, in a datasets.Dataset cast to torch/np using `set_format` or `with_format`, we get a memory explosion if the item contains 4d (or above) tensors.
### Steps to reproduce the bug
MWE:
```python
from datasets import load_dataset
import numpy as np
de... | clefourrier | https://github.com/huggingface/datasets/issues/5165 | null | false |
1,422,813,247 | 5,164 | WIP: drop labels in Image and Audio folders by default | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"close in favor of https://github.com/huggingface/datasets/pull/5192"
] | 2022-10-25T17:21:49 | 2022-11-16T14:21:16 | 2022-11-02T14:03:02 | will fix https://github.com/huggingface/datasets/issues/5153 and redundant labels displaying for most of the images datasets on the Hub (which are used just to store files)
TODO: discuss adding `drop_labels` (and `drop_metadata`) params to yaml | polinaeterna | https://github.com/huggingface/datasets/pull/5164 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5164",
"html_url": "https://github.com/huggingface/datasets/pull/5164",
"diff_url": "https://github.com/huggingface/datasets/pull/5164.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5164.patch",
"merged_at": null
} | true |
1,422,540,337 | 5,163 | Reduce default max `writer_batch_size` | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-25T14:14:52 | 2022-10-27T12:19:27 | 2022-10-27T12:16:47 | Reduce the default writer_batch_size from 10k to 1k examples. Additionally, align the default values of `batch_size` and `writer_batch_size` in `Dataset.cast` with the values from the corresponding docstring. | mariosasko | https://github.com/huggingface/datasets/pull/5163 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5163",
"html_url": "https://github.com/huggingface/datasets/pull/5163",
"diff_url": "https://github.com/huggingface/datasets/pull/5163.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5163.patch",
"merged_at": "2022-10-27T12:16... | true |
1,422,461,112 | 5,162 | Pip-compile: Could not find a version that matches dill<0.3.6,>=0.3.6 | closed | [
"Thanks for reporting, @Rijgersberg.\r\n\r\nWe were waiting for the release of `dill` 0.3.6, that happened 2 days ago (24 Oct 2022): https://github.com/uqfoundation/dill/releases/tag/dill-0.3.6\r\n- See comment: https://github.com/huggingface/datasets/pull/4397#discussion_r880629543\r\n\r\nAlso `multiprocess` 0.70.... | 2022-10-25T13:23:50 | 2022-11-14T08:25:37 | 2022-10-28T05:38:15 | ### Describe the bug
When using `pip-compile` (part of `pip-tools`) to generate a pinned requirements file that includes `datasets`, a version conflict of `dill` appears.
It is caused by a transitive dependency conflict between `datasets` and `multiprocess`.
### Steps to reproduce the bug
```bash
$ echo "dataset... | Rijgersberg | https://github.com/huggingface/datasets/issues/5162 | null | false |
1,422,371,748 | 5,161 | Dataset canβt cache modelβs outputs | closed | [
"Addressed in https://github.com/huggingface/datasets/pull/5191 (torch.Tensor objects now produce deterministic hashes)"
] | 2022-10-25T12:19:00 | 2022-11-03T16:12:52 | 2022-11-03T16:12:51 | ### Describe the bug
Hi,
I try to cache some outputs of teacher model( Knowledge Distillation ) by using map function of Dataset library, while every time I run my code, I still recompute all the sequences. I tested Bert Model like this, I got different hash every single run, so any idea to deal with this?
### Ste... | jongjyh | https://github.com/huggingface/datasets/issues/5161 | null | false |
1,422,193,938 | 5,160 | Automatically add filename for image/audio folder | open | [
"Also cc @anton-l ",
"BTW the exact same holds true for the audio folder",
"I'm fine with adding a new column with the file name personally. Not sure how breaking this is though",
"@patrickvonplaten do you mean just filename or full relative path inside the repo?\r\nI think it shouldn't be breaking, at least ... | 2022-10-25T09:56:49 | 2022-10-26T16:51:46 | null | ### Feature request
When creating a custom audio of image dataset, it would be great to automatically have access to the filename. It should be both:
a) Automatically displayed in the viewer
b) Automatically added as a column to the dataset when doing `load_dataset`
In `diffusers` our test rely quite heavily on i... | patrickvonplaten | https://github.com/huggingface/datasets/issues/5160 | null | false |
1,422,172,080 | 5,159 | fsspec lock reset in multiprocessing | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-25T09:41:59 | 2022-11-03T20:51:15 | 2022-11-03T20:48:53 | `fsspec` added a clean way of resetting its lock - instead of doing it manually | lhoestq | https://github.com/huggingface/datasets/pull/5159 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5159",
"html_url": "https://github.com/huggingface/datasets/pull/5159",
"diff_url": "https://github.com/huggingface/datasets/pull/5159.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5159.patch",
"merged_at": "2022-11-03T20:48... | true |
1,422,059,287 | 5,158 | Fix language and license tag names in all Hub datasets | closed | [
"There are currently 402 datasets with deprecated \"languages\" or \"licenses\".",
"hey @albertvillanova ,i would love to work on this issue if you like.",
"Hi @ayushthe1, thanks for your offer.\r\n\r\nBut as you can see, I self-assigned this issue.\r\n\r\nI have already fixed 200 out of the 402 datasets. My sc... | 2022-10-25T08:19:29 | 2022-10-25T11:27:26 | 2022-10-25T10:42:19 | While working on this:
- #5137
we realized there are still many datasets with deprecated "languages" and "licenses" tag names (instead of "language" and "license").
This is a blocking issue: no subsequent PR can be opened to modify their metadata: a ValueError will be thrown.
We should fix the "language" and ... | albertvillanova | https://github.com/huggingface/datasets/issues/5158 | null | false |
1,421,703,577 | 5,157 | Consistent caching between python and jupyter | closed | [
"Hi ! Maybe it's possible to have a consistent hash for a function defined in `__main__` and a function define in a notebook.\r\n\r\nHowever for functions imported from another location, pickle uses the location to identify the code, so in that case we can't do much I believe.\r\n\r\nWould it be ok for you if we on... | 2022-10-25T01:34:33 | 2022-11-02T15:43:22 | 2022-11-02T15:43:22 | ### Feature request
I hope this is not my mistake, currently if I use `load_dataset` from a python session on a custom dataset to do the preprocessing, it will be saved in the cache and in other python sessions it will be loaded from the cache, however calling the same from a jupyter notebook does not work, meaning th... | gpucce | https://github.com/huggingface/datasets/issues/5157 | null | false |
1,421,667,125 | 5,156 | Unable to download dataset using Azure Data Lake Gen 2 | closed | [
"Hi ! From the `adlfs` docs, there are two filesystems you can use:\r\n> To use the Gen1 filesystem:\r\n> - known_implementations[βadlβ] = {βclassβ: βadlfs.AzureDatalakeFileSystemβ}\r\n> \r\n> To use the Gen2 filesystem:\r\n> - known_implementations[βabfsβ] = {βclassβ: βadlfs.AzureBlobFileSystemβ}\r\n\r\nIf I'm no... | 2022-10-25T00:43:18 | 2024-02-15T09:48:36 | 2022-11-17T23:37:08 | ### Describe the bug
When using the DatasetBuilder method with the credentials for the cloud storage Azure Data Lake (adl) Gen2, the following error is showed:
```
Traceback (most recent call last):
File "download_hf_dataset.py", line 143, in <module>
main()
File "download_hf_dataset.py", line 102, in mai... | clarissesimoes | https://github.com/huggingface/datasets/issues/5156 | null | false |
1,421,278,748 | 5,155 | TextConfig: added "errors" | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thanks for adding this ! You can fix the CI by formatting your code using the `make style` command :)",
"[**@lhoestq**](https://github.com/lhoestq) commented on [Oct 27, 2022, 4:08 PM GMT+3:30](https://github.com/huggingface/datase... | 2022-10-24T18:56:52 | 2022-11-03T13:38:13 | 2022-11-03T13:35:35 | This patch adds the ability to set the `errors` option of `open` for loading text datasets. I needed it because some data I had scraped had bad bytes in it, so I needed `errors='ignore'`. | NightMachinery | https://github.com/huggingface/datasets/pull/5155 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5155",
"html_url": "https://github.com/huggingface/datasets/pull/5155",
"diff_url": "https://github.com/huggingface/datasets/pull/5155.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5155.patch",
"merged_at": "2022-11-03T13:35... | true |
1,421,161,992 | 5,154 | Test latest fsspec in CI | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"actually the latest fsspec is already installed "
] | 2022-10-24T17:18:13 | 2023-09-24T10:06:06 | 2022-10-25T09:30:45 | Following the discussion in https://discuss.huggingface.co/t/attributeerror-module-fsspec-has-no-attribute-asyn/19255 I think we need to test the latest fsspec in the CI | lhoestq | https://github.com/huggingface/datasets/pull/5154 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5154",
"html_url": "https://github.com/huggingface/datasets/pull/5154",
"diff_url": "https://github.com/huggingface/datasets/pull/5154.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5154.patch",
"merged_at": null
} | true |
1,420,833,457 | 5,153 | default Image/AudioFolder infers labels when there is no metadata files even if there is only one dir | closed | [
"Makes sense! For the last structure, we could count the path segments (delimited by \"/\" for URLs and `os.sep` for local paths) to ensure all inferred labels are on the same level. Otherwise, I think it's safe to assume they are meaningless and ignore them.\r\n"
] | 2022-10-24T13:28:18 | 2022-11-15T16:31:10 | 2022-11-15T16:31:09 | ### Describe the bug
By default FolderBasedBuilder infers labels if there is not metadata files, even if it's meaningless (for example, they are in a single directory or in the root folder, see this repo as an example: https://huggingface.co/datasets/patrickvonplaten/audios
As this is a corner case for quick expl... | polinaeterna | https://github.com/huggingface/datasets/issues/5153 | null | false |
1,420,808,919 | 5,152 | refactor FolderBasedBuilder and Image/AudioFolder tests | open | [] | 2022-10-24T13:11:52 | 2022-10-24T13:11:52 | null | Tests for FolderBasedBuilder, ImageFolder and AudioFolder are mostly duplicating each other. They need to be refactored and Audio/ImageFolder should have only tests specific to the loader. | polinaeterna | https://github.com/huggingface/datasets/issues/5152 | null | false |
1,420,791,163 | 5,151 | Add support to create different configs with `push_to_hub` (+ inferring configs from directories with package managers?) | open | [
"also asked in https://discuss.huggingface.co/t/create-multiple-dataset-configs-with-push-to-hub-method/25480"
] | 2022-10-24T12:59:18 | 2022-11-04T14:55:20 | null | Now one can push only different splits within one default config of a dataset.
Would be nice to allow something like:
```
ds.push_to_hub(repo_name, config=config_name)
```
I'm not sure, but this will probably require changes in `data_files.py` patterns. If so, it would also allow to create different configs fo... | polinaeterna | https://github.com/huggingface/datasets/issues/5151 | null | false |
1,420,684,999 | 5,150 | Problems after upgrading to 2.6.1 | open | [
"Hi! I can't reproduce the error following these steps. Can you please provide a reproducible example?",
"I faced the same issue:\r\n\r\n### Repro\r\n```\r\n!pip install datasets==2.6.1\r\nimport datasets as Dataset\r\ndataset = Dataset.from_pandas(dataframe)\r\ndataset.save_to_disk(local)\r\n\r\n!pip install dat... | 2022-10-24T11:32:36 | 2024-05-12T07:40:03 | null | ### Describe the bug
Loading a dataset_dict from disk with `load_from_disk` is now creating a `KeyError "length"` that was not occurring in v2.5.2.
Context:
- Each individual dataset in the dict is created with `Dataset.from_pandas`
- The dataset_dict is create from a dict of `Dataset`s, e.g., `DatasetDict({"tr... | pietrolesci | https://github.com/huggingface/datasets/issues/5150 | null | false |
1,420,415,639 | 5,149 | Make iter_files deterministic | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-24T08:16:27 | 2022-10-27T09:53:23 | 2022-10-27T09:51:09 | Fix #5145. | albertvillanova | https://github.com/huggingface/datasets/pull/5149 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5149",
"html_url": "https://github.com/huggingface/datasets/pull/5149",
"diff_url": "https://github.com/huggingface/datasets/pull/5149.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5149.patch",
"merged_at": "2022-10-27T09:51... | true |
1,420,219,222 | 5,148 | Cannot find the rvl_cdip dataset | closed | [
"Hi, @santule.\r\n\r\nWe have transferred all dataset scripts from GitHub to the Hugging Face Hub: https://huggingface.co/datasets\r\n- Concretely, you have \"rvl_cdip\" here: https://huggingface.co/datasets/rvl_cdip\r\n\r\nTo be able to load them, you should update your `datasets` library:\r\n```\r\npip install -U... | 2022-10-24T04:57:42 | 2022-10-24T12:23:47 | 2022-10-24T06:25:28 | Hi,
I am trying to use load_dataset to load the official "rvl_cdip" dataset but getting an error.
dataset = load_dataset("rvl_cdip")
Couldn't find 'rvl_cdip' on the Hugging Face Hub either: FileNotFoundError: Couldn't find the file at https://raw.githubusercontent.com/huggingface/datasets/master/datasets/rvl_cdi... | santule | https://github.com/huggingface/datasets/issues/5148 | null | false |
1,419,522,275 | 5,147 | Allow ignoring kwargs inside fn_kwargs during dataset.map's fingerprinting | open | [
"Hi ! In the `transformers` issue the object to not hash is a `Pool` - I think you can instantiate it inside your function instead of passing it as a parameter. It's good practice that your function and all its fn_kwargs are picklable, in case you want to parallelize `map` using `num_proc>1`\r\n\r\nFor the other ca... | 2022-10-22T21:46:38 | 2022-11-01T22:19:07 | null | ### Feature request
`dataset.map` accepts a `fn_kwargs` that is passed to `fn`. Currently, the whole `fn_kwargs` is used by `fingerprint_transform` to calculate the new fingerprint.
I'd like to be able to inform `fingerprint_transform` which `fn_kwargs` shoud/shouldn't be taken into account during hashing.
Of co... | falcaopetri | https://github.com/huggingface/datasets/issues/5147 | null | false |
1,418,331,282 | 5,146 | Delete duplicate issue template file | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-21T13:18:46 | 2022-10-21T13:52:30 | 2022-10-21T13:50:04 | A conflict between two PRs:
- #5116
- #5136
was not properly resolved, resulting in a duplicate issue template.
This PR removes the duplicate template. | albertvillanova | https://github.com/huggingface/datasets/pull/5146 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5146",
"html_url": "https://github.com/huggingface/datasets/pull/5146",
"diff_url": "https://github.com/huggingface/datasets/pull/5146.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5146.patch",
"merged_at": "2022-10-21T13:50... | true |
1,418,005,452 | 5,145 | Dataset order is not deterministic with ZIP archives and `iter_files` | closed | [
"Thanks for reporting ! The issue doesn't come from shuffling, but from `beans` row order not being deterministic:\r\n\r\nhttps://huggingface.co/datasets/beans/blob/main/beans.py uses `dl_manager.iter_files` on ZIP archives and the file order doesn't seen to be deterministic and changes across machines",
"Thank y... | 2022-10-21T09:00:03 | 2022-10-27T09:51:49 | 2022-10-27T09:51:10 | ### Describe the bug
For the `beans` dataset (did not try on other), the order of samples is not the same on different machines. Tested on my local laptop, github actions machine, and ec2 instance. The three yield a different order.
### Steps to reproduce the bug
In a clean docker container or conda environmen... | fxmarty | https://github.com/huggingface/datasets/issues/5145 | null | false |
1,417,974,731 | 5,144 | Inconsistent documentation on map remove_columns | closed | [
"Thanks for reporting, @zhaowei-wang-nlp.\r\n\r\nYou are right, the documentation is confusing on the behavior of `remove_columns`. We should better explain it. ",
"This is a duplicate of https://github.com/huggingface/datasets/issues/2343.",
"I'm closing this issue because as @mariosasko pointed out, it is a d... | 2022-10-21T08:37:53 | 2022-11-15T14:15:10 | 2022-11-15T14:15:10 | ### Describe the bug
The page [process](https://huggingface.co/docs/datasets/process) says this about the parameter `remove_columns` of the function `map`:
When you remove a column, it is only removed after the example has been provided to the mapped function.
So it seems that the `remove_columns` parameter remo... | zhaowei-wang-nlp | https://github.com/huggingface/datasets/issues/5144 | null | false |
1,416,837,186 | 5,143 | DownloadManager Git LFS support | closed | [
"Hey ! Actually it works, just pass the right URL ;)\r\nThe URL must be the one with β/resolve/β\r\n\r\ne.g. https://huggingface.co/datasets/imagenet-1k/resolve/main/data/test_images.tar.gz\r\n\r\nYou can even pass a relative path to the dl_manager instead, like `dl_manager.download(\"data/test_images.tar.gz\")`",
... | 2022-10-20T15:29:29 | 2022-10-20T17:17:10 | 2022-10-20T17:17:10 | ### Feature request
Maybe I'm mistaken but the `DownloadManager` does not support extracting git lfs files out of the box right?
Using `dl_manager.download()` or `dl_manager.download_and_extract()` still returns lfs files afaict.
Is there a good way to write a dataset loading script for a repo with lfs files?
##... | Muennighoff | https://github.com/huggingface/datasets/issues/5143 | null | false |
1,416,317,678 | 5,142 | Deprecate num_proc parameter in DownloadManager.extract | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Hey @mariosasko . Can you please help me with why the tests keep failing. I have reviewed the code changes multiple times but can't spot any mistakes. ",
"You can fix this failure by formatting your code with the `make style` comm... | 2022-10-20T09:52:52 | 2022-10-25T18:06:56 | 2022-10-25T15:56:45 | fixes #5132 : Deprecated the `num_proc` parameter in `DownloadManager.extract` by passing `num_proc` parameter to `map_nested` . | ayushthe1 | https://github.com/huggingface/datasets/pull/5142 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5142",
"html_url": "https://github.com/huggingface/datasets/pull/5142",
"diff_url": "https://github.com/huggingface/datasets/pull/5142.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5142.patch",
"merged_at": "2022-10-25T15:56... | true |
1,415,479,438 | 5,141 | Raise ImportError instead of OSError | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thanks @mariosasko ,i commited the changes as you said.\r\n\r\n"
] | 2022-10-19T19:30:05 | 2022-10-25T15:59:25 | 2022-10-25T15:56:58 | fixes #5134 : Replaced OSError with ImportError if required extraction library is not installed. | ayushthe1 | https://github.com/huggingface/datasets/pull/5141 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5141",
"html_url": "https://github.com/huggingface/datasets/pull/5141",
"diff_url": "https://github.com/huggingface/datasets/pull/5141.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5141.patch",
"merged_at": "2022-10-25T15:56... | true |
1,415,075,530 | 5,140 | Make the KeyHasher FIPS compliant | closed | [] | 2022-10-19T14:25:52 | 2022-11-07T16:20:43 | 2022-11-07T16:20:43 | MD5 is not FIPS compliant thus I am proposing this minimal change to make datasets package FIPS compliant | vvalouch | https://github.com/huggingface/datasets/pull/5140 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5140",
"html_url": "https://github.com/huggingface/datasets/pull/5140",
"diff_url": "https://github.com/huggingface/datasets/pull/5140.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5140.patch",
"merged_at": null
} | true |
1,414,642,723 | 5,137 | Align task tags in dataset metadata | closed | [
"I removed all the invalid task_ids in datasts without namespace, based on the <s>(internal)</s> types.ts",
"(Types.ts is not internal it's public)",
"I have opened PRs to fix the task_ids in all datasets within a namespace as well.\r\n\r\nWorking on task_categories...",
"For future reference: this fix had so... | 2022-10-19T09:41:42 | 2022-11-10T05:25:58 | 2022-10-25T06:17:00 | ## Describe
Once we have agreed on a common naming for task tags for all open source projects, we should align on them.
## Steps
- [x] Align task tags in canonical datasets
- [x] task_categories: 4 datasets
- [x] task_ids (by @lhoestq)
- [x] Open PRs in community datasets
- [x] task_categories: 451 datas... | albertvillanova | https://github.com/huggingface/datasets/issues/5137 | null | false |
1,414,492,139 | 5,136 | Update docs once dataset scripts transferred to the Hub | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-19T07:58:27 | 2022-10-20T08:12:21 | 2022-10-20T08:10:00 | Todo:
- [x] Update docs:
- [x] Datasets on GitHub (legacy)
- [x] Load: offline
- [x] About dataset load:
- [x] Maintaining integrity
- [x] Security
- [x] Update docstrings:
- [x] Inspect:
- [x] get_dataset_config_info
- [x] get_dataset_split_names
- [x] Load:
- [x] dataset_modu... | albertvillanova | https://github.com/huggingface/datasets/pull/5136 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5136",
"html_url": "https://github.com/huggingface/datasets/pull/5136",
"diff_url": "https://github.com/huggingface/datasets/pull/5136.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5136.patch",
"merged_at": "2022-10-20T08:10... | true |
1,414,413,519 | 5,135 | Update docs once dataset scripts transferred to the Hub | closed | [] | 2022-10-19T06:58:19 | 2022-10-20T08:10:01 | 2022-10-20T08:10:01 | ## Describe the bug
As discussed in:
- https://github.com/huggingface/hub-docs/pull/423#pullrequestreview-1146083701
we should update our docs once dataset scripts have been transferred to the Hub (and removed from GitHub):
- #4974
Concretely:
- [x] Datasets on GitHub (legacy): https://huggingface.co/docs/dat... | albertvillanova | https://github.com/huggingface/datasets/issues/5135 | null | false |
1,413,623,687 | 5,134 | Raise ImportError instead of OSError if required extraction library is not installed | closed | [
"hey ,i would like to work on this issue . Please assign it to me.",
"hey @mariosasko , i made a pr for this issue. Could you please review it.\r\nAlso i found multiple `OSError` in `extract.py` file which i thought could be replaced too but wasn't sure about them.\r\nPlease do tell if that also needs to be done.... | 2022-10-18T17:53:46 | 2022-10-25T15:56:59 | 2022-10-25T15:56:59 | According to the official Python docs, `OSError` should be thrown in the following situations:
> This exception is raised when a system function returns a system-related error, including I/O failures such as βfile not foundβ or βdisk fullβ (not for illegal argument types or other incidental errors).
Hence, it makes... | mariosasko | https://github.com/huggingface/datasets/issues/5134 | null | false |
1,413,623,462 | 5,133 | Tensor operation not functioning in dataset mapping | closed | [
"Hi! The Torch ops in your snippet are not equivalent to the NumPy ones, hence the difference. You can get the same behavior by replacing the line `feature = torch.mean(feature, dim=1)` with `feature = feature.squeeze().mean(1)` .",
"> Hi! The Torch ops in your snippet are not equivalent to the NumPy ones, hence ... | 2022-10-18T17:53:35 | 2022-10-19T04:15:45 | 2022-10-19T04:15:44 | ## Describe the bug
I'm doing a torch.mean() operation in data preprocessing, and it's not working.
## Steps to reproduce the bug
```
from transformers import pipeline
import torch
import numpy as np
from datasets import load_dataset
device = 'cuda:0'
raw_dataset = load_dataset("glue", "sst2")
feature_extra... | xinghaow99 | https://github.com/huggingface/datasets/issues/5133 | null | false |
1,413,607,306 | 5,132 | Depracate `num_proc` parameter in `DownloadManager.extract` | closed | [
"I can take this! #self-assign",
"#self-assign",
"@lazarust i'm already working on this issue :smile: ",
"#self-assign",
"hey @mariosasko , i made a pr for this issue. Could you please review it."
] | 2022-10-18T17:41:05 | 2022-10-25T15:56:46 | 2022-10-25T15:56:46 | The `num_proc` parameter is only present in `DownloadManager.extract` but not in `StreamingDownloadManager.extract`, making it impossible to support streaming in the dataset scripts that use it (`openwebtext` and `the_pile_stack_exchange`). We can avoid this situation by deprecating this parameter and passing `Download... | mariosasko | https://github.com/huggingface/datasets/issues/5132 | null | false |
1,413,534,863 | 5,131 | WikiText 103 tokenizer hangs | closed | [
"any updates on this? It happens to me on [OpenWikiText-20%](https://huggingface.co/datasets/Bingsu/openwebtext_20p) dataset, but not on [OpenWebText-10k](https://huggingface.co/datasets/stas/openwebtext-10k). This is really strange because I don't change anything else in my running script.\r\n\r\ntransformers vers... | 2022-10-18T16:44:00 | 2023-08-08T08:42:40 | 2023-07-21T14:41:51 | See issue here: https://github.com/huggingface/transformers/issues/19702 | TrentBrick | https://github.com/huggingface/datasets/issues/5131 | null | false |
1,413,435,000 | 5,130 | Avoid extra cast in `class_encode_column` | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-18T15:31:24 | 2022-10-19T11:53:02 | 2022-10-19T11:50:46 | Pass the updated features to `map` to avoid the `cast` in `class_encode_column`. | mariosasko | https://github.com/huggingface/datasets/pull/5130 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5130",
"html_url": "https://github.com/huggingface/datasets/pull/5130",
"diff_url": "https://github.com/huggingface/datasets/pull/5130.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5130.patch",
"merged_at": "2022-10-19T11:50... | true |
1,413,031,664 | 5,129 | unexpected `cast` or `class_encode_column` result after `rename_column` | closed | [
"Hi! Unfortunately, I can't reproduce this issue locally (in Python 3.7/3.10) or in Colab. I would assume this is due to a bug we fixed in the latest release, but your version is up-to-date, so I'm not sure if there is something we can do to help...",
"Hi, ζΉεδΈ. I tried running the code with exact the same configu... | 2022-10-18T11:15:24 | 2022-10-19T03:02:26 | 2022-10-19T03:02:26 | ## Describe the bug
When invoke `cast` or `class_encode_column` to a colunm renamed by `rename_column` , it will convert all the variables in this column into one variable. I also run this script in version 2.5.2, this bug does not appear. So I switched to the older version.
## Steps to reproduce the bug
```python... | quaeast | https://github.com/huggingface/datasets/issues/5129 | null | false |
1,412,783,855 | 5,128 | Make filename matching more robust | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"> I think we should also modify one of the metadata files in the `folder_based_builder` tests to make sure \"./\" is ignored now in the `file_name`\r\n\r\n@mariosasko what do you mean here? I'm not sure which metadata file I should m... | 2022-10-18T08:22:48 | 2022-10-28T13:07:38 | 2022-10-28T13:05:06 | Fix #5046 | riccardobucco | https://github.com/huggingface/datasets/pull/5128 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5128",
"html_url": "https://github.com/huggingface/datasets/pull/5128",
"diff_url": "https://github.com/huggingface/datasets/pull/5128.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5128.patch",
"merged_at": "2022-10-28T13:05... | true |
1,411,897,544 | 5,127 | [WIP] WebDataset export | closed | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5127). All of your documentation changes will be reflected on that endpoint.",
"Should we close this PR?"
] | 2022-10-17T16:50:22 | 2024-01-11T06:27:04 | 2024-01-08T14:25:43 | I added a first draft of the `IterableDataset.to_wds` method.
You can use it to savea dataset loaded in streamign mode as a webdataset locally.
The API can be further improved to allow to export to a cloud storage like the HF Hub.
I also included sharding with a default max shard size of 500MB (uncompressed), an... | lhoestq | https://github.com/huggingface/datasets/pull/5127 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5127",
"html_url": "https://github.com/huggingface/datasets/pull/5127",
"diff_url": "https://github.com/huggingface/datasets/pull/5127.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5127.patch",
"merged_at": null
} | true |
1,411,757,124 | 5,126 | Fix class name of symbolic link | closed | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5126). All of your documentation changes will be reflected on that endpoint.",
"I have removed the reference to the Issue in the PR title, so that we avoid to have both references (to the issue and to the PR) in the merge commi... | 2022-10-17T15:11:02 | 2022-11-14T14:40:18 | 2022-11-14T14:40:18 | Fix #5098 | riccardobucco | https://github.com/huggingface/datasets/pull/5126 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5126",
"html_url": "https://github.com/huggingface/datasets/pull/5126",
"diff_url": "https://github.com/huggingface/datasets/pull/5126.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5126.patch",
"merged_at": "2022-11-14T14:40... | true |
1,411,602,813 | 5,125 | Add `pyproject.toml` for `black` | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-17T13:38:47 | 2024-11-20T13:36:11 | 2022-10-17T14:21:09 | Add `pyproject.toml` as a config file for the `black` tool to support VS Code's auto-formatting on save (and to be more consistent with the other HF projects).
| mariosasko | https://github.com/huggingface/datasets/pull/5125 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5125",
"html_url": "https://github.com/huggingface/datasets/pull/5125",
"diff_url": "https://github.com/huggingface/datasets/pull/5125.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5125.patch",
"merged_at": "2022-10-17T14:21... | true |
1,411,159,725 | 5,124 | Install tensorflow-macos dependency conditionally | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-17T08:45:08 | 2022-10-19T09:12:17 | 2022-10-19T09:10:06 | Fix #5118. | albertvillanova | https://github.com/huggingface/datasets/pull/5124 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5124",
"html_url": "https://github.com/huggingface/datasets/pull/5124",
"diff_url": "https://github.com/huggingface/datasets/pull/5124.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5124.patch",
"merged_at": "2022-10-19T09:10... | true |
1,410,828,756 | 5,123 | datasets freezes with streaming mode in multiple-gpu | open | [
"@lhoestq I tested the script without accelerator, and I confirm this is due to datasets part as this gets similar results without accelerator.",
"Hi ! You said it works on 1 GPU but doesn't wortk without accelerator - what's the difference between running on 1 GPU and running without accelerator in your case ?"... | 2022-10-17T03:28:16 | 2023-05-14T06:55:20 | null | ## Describe the bug
Hi. I am using this dataloader, which is for processing large datasets in streaming mode mentioned in one of examples of huggingface. I am using it to read c4: https://github.com/huggingface/transformers/blob/b48ac1a094e572d6076b46a9e4ed3e0ebe978afc/examples/research_projects/codeparrot/scripts/cod... | jackfeinmann5 | https://github.com/huggingface/datasets/issues/5123 | null | false |
1,410,732,403 | 5,122 | Add warning | closed | [
"As mentioned in https://github.com/huggingface/datasets/issues/5105 I think we just need to keep the existing files instead of deleting them.\r\nThe `dataset_info.json` file contains the split names anyway, so we know which files belong to the dataset, and which ones don't."
] | 2022-10-17T01:30:37 | 2022-11-05T12:23:53 | 2022-11-05T12:23:53 | Fixes: #5105
I think removing the directory with warning is a better solution for this issue. Because if we decide to keep existing files in directory, then we should deal with the case providing same directory for several datasets! Which we know is not possible since `dataset_info.json` exists in that directory. | Salehbigdeli | https://github.com/huggingface/datasets/pull/5122 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5122",
"html_url": "https://github.com/huggingface/datasets/pull/5122",
"diff_url": "https://github.com/huggingface/datasets/pull/5122.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5122.patch",
"merged_at": null
} | true |
1,410,681,067 | 5,121 | Bugfix ignore function when creating new_fingerprint for caching | closed | [
"Adding \"function\" to the kwargs to ignore when computing the fingerprint will break `map` caching. Indeed passing two different function would result in two different datasets that have the same fingerprint - and the cache wouldn't be able to distinguish them.\r\n\r\nE.g this code would reload ds1 from the cache... | 2022-10-17T00:03:43 | 2022-10-17T12:39:36 | 2022-10-17T12:39:36 | maybe fixes: #5109 | Salehbigdeli | https://github.com/huggingface/datasets/pull/5121 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5121",
"html_url": "https://github.com/huggingface/datasets/pull/5121",
"diff_url": "https://github.com/huggingface/datasets/pull/5121.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5121.patch",
"merged_at": null
} | true |
1,410,641,221 | 5,120 | Fix `tqdm` zip bug | closed | [
"@albertvillanova Thanks for your comment. What do you think about creating 2 `pbar` for each case? I see the `pbar_iterable` is initialized differently. Maybe `pbar` can also be initialized like that.",
"@albertvillanova Another solution I implemented is to change `pbar_iterable` and add the `zip` to it. I updat... | 2022-10-16T22:19:18 | 2022-10-23T10:27:53 | 2022-10-19T08:53:17 | This PR solves #5117, by wrapping the entire `zip` clause in tqdm.
For more information, please checkout this Stack Overflow thread:
https://stackoverflow.com/questions/41171191/tqdm-progressbar-and-zip-built-in-do-not-work-together | david1542 | https://github.com/huggingface/datasets/pull/5120 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5120",
"html_url": "https://github.com/huggingface/datasets/pull/5120",
"diff_url": "https://github.com/huggingface/datasets/pull/5120.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5120.patch",
"merged_at": "2022-10-19T08:53... | true |
1,410,561,363 | 5,119 | [TYPO] Update new_dataset_script.py | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-16T17:36:49 | 2022-10-19T09:48:19 | 2022-10-19T09:45:59 | null | cakiki | https://github.com/huggingface/datasets/pull/5119 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5119",
"html_url": "https://github.com/huggingface/datasets/pull/5119",
"diff_url": "https://github.com/huggingface/datasets/pull/5119.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5119.patch",
"merged_at": "2022-10-19T09:45... | true |
1,410,547,373 | 5,118 | Installing `datasets` on M1 computers | closed | [
"Thanks for reporting, @david1542."
] | 2022-10-16T16:50:08 | 2022-10-19T09:10:08 | 2022-10-19T09:10:08 | ## Describe the bug
I wanted to install `datasets` dependencies on my M1 (in order to start contributing to the project). However, I got an error regarding `tensorflow`.
On M1, `tensorflow-macos` needs to be installed instead. Can we add a conditional requirement, so that `tensorflow-macos` would be installed on M1... | david1542 | https://github.com/huggingface/datasets/issues/5118 | null | false |
1,409,571,346 | 5,117 | Progress bars have color red and never completed to 100% | closed | [
"Hi @echatzikyriakidis, thanks for submitting the issue.\r\nWhich shell are you using exactly? I tried to run the command you sent, but I don't see colors at all π§\r\n\r\nI tried from bash and zsh as well.",
"Hi @david1542 ,\r\n\r\nI use Google Colab.\r\n",
"Got it. I [created a PR](https://github.com/huggingf... | 2022-10-14T16:12:30 | 2024-06-19T19:03:42 | 2022-10-23T12:58:41 | ## Describe the bug
Progress bars after transformative operations turn in red and never be completed to 100%
## Steps to reproduce the bug
```python
from datasets import load_dataset
load_dataset('rotten_tomatoes', split='test').filter(lambda o: True)
```
## Expected results
Progress bar should be 100% an... | echatzikyriakidis | https://github.com/huggingface/datasets/issues/5117 | null | false |
1,409,549,471 | 5,116 | Use yaml for issue templates + revamp | closed | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-10-14T15:53:13 | 2022-10-19T13:05:49 | 2022-10-19T13:03:22 | Use YAML instead of markdown (more expressive) for the issue templates. In addition, update their structure/fields to be more aligned with Transformers.
PS: also removes the "add_dataset" PR template, as we no longer accept such PRs. | mariosasko | https://github.com/huggingface/datasets/pull/5116 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5116",
"html_url": "https://github.com/huggingface/datasets/pull/5116",
"diff_url": "https://github.com/huggingface/datasets/pull/5116.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5116.patch",
"merged_at": "2022-10-19T13:03... | true |
1,409,250,020 | 5,115 | Fix iter_batches | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I also ran the code in https://github.com/huggingface/datasets/issues/5111 and it works fine now :)",
"This is ready for review :)"
] | 2022-10-14T12:06:14 | 2022-10-14T15:02:15 | 2022-10-14T14:59:58 | The `pa.Table.to_reader()` method available in `pyarrow>=8.0.0` may return chunks of size < `max_chunksize`, therefore `iter_batches` can return batches smaller than the `batch_size` specified by the user
Therefore batched `map` couldn't always use batches of the right size, e.g. this fails because it runs only on o... | lhoestq | https://github.com/huggingface/datasets/pull/5115 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5115",
"html_url": "https://github.com/huggingface/datasets/pull/5115",
"diff_url": "https://github.com/huggingface/datasets/pull/5115.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5115.patch",
"merged_at": "2022-10-14T14:59... | true |
1,409,236,738 | 5,114 | load_from_disk with remote filesystem fails due to a wrong temporary local folder path | open | [
"Hi Hubert! Could you please probably create a publicly available `gs://` dataset link? I think this would be easier for others to directly start to debug.",
"What seems to work is to change the line to:\r\n```\r\nfs.download(src_dataset_path, dataset_path.parent.as_posix(), recursive=True)\r\n```"
] | 2022-10-14T11:54:53 | 2022-11-19T07:13:10 | null | ## Describe the bug
The function load_from_disk fails when using a remote filesystem because of a wrong temporary path generation in the load_from_disk method of arrow_dataset.py:
```python
if is_remote_filesystem(fs):
src_dataset_path = extract_path_from_uri(dataset_path)
dataset_path = Dataset._build... | bruno-hays | https://github.com/huggingface/datasets/issues/5114 | null | false |
1,409,207,607 | 5,113 | Fix filter indices when batched | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I think a patch release will be necessary.",
"I'm also fixing https://github.com/huggingface/datasets/issues/5111 which will lalso require a patch release"
] | 2022-10-14T11:30:03 | 2022-10-24T06:21:09 | 2022-10-14T12:11:44 | This PR fixes a bug introduced by:
- #5030
Fix #5112. | albertvillanova | https://github.com/huggingface/datasets/pull/5113 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5113",
"html_url": "https://github.com/huggingface/datasets/pull/5113",
"diff_url": "https://github.com/huggingface/datasets/pull/5113.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5113.patch",
"merged_at": "2022-10-14T12:11... | true |
1,409,143,409 | 5,112 | Bug with filtered indices | closed | [
"The issue is here:\r\nhttps://github.com/huggingface/datasets/blob/3ad9644b9a2e4558dd1d0f1e43c67658674e6228/src/datasets/arrow_dataset.py#L2964",
"@PartiallyTyped, @Muennighoff: the issue is fixed.\r\n\r\nWe are planning to make a patch release today.",
"Thanks a lot for the swift response!Β For a brief moment ... | 2022-10-14T10:35:47 | 2022-10-14T13:55:03 | 2022-10-14T12:11:45 | ## Describe the bug
As reported by @PartiallyTyped (and by @Muennighoff):
- https://github.com/huggingface/datasets/issues/5111#issuecomment-1278652524
There is an issue with the indices of a filtered dataset.
## Steps to reproduce the bug
```python
ds = Dataset.from_dict({"num": [0, 1, 2, 3]})
ds = ds.filte... | albertvillanova | https://github.com/huggingface/datasets/issues/5112 | null | false |
1,408,143,170 | 5,111 | map and filter not working properly in multiprocessing with the new release 2.6.0 | closed | [
"Same bug exists with `num_proc=1` on colab. `3.7.14 (default, Sep 8 2022, 00:06:44) [GCC 7.5.0]` ",
"Thanks for reporting, @loubnabnl and for the additional information, @PartiallyTyped.\r\n\r\nHowever, I'm not able to reproduce this issue, neither locally nor on Colab:\r\n```\r\nDataset({\r\n features: ['re... | 2022-10-13T17:00:55 | 2022-10-17T08:26:59 | 2022-10-14T14:59:59 | ## Describe the bug
When mapping is used on a dataset with more than one process, there is a weird behavior when trying to use `filter` , it's like only the samples from one worker are retrieved, one needs to specify the same `num_proc` in filter for it to work properly. This doesn't happen with `datasets` version 2.5... | loubnabnl | https://github.com/huggingface/datasets/issues/5111 | null | false |
1,407,434,706 | 5,109 | Map caching not working for some class methods | closed | [
"The hash used for caching is computed by pickling recursively the function passed to `map`. Maybe some objects don't have the same hash across sessions. In particular you can check the hash of your model using\r\n```python\r\nfrom datasets.fingerprint import Hasher\r\nobj = AutoModel.from_config(config=config, ad... | 2022-10-13T09:12:58 | 2022-10-17T10:38:45 | 2022-10-17T10:38:45 | ## Describe the bug
The cache loading is not working as expected for some class methods with a model stored in an attribute.
The new fingerprint for `_map_single` is not the same at each run. The hasher generate a different hash for the class method.
This comes from `dumps` function in `datasets.utils.py_utils` whic... | Mouhanedg56 | https://github.com/huggingface/datasets/issues/5109 | null | false |
1,407,044,107 | 5,108 | Fix a typo in arrow_dataset.py | closed | [] | 2022-10-13T02:33:55 | 2022-10-14T09:47:28 | 2022-10-14T09:47:27 | null | yangky11 | https://github.com/huggingface/datasets/pull/5108 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5108",
"html_url": "https://github.com/huggingface/datasets/pull/5108",
"diff_url": "https://github.com/huggingface/datasets/pull/5108.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5108.patch",
"merged_at": "2022-10-14T09:47... | true |
1,406,736,710 | 5,107 | Multiprocessed dataset builder | closed | [
"I would also like to add a test, but am not sure whether it should go into `test_builder` (more natural imo) or `test_load` (which already contains a lot of the things I have to import to run my current testing setup). For reference, what I run to test that it works looks like:\r\n\r\n```\r\nimport os\r\nfrom path... | 2022-10-12T19:59:17 | 2022-12-01T15:37:09 | 2022-11-09T17:11:43 | This PR adds the multiprocessing part of #2650 (but not the caching of already-computed arrow files). On the other side, loading of sharded arrow files still needs to be implemented (sharded parquet files can already be loaded). | TevenLeScao | https://github.com/huggingface/datasets/pull/5107 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5107",
"html_url": "https://github.com/huggingface/datasets/pull/5107",
"diff_url": "https://github.com/huggingface/datasets/pull/5107.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5107.patch",
"merged_at": "2022-11-09T17:11... | true |
1,406,635,758 | 5,106 | Fix task template reload from dict | closed | [
"_The documentation is not available anymore as the PR was closed or merged._",
"> Just wondering if there might be other data classes default values missed that could cause an issue... Apart from feature-like classes and tasks, I don't see any others though...\r\n\r\nI think we're good ! `asdict` is used on the ... | 2022-10-12T18:33:49 | 2022-10-13T09:59:07 | 2022-10-13T09:56:51 | Since #4926 the JSON dumps are simplified and it made task template dicts empty by default.
I fixed this by always including the task name which is needed to reload a task from a dict | lhoestq | https://github.com/huggingface/datasets/pull/5106 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5106",
"html_url": "https://github.com/huggingface/datasets/pull/5106",
"diff_url": "https://github.com/huggingface/datasets/pull/5106.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/5106.patch",
"merged_at": "2022-10-13T09:56... | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.