Dataset Viewer
Auto-converted to Parquet Duplicate
url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.83B
node_id
stringlengths
18
32
number
int64
1
6.09k
title
stringlengths
1
290
labels
list
state
stringclasses
2 values
locked
bool
1 class
milestone
dict
comments
int64
0
54
created_at
stringlengths
20
20
updated_at
stringlengths
20
20
closed_at
stringlengths
20
20
active_lock_reason
null
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
3 values
draft
bool
2 classes
pull_request
dict
is_pull_request
bool
2 classes
comments_text
sequence
https://api.github.com/repos/huggingface/datasets/issues/2771
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2771/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2771/comments
https://api.github.com/repos/huggingface/datasets/issues/2771/events
https://github.com/huggingface/datasets/pull/2771
963,257,036
MDExOlB1bGxSZXF1ZXN0NzA1OTExMDMw
2,771
[WIP][Common Voice 7] Add common voice 7.0
[]
closed
false
null
2
2021-08-07T16:01:10Z
2021-12-06T23:24:02Z
2021-12-06T23:24:02Z
null
This PR allows to load the new common voice dataset manually as explained when doing: ```python from datasets import load_dataset ds = load_dataset("./datasets/datasets/common_voice_7", "ab") ``` => ``` Please follow the manual download instructions: You need t...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2771/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2771/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2771.diff", "html_url": "https://github.com/huggingface/datasets/pull/2771", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2771.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2771" }
true
[ "Hi ! I think the name `common_voice_7` is fine :)\r\nMoreover if the dataset_infos.json is missing I'm pretty sure you don't need to specify `ignore_verifications=True`", "Hi, how about to add a new parameter \"version\" in the function load_dataset, something like: \r\n`load_dataset(\"common_voice\", \"lg\", ve...
https://api.github.com/repos/huggingface/datasets/issues/5603
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5603/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5603/comments
https://api.github.com/repos/huggingface/datasets/issues/5603/events
https://github.com/huggingface/datasets/pull/5603
1,607,143,509
PR_kwDODunzps5LJZzG
5,603
Don't compute checksums if not necessary in `datasets-cli test`
[]
closed
false
null
3
2023-03-02T16:42:39Z
2023-03-03T15:45:32Z
2023-03-03T15:38:28Z
null
we only need them if there exists a `dataset_infos.json`
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5603/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5603/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/5603.diff", "html_url": "https://github.com/huggingface/datasets/pull/5603", "merged_at": "2023-03-03T15:38:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/5603.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
https://api.github.com/repos/huggingface/datasets/issues/5111
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5111/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5111/comments
https://api.github.com/repos/huggingface/datasets/issues/5111/events
https://github.com/huggingface/datasets/issues/5111
1,408,143,170
I_kwDODunzps5T7o9C
5,111
map and filter not working properly in multiprocessing with the new release 2.6.0
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
14
2022-10-13T17:00:55Z
2022-10-17T08:26:59Z
2022-10-14T14:59:59Z
null
## Describe the bug When mapping is used on a dataset with more than one process, there is a weird behavior when trying to use `filter` , it's like only the samples from one worker are retrieved, one needs to specify the same `num_proc` in filter for it to work properly. This doesn't happen with `datasets` version 2.5...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5111/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5111/timeline
null
completed
null
null
false
[ "Same bug exists with `num_proc=1` on colab. `3.7.14 (default, Sep 8 2022, 00:06:44) [GCC 7.5.0]` ", "Thanks for reporting, @loubnabnl and for the additional information, @PartiallyTyped.\r\n\r\nHowever, I'm not able to reproduce this issue, neither locally nor on Colab:\r\n```\r\nDataset({\r\n features: ['re...
https://api.github.com/repos/huggingface/datasets/issues/2344
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2344/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2344/comments
https://api.github.com/repos/huggingface/datasets/issues/2344/events
https://github.com/huggingface/datasets/issues/2344
885,331,505
MDU6SXNzdWU4ODUzMzE1MDU=
2,344
Is there a way to join multiple datasets in one?
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
2
2021-05-10T23:16:10Z
2022-10-05T17:27:05Z
null
null
**Is your feature request related to a problem? Please describe.** I need to join 2 datasets, one that is in the hub and another I've created from my files. Is there an easy way to join these 2? **Describe the solution you'd like** Id like to join them with a merge or join method, just like pandas dataframes. **Add...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2344/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2344/timeline
null
null
null
null
false
[ "Hi ! We don't have `join`/`merge` on a certain column as in pandas.\r\nMaybe you can just use the [concatenate_datasets](https://huggingface.co/docs/datasets/package_reference/main_classes.html?highlight=concatenate#datasets.concatenate_datasets) function.\r\n", "Hi! You can use `datasets_sql` for that now. As o...
https://api.github.com/repos/huggingface/datasets/issues/3848
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3848/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3848/comments
https://api.github.com/repos/huggingface/datasets/issues/3848/events
https://github.com/huggingface/datasets/issues/3848
1,162,076,902
I_kwDODunzps5FQ-Lm
3,848
NonMatchingChecksumError when checksum is None
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
7
2022-03-08T00:24:12Z
2022-03-15T14:37:26Z
2022-03-15T12:28:23Z
null
I ran into the following error when adding a new dataset: ```bash expected_checksums = {'https://adversarialglue.github.io/dataset/dev.zip': {'checksum': None, 'num_bytes': 40662}} recorded_checksums = {'https://adversarialglue.github.io/dataset/dev.zip': {'checksum': 'efb4cbd3aa4a87bfaffc310ae951981cc0a36c6c71c64...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3848/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3848/timeline
null
completed
null
null
false
[ "Hi @jxmorris12, thanks for reporting.\r\n\r\nThe objective of `verify_checksums` is to check that both checksums are equal. Therefore if one is None and the other is non-None, they are not equal, and the function accordingly raises a NonMatchingChecksumError. That behavior is expected.\r\n\r\nThe question is: how ...
https://api.github.com/repos/huggingface/datasets/issues/3716
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3716/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3716/comments
https://api.github.com/repos/huggingface/datasets/issues/3716/events
https://github.com/huggingface/datasets/issues/3716
1,136,831,092
I_kwDODunzps5Dwqp0
3,716
`FaissIndex` to support multiple GPU and `custom_index`
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
2
2022-02-14T06:21:43Z
2022-03-07T16:28:56Z
2022-03-07T16:28:56Z
null
**Is your feature request related to a problem? Please describe.** Currently, because `device` is of the type `int | None`, to leverage `faiss-gpu`'s multi-gpu support, you need to create a `custom_index`. However, if using a `custom_index` created by e.g. `faiss.index_cpu_to_all_gpus`, then `FaissIndex.save` does not ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3716/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3716/timeline
null
completed
null
null
false
[ "Hi @rentruewang, thansk for reporting and for your PR!!! We should definitely support this. ", "@albertvillanova Great! :)" ]
https://api.github.com/repos/huggingface/datasets/issues/5958
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5958/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5958/comments
https://api.github.com/repos/huggingface/datasets/issues/5958/events
https://github.com/huggingface/datasets/pull/5958
1,757,265,971
PR_kwDODunzps5TA3__
5,958
set dev version
[]
closed
false
null
3
2023-06-14T16:26:34Z
2023-06-14T16:34:55Z
2023-06-14T16:26:51Z
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5958/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5958/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/5958.diff", "html_url": "https://github.com/huggingface/datasets/pull/5958", "merged_at": "2023-06-14T16:26:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/5958.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5958). All of your documentation changes will be reflected on that endpoint.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma...
https://api.github.com/repos/huggingface/datasets/issues/4845
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4845/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4845/comments
https://api.github.com/repos/huggingface/datasets/issues/4845/events
https://github.com/huggingface/datasets/pull/4845
1,337,928,283
PR_kwDODunzps49IOjf
4,845
Mark CI tests as xfail if Hub HTTP error
[]
closed
false
null
1
2022-08-13T10:45:11Z
2022-08-23T04:57:12Z
2022-08-23T04:42:26Z
null
In order to make testing more robust (and avoid merges to master with red tests), we could mark tests as xfailed (instead of failed) when the Hub raises some temporary HTTP errors. This PR: - marks tests as xfailed only if the Hub raises a 500 error for: - test_upstream_hub - makes pytest report the xfailed/xpa...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4845/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4845/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4845.diff", "html_url": "https://github.com/huggingface/datasets/pull/4845", "merged_at": "2022-08-23T04:42:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/4845.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/241
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/241/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/241/comments
https://api.github.com/repos/huggingface/datasets/issues/241/events
https://github.com/huggingface/datasets/pull/241
631,703,079
MDExOlB1bGxSZXF1ZXN0NDI4NTQwMDM0
241
Fix empty cache dir
[]
closed
false
null
2
2020-06-05T15:45:22Z
2020-06-08T08:35:33Z
2020-06-08T08:35:31Z
null
If the cache dir of a dataset is empty, the dataset fails to load and throws a FileNotFounfError. We could end up with empty cache dir because there was a line in the code that created the cache dir without using a temp dir. Using a temp dir is useful as it gets renamed to the real cache dir only if the full process is...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/241/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/241/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/241.diff", "html_url": "https://github.com/huggingface/datasets/pull/241", "merged_at": "2020-06-08T08:35:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/241.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/241...
true
[ "Looks great! Will this change force all cached datasets to be redownloaded? But even if it does, it shoud not be a big problem, I think", "> Looks great! Will this change force all cached datasets to be redownloaded? But even if it does, it shoud not be a big problem, I think\r\n\r\nNo it shouldn't force to redo...
https://api.github.com/repos/huggingface/datasets/issues/681
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/681/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/681/comments
https://api.github.com/repos/huggingface/datasets/issues/681/events
https://github.com/huggingface/datasets/pull/681
710,075,721
MDExOlB1bGxSZXF1ZXN0NDkzOTkwMjEz
681
Adding missing @property (+2 small flake8 fixes).
[]
closed
false
null
0
2020-09-28T08:53:53Z
2020-09-28T10:26:13Z
2020-09-28T10:26:09Z
null
Fixes #678
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/681/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/681/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/681.diff", "html_url": "https://github.com/huggingface/datasets/pull/681", "merged_at": "2020-09-28T10:26:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/681.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/681...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/412
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/412/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/412/comments
https://api.github.com/repos/huggingface/datasets/issues/412/events
https://github.com/huggingface/datasets/issues/412
660,047,139
MDU6SXNzdWU2NjAwNDcxMzk=
412
Unable to load XTREME dataset from disk
[]
closed
false
null
3
2020-07-18T09:55:00Z
2020-07-21T08:15:44Z
2020-07-21T08:15:44Z
null
Hi 🤗 team! ## Description of the problem Following the [docs](https://huggingface.co/nlp/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https://github.com/google-research/xtreme) benchmark. I have manually downloaded the `AmazonPho...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/412/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/412/timeline
null
completed
null
null
false
[ "Hi @lewtun, you have to provide the full path to the downloaded file for example `/home/lewtum/..`", "I was able to repro. Opening a PR to fix that.\r\nThanks for reporting this issue !", "Thanks for the rapid fix @lhoestq!" ]
https://api.github.com/repos/huggingface/datasets/issues/4528
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4528/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4528/comments
https://api.github.com/repos/huggingface/datasets/issues/4528/events
https://github.com/huggingface/datasets/issues/4528
1,276,679,155
I_kwDODunzps5MGJPz
4,528
Memory leak when iterating a Dataset
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
5
2022-06-20T10:03:14Z
2022-09-12T08:51:39Z
2022-09-12T08:51:39Z
null
e## Describe the bug It seems that memory never gets freed after iterating a `Dataset` (using `.map()` or a simple `for` loop) ## Steps to reproduce the bug ```python import gc import logging import time import pyarrow from datasets import load_dataset from tqdm import trange import os, psutil logging.ba...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4528/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4528/timeline
null
completed
null
null
false
[ "Is someone assigned to this issue?", "The same issue is being debugged here: https://github.com/huggingface/datasets/issues/4883\r\n", "Here is a modified repro example that makes it easier to see the leak:\r\n\r\n```\r\n$ cat ds2.py\r\nimport gc, sys\r\nimport time\r\nfrom datasets import load_dataset\r\nimpo...
https://api.github.com/repos/huggingface/datasets/issues/2442
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2442/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2442/comments
https://api.github.com/repos/huggingface/datasets/issues/2442/events
https://github.com/huggingface/datasets/pull/2442
909,677,029
MDExOlB1bGxSZXF1ZXN0NjYwMjE1ODY1
2,442
add english language tags for ~100 datasets
[]
closed
false
null
1
2021-06-02T16:24:56Z
2021-06-04T09:51:40Z
2021-06-04T09:51:39Z
null
As discussed on Slack, I have manually checked for ~100 datasets that they have at least one subset in English. This information was missing so adding into the READMEs. Note that I didn't check all the subsets so it's possible that some of the datasets have subsets in other languages than English...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2442/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2442/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2442.diff", "html_url": "https://github.com/huggingface/datasets/pull/2442", "merged_at": "2021-06-04T09:51:39Z", "patch_url": "https://github.com/huggingface/datasets/pull/2442.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "Fixing the tags of all the datasets is out of scope for this PR so I'm merging even though the CI fails because of the missing tags" ]
https://api.github.com/repos/huggingface/datasets/issues/2505
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2505/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2505/comments
https://api.github.com/repos/huggingface/datasets/issues/2505/events
https://github.com/huggingface/datasets/pull/2505
921,234,797
MDExOlB1bGxSZXF1ZXN0NjcwMjY2NjQy
2,505
Make numpy arrow extractor faster
[]
closed
false
null
5
2021-06-15T10:11:32Z
2021-06-28T09:53:39Z
2021-06-28T09:53:38Z
null
I changed the NumpyArrowExtractor to call directly to_numpy and see if it can lead to speed-ups as discussed in https://github.com/huggingface/datasets/issues/2498 This could make the numpy/torch/tf/jax formatting faster
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2505/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2505/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2505.diff", "html_url": "https://github.com/huggingface/datasets/pull/2505", "merged_at": "2021-06-28T09:53:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/2505.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "Looks like we have a nice speed up in some benchmarks. For example:\r\n- `read_formatted numpy 5000`: 4.584777 sec -> 0.487113 sec\r\n- `read_formatted torch 5000`: 4.565676 sec -> 1.289514 sec", "Can we convert this draft to PR @lhoestq ?", "Ready for review ! cc @vblagoje", "@lhoestq I tried the branch a...
https://api.github.com/repos/huggingface/datasets/issues/4374
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4374/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4374/comments
https://api.github.com/repos/huggingface/datasets/issues/4374/events
https://github.com/huggingface/datasets/issues/4374
1,241,860,535
I_kwDODunzps5KBUm3
4,374
extremely slow processing when using a custom dataset
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "d876e3", "default": true, "descript...
closed
false
null
2
2022-05-19T14:18:05Z
2023-07-25T15:07:17Z
2023-07-25T15:07:16Z
null
## processing a custom dataset loaded as .txt file is extremely slow, compared to a dataset of similar volume from the hub I have a large .txt file of 22 GB which i load into HF dataset `lang_dataset = datasets.load_dataset("text", data_files="hi.txt")` further i use a pre-processing function to clean the d...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4374/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4374/timeline
null
completed
null
null
false
[ "Hi !\r\n\r\nMy guess is that some examples in your dataset are bigger than your RAM, and therefore loading them in RAM to pass them to `remove_non_indic_sentences` takes forever because it might use SWAP memory.\r\n\r\nMaybe several examples in your dataset are grouped together, can you check `len(lang_dataset[\"t...
https://api.github.com/repos/huggingface/datasets/issues/1280
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1280/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1280/comments
https://api.github.com/repos/huggingface/datasets/issues/1280/events
https://github.com/huggingface/datasets/pull/1280
759,151,028
MDExOlB1bGxSZXF1ZXN0NTM0MTk2MDc0
1,280
disaster response messages dataset
[]
closed
false
null
2
2020-12-08T07:27:16Z
2020-12-09T16:21:57Z
2020-12-09T16:21:57Z
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1280/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1280/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1280.diff", "html_url": "https://github.com/huggingface/datasets/pull/1280", "merged_at": "2020-12-09T16:21:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/1280.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "I have added the Readme.md as well, the PR is ready for review. \r\n\r\nThank you ", "Hi @lhoestq I have updated the code and files. Please if you could check once.\r\n\r\nThank you" ]
https://api.github.com/repos/huggingface/datasets/issues/1758
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1758/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1758/comments
https://api.github.com/repos/huggingface/datasets/issues/1758/events
https://github.com/huggingface/datasets/issues/1758
790,626,116
MDU6SXNzdWU3OTA2MjYxMTY=
1,758
dataset.search() (elastic) cannot reliably retrieve search results
[]
closed
false
null
2
2021-01-21T02:26:37Z
2021-01-22T00:25:50Z
2021-01-22T00:25:50Z
null
I am trying to use elastic search to retrieve the indices of items in the dataset in their precise order, given shuffled training indices. The problem I have is that I cannot retrieve reliable results with my data on my first search. I have to run the search **twice** to get the right answer. I am indexing data t...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1758/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1758/timeline
null
completed
null
null
false
[ "Hi !\r\nI tried your code on my side and I was able to workaround this issue by waiting a few seconds before querying the index.\r\nMaybe this is because the index is not updated yet on the ElasticSearch side ?", "Thanks for the feedback! I added a 30 second \"sleep\" and that seemed to work well!" ]
https://api.github.com/repos/huggingface/datasets/issues/922
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/922/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/922/comments
https://api.github.com/repos/huggingface/datasets/issues/922/events
https://github.com/huggingface/datasets/pull/922
753,559,130
MDExOlB1bGxSZXF1ZXN0NTI5NjEzOTA4
922
Add XOR QA Dataset
[]
closed
false
null
4
2020-11-30T15:10:54Z
2020-12-02T03:12:21Z
2020-12-02T03:12:21Z
null
Added XOR Question Answering Dataset. The link to the dataset can be found [here](https://nlp.cs.washington.edu/xorqa/) - [x] Followed the instructions in CONTRIBUTING.md - [x] Ran the tests successfully - [x] Created the dummy data
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/922/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/922/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/922.diff", "html_url": "https://github.com/huggingface/datasets/pull/922", "merged_at": "2020-12-02T03:12:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/922.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/922...
true
[ "Hi @sumanthd17 \r\n\r\nLooks like a good start! You will also need to add a Dataset card, following the instructions given [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md#manually-tag-the-dataset-and-write-the-dataset-card)", "I followed the instructions mentioned there but my datas...
https://api.github.com/repos/huggingface/datasets/issues/1252
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1252/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1252/comments
https://api.github.com/repos/huggingface/datasets/issues/1252/events
https://github.com/huggingface/datasets/pull/1252
758,511,388
MDExOlB1bGxSZXF1ZXN0NTMzNjczMDcx
1,252
Add Naver sentiment movie corpus
[]
closed
false
null
0
2020-12-07T13:33:45Z
2020-12-08T14:32:33Z
2020-12-08T14:21:37Z
null
Supersedes #1168 > This PR adds the [Naver sentiment movie corpus](https://github.com/e9t/nsmc), a dataset containing Korean movie reviews from Naver, the most commonly used search engine in Korea. This dataset is often used to benchmark models on Korean NLP tasks, as seen in [this paper](https://www.aclweb.org/ant...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1252/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1252/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1252.diff", "html_url": "https://github.com/huggingface/datasets/pull/1252", "merged_at": "2020-12-08T14:21:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/1252.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/2560
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2560/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2560/comments
https://api.github.com/repos/huggingface/datasets/issues/2560/events
https://github.com/huggingface/datasets/pull/2560
932,143,634
MDExOlB1bGxSZXF1ZXN0Njc5NTMyODk4
2,560
fix Dataset.map when num_procs > num rows
[]
closed
false
null
3
2021-06-29T02:24:11Z
2021-06-29T15:00:18Z
2021-06-29T14:53:31Z
null
closes #2470 ## Testing notes To run updated tests: ```sh pytest tests/test_arrow_dataset.py -k "BaseDatasetTest and test_map_multiprocessing" -s ``` With Python code (to view warning): ```python from datasets import Dataset dataset = Dataset.from_dict({"x": ["sample"]}) print(len(dataset)) dataset.map...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2560/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2560/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2560.diff", "html_url": "https://github.com/huggingface/datasets/pull/2560", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2560.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2560" }
true
[ "Hi ! Thanks for fixing this :)\r\n\r\nLooks like you have tons of changes due to code formatting.\r\nWe're using `black` for this, with a custom line length. To run our code formatting, you just need to run\r\n```\r\nmake style\r\n```\r\n\r\nThen for the windows error in the CI, I'm looking into it. It's probably ...
https://api.github.com/repos/huggingface/datasets/issues/622
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/622/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/622/comments
https://api.github.com/repos/huggingface/datasets/issues/622/events
https://github.com/huggingface/datasets/issues/622
700,225,826
MDU6SXNzdWU3MDAyMjU4MjY=
622
load_dataset for text files not working
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
41
2020-09-12T12:49:28Z
2020-10-28T11:07:31Z
2020-10-28T11:07:30Z
null
Trying the following snippet, I get different problems on Linux and Windows. ```python dataset = load_dataset("text", data_files="data.txt") # or dataset = load_dataset("text", data_files=["data.txt"]) ``` (ps [This example](https://huggingface.co/docs/datasets/loading_datasets.html#json-files) shows that ...
{ "+1": 3, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/622/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/622/timeline
null
completed
null
null
false
[ "Can you give us more information on your os and pip environments (pip list)?", "@thomwolf Sure. I'll try downgrading to 3.7 now even though Arrow say they support >=3.5.\r\n\r\nLinux (Ubuntu 18.04) - Python 3.8\r\n======================\r\nPackage - Version\r\n---------------------\r\ncertifi 2...
https://api.github.com/repos/huggingface/datasets/issues/4547
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4547/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4547/comments
https://api.github.com/repos/huggingface/datasets/issues/4547/events
https://github.com/huggingface/datasets/pull/4547
1,282,160,517
PR_kwDODunzps46Ot5u
4,547
[CI] Fix some warnings
[]
closed
false
null
4
2022-06-23T10:10:49Z
2022-06-28T14:10:57Z
2022-06-28T13:59:54Z
null
There are some warnings in the CI that are annoying, I tried to remove most of them
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4547/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4547/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4547.diff", "html_url": "https://github.com/huggingface/datasets/pull/4547", "merged_at": "2022-06-28T13:59:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/4547.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "There is a CI failure only related to the missing content of the universal_dependencies dataset card, we can ignore this failure in this PR", "good catch, I thought I resolved them all sorry", "Alright it should be good now" ]
https://api.github.com/repos/huggingface/datasets/issues/1282
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1282/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1282/comments
https://api.github.com/repos/huggingface/datasets/issues/1282/events
https://github.com/huggingface/datasets/pull/1282
759,208,335
MDExOlB1bGxSZXF1ZXN0NTM0MjQ4NzI5
1,282
add thaiqa_squad
[]
closed
false
null
0
2020-12-08T08:14:38Z
2020-12-08T18:36:18Z
2020-12-08T18:36:18Z
null
Example format is a little different from SQuAD since `thaiqa` always have one answer per question so I added a check to convert answers to lists if they are not already one to future-proof additional questions that might have multiple answers. `thaiqa_squad` is an open-domain, extractive question answering dataset ...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/1282/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1282/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1282.diff", "html_url": "https://github.com/huggingface/datasets/pull/1282", "merged_at": "2020-12-08T18:36:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/1282.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/5365
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5365/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5365/comments
https://api.github.com/repos/huggingface/datasets/issues/5365/events
https://github.com/huggingface/datasets/pull/5365
1,498,422,466
PR_kwDODunzps5Fi6ZD
5,365
fix: image array should support other formats than uint8
[]
closed
false
null
4
2022-12-15T13:17:50Z
2023-01-26T18:46:45Z
2023-01-26T18:39:36Z
null
Currently images that are provided as ndarrays, but not in `uint8` format are going to loose data. Namely, for example in a depth image where the data is in float32 format, the type-casting to uint8 will basically make the whole image blank. `PIL.Image.fromarray` [does support mode `F`](https://pillow.readthedocs.io/e...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5365/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5365/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/5365.diff", "html_url": "https://github.com/huggingface/datasets/pull/5365", "merged_at": "2023-01-26T18:39:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/5365.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hi, thanks for working on this! \r\n\r\nI agree that the current type-casting (always cast to `np.uint8` as Tensorflow Datasets does) is a bit too harsh. However, not all dtypes are supported in `Image.fromarray` (e.g. np.int64), so ...
https://api.github.com/repos/huggingface/datasets/issues/4185
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4185/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4185/comments
https://api.github.com/repos/huggingface/datasets/issues/4185/events
https://github.com/huggingface/datasets/issues/4185
1,209,429,743
I_kwDODunzps5IFm7v
4,185
Librispeech documentation, clarification on format
[]
open
false
null
8
2022-04-20T09:35:55Z
2022-04-21T11:00:53Z
null
null
https://github.com/huggingface/datasets/blob/cd3ce34ab1604118351e1978d26402de57188901/datasets/librispeech_asr/librispeech_asr.py#L53 > Note that in order to limit the required storage for preparing this dataset, the audio > is stored in the .flac format and is not converted to a float32 array. To convert, the audi...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4185/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4185/timeline
null
null
null
null
false
[ "(@patrickvonplaten )", "Also cc @lhoestq here", "The documentation in the code is definitely outdated - thanks for letting me know, I'll remove it in https://github.com/huggingface/datasets/pull/4184 .\r\n\r\nYou're exactly right `audio` `array` already decodes the audio file to the correct waveform. This is d...
https://api.github.com/repos/huggingface/datasets/issues/3754
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3754/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3754/comments
https://api.github.com/repos/huggingface/datasets/issues/3754/events
https://github.com/huggingface/datasets/issues/3754
1,142,886,536
I_kwDODunzps5EHxCI
3,754
Overflowing indices in `select`
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
2
2022-02-18T11:30:52Z
2022-02-18T11:38:23Z
2022-02-18T11:38:23Z
null
## Describe the bug The `Dataset.select` function seems to accept indices that are larger than the dataset size and seems to effectively use `index %len(ds)`. ## Steps to reproduce the bug ```python from datasets import Dataset ds = Dataset.from_dict({"test": [1,2,3]}) ds = ds.select(range(5)) print(ds) p...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3754/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3754/timeline
null
completed
null
null
false
[ "Fixed on master (see https://github.com/huggingface/datasets/pull/3719).", "Awesome, I did not find that one! Thanks." ]
https://api.github.com/repos/huggingface/datasets/issues/4045
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4045/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4045/comments
https://api.github.com/repos/huggingface/datasets/issues/4045/events
https://github.com/huggingface/datasets/pull/4045
1,183,661,091
PR_kwDODunzps41KtfV
4,045
Fix CLI dummy data generation
[]
closed
false
null
1
2022-03-28T16:09:15Z
2022-03-31T15:04:12Z
2022-03-31T14:59:06Z
null
PR: - #3868 broke the CLI dummy data generation. Fix #4044.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4045/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4045/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4045.diff", "html_url": "https://github.com/huggingface/datasets/pull/4045", "merged_at": "2022-03-31T14:59:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/4045.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/5130
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5130/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5130/comments
https://api.github.com/repos/huggingface/datasets/issues/5130/events
https://github.com/huggingface/datasets/pull/5130
1,413,435,000
PR_kwDODunzps5BBxXX
5,130
Avoid extra cast in `class_encode_column`
[]
closed
false
null
1
2022-10-18T15:31:24Z
2022-10-19T11:53:02Z
2022-10-19T11:50:46Z
null
Pass the updated features to `map` to avoid the `cast` in `class_encode_column`.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5130/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5130/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/5130.diff", "html_url": "https://github.com/huggingface/datasets/pull/5130", "merged_at": "2022-10-19T11:50:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/5130.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/2027
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2027/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2027/comments
https://api.github.com/repos/huggingface/datasets/issues/2027/events
https://github.com/huggingface/datasets/pull/2027
828,490,444
MDExOlB1bGxSZXF1ZXN0NTkwMjkzNDA1
2,027
Update format columns in Dataset.rename_columns
[]
closed
false
null
0
2021-03-10T23:50:59Z
2021-03-11T14:38:40Z
2021-03-11T14:38:40Z
null
Fixes #2026
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2027/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2027/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2027.diff", "html_url": "https://github.com/huggingface/datasets/pull/2027", "merged_at": "2021-03-11T14:38:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/2027.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/3104
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3104/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3104/comments
https://api.github.com/repos/huggingface/datasets/issues/3104/events
https://github.com/huggingface/datasets/issues/3104
1,029,080,412
I_kwDODunzps49VoVc
3,104
Missing Zenodo 1.13.3 release
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
1
2021-10-18T12:57:18Z
2021-10-22T13:22:25Z
2021-10-22T13:22:24Z
null
After `datasets` 1.13.3 release, this does not appear in Zenodo releases: https://zenodo.org/record/5570305 TODO: - [x] Contact Zenodo support - [x] Check it is fixed
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3104/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3104/timeline
null
completed
null
null
false
[ "Zenodo has fixed on their side the 1.13.3 release: https://zenodo.org/record/5589150" ]
https://api.github.com/repos/huggingface/datasets/issues/455
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/455/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/455/comments
https://api.github.com/repos/huggingface/datasets/issues/455/events
https://github.com/huggingface/datasets/pull/455
668,037,965
MDExOlB1bGxSZXF1ZXN0NDU4NTk4NTUw
455
Add bleurt
[]
closed
false
null
4
2020-07-29T18:08:32Z
2020-07-31T13:56:14Z
2020-07-31T13:56:14Z
null
This PR adds the BLEURT metric to the library. The BLEURT `Metric` downloads a TF checkpoint corresponding to its `config_name` at creation (in the `_info` function). Default is set to `bleurt-base-128`. Note that the default in the original package is `bleurt-tiny-128`, but they throw a warning and recommend usi...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/455/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/455/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/455.diff", "html_url": "https://github.com/huggingface/datasets/pull/455", "merged_at": "2020-07-31T13:56:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/455.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/455...
true
[ "Sorry one nit: Could we use named arguments for the call to BLEURT?\r\n\r\ni.e. \r\n scores = self.scorer.score(references=references, candidates=predictions)\r\n\r\n(i.e. so it is less bug prone)\r\n", "Following up on Ankur's comment---we are going to drop support for\npositional (not named) arguments i...
https://api.github.com/repos/huggingface/datasets/issues/4668
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4668/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4668/comments
https://api.github.com/repos/huggingface/datasets/issues/4668/events
https://github.com/huggingface/datasets/issues/4668
1,299,735,893
I_kwDODunzps5NeGVV
4,668
Dataset Viewer issue for hungnm/multilingual-amazon-review-sentiment-processed
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
closed
false
null
1
2022-07-09T18:04:13Z
2022-07-11T07:47:47Z
2022-07-11T07:47:47Z
null
### Link https://huggingface.co/hungnm/multilingual-amazon-review-sentiment ### Description _No response_ ### Owner Yes
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4668/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4668/timeline
null
completed
null
null
false
[ "It seems like a private dataset. The viewer is currently not supported on the private datasets." ]
https://api.github.com/repos/huggingface/datasets/issues/3622
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3622/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3622/comments
https://api.github.com/repos/huggingface/datasets/issues/3622/events
https://github.com/huggingface/datasets/issues/3622
1,112,831,661
I_kwDODunzps5CVHat
3,622
Extend support for streaming datasets that use os.path.relpath
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
0
2022-01-24T15:58:23Z
2022-02-04T14:03:54Z
2022-02-04T14:03:54Z
null
Extend support for streaming datasets that use `os.path.relpath`. This feature will also be useful to yield the relative path of audio or image files.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3622/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3622/timeline
null
completed
null
null
false
[]
End of preview. Expand in Data Studio

Dataset Card for "github-issues_huggingface-datasets"

Dataset Name: GitHub Issues from Hugging Face Datasets

Description: The "github-issues_huggingface-datasets" dataset is a corpus of GitHub issues extracted from the Hugging Face Datasets repository. It includes valuable information and metadata related to the issues, such as titles, descriptions, labels, states, comments, and whether they are pull requests. The dataset was compiled using the GitHub REST API, which enabled the retrieval of issues and their corresponding comments. Additionally, a new column was added to indicate whether an issue is a pull request.

Dataset Contents: The dataset consists of two splits:

  1. Train Split: Contains 4,863 records, each with features like URL, repository URL, labels URL, comments URL, HTML URL, ID, node ID, issue number, Title, labels, state, locked status, milestone, comments, creation date, update date, closing date, reactions, timeline URL, and more.
  2. Test Split: Comprises 1,216 records with the same features as the train split.

Potential Uses: This dataset is valuable for various purposes, such as:

  • Semantic search: Analyzing and retrieving issues based on semantic similarity.
  • Multilabel classification: Classifying issues into multiple categories based on their labels.
  • Exploratory analysis: Gaining insights into the trends and patterns within GitHub issues.

Limitations and Risks: Users of this dataset should be aware of potential limitations, such as data incompleteness, bias in issue labeling, or outdated information. Additionally, data privacy and ethical considerations should be taken into account when using GitHub issues data.

Access: The dataset is openly accessible to anyone interested in using it for research, analysis, or any other suitable applications. The dataset is publicly available for download and usage.

Note: Certain user-specific features, such as "user", "author_association", "assignee" and "assignees" have been excluded from the dataset to protect individual privacy and mitigate the risk of identifying users or contributors.

More Information needed

Downloads last month
7