url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
3.64B
node_id
stringlengths
18
32
number
int64
1
7.87k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
int64
0
70
created_at
stringdate
2020-04-14 10:18:02
2025-11-18 08:33:04
updated_at
stringdate
2020-04-27 16:04:17
2025-11-18 16:07:04
closed_at
stringlengths
3
25
author_association
stringclasses
4 values
type
float64
active_lock_reason
float64
sub_issues_summary
dict
issue_dependencies_summary
dict
body
stringlengths
0
228k
closed_by
dict
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
float64
state_reason
stringclasses
4 values
draft
float64
0
1
pull_request
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/288
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/288/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/288/comments
https://api.github.com/repos/huggingface/datasets/issues/288/events
https://github.com/huggingface/datasets/issues/288
641,888,610
MDU6SXNzdWU2NDE4ODg2MTA=
288
Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'
{ "avatar_url": "https://avatars.githubusercontent.com/u/14964542?v=4", "events_url": "https://api.github.com/users/wutong8023/events{/privacy}", "followers_url": "https://api.github.com/users/wutong8023/followers", "following_url": "https://api.github.com/users/wutong8023/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
5
2020-06-19 11:01:22+00:00
2020-06-21 09:05:11+00:00
2020-06-21 09:05:11+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
/Users/parasol_tree/anaconda3/lib/python3.6/site-packages/tensorflow/python/framework/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'. _np_qint8 = np.dtype([("qint8", np.int8, 1)]) /Users/...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/288/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/288/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/287
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/287/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/287/comments
https://api.github.com/repos/huggingface/datasets/issues/287/events
https://github.com/huggingface/datasets/pull/287
641,800,227
MDExOlB1bGxSZXF1ZXN0NDM2OTY0NTg0
287
fix squad_v2 metric
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-06-19 08:24:46+00:00
2020-06-19 08:33:43+00:00
2020-06-19 08:33:41+00:00
MEMBER
null
null
null
null
Fix #280 The imports were wrong
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/287/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/287/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/287.diff", "html_url": "https://github.com/huggingface/datasets/pull/287", "merged_at": "2020-06-19T08:33:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/287.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/287...
true
https://api.github.com/repos/huggingface/datasets/issues/286
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/286/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/286/comments
https://api.github.com/repos/huggingface/datasets/issues/286/events
https://github.com/huggingface/datasets/pull/286
641,585,758
MDExOlB1bGxSZXF1ZXN0NDM2NzkzMjI4
286
Add ANLI dataset.
{ "avatar_url": "https://avatars.githubusercontent.com/u/11016329?v=4", "events_url": "https://api.github.com/users/easonnie/events{/privacy}", "followers_url": "https://api.github.com/users/easonnie/followers", "following_url": "https://api.github.com/users/easonnie/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
1
2020-06-18 22:27:30+00:00
2020-06-22 12:23:27+00:00
2020-06-22 12:23:27+00:00
CONTRIBUTOR
null
null
null
null
I completed all the steps in https://github.com/huggingface/nlp/blob/master/CONTRIBUTING.md#how-to-add-a-dataset and push the code for ANLI. Please let me know if there are any errors.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/286/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/286/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/286.diff", "html_url": "https://github.com/huggingface/datasets/pull/286", "merged_at": "2020-06-22T12:23:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/286.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/286...
true
https://api.github.com/repos/huggingface/datasets/issues/285
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/285/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/285/comments
https://api.github.com/repos/huggingface/datasets/issues/285/events
https://github.com/huggingface/datasets/pull/285
641,360,702
MDExOlB1bGxSZXF1ZXN0NDM2NjAyMjk4
285
Consistent formatting of citations
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
1
2020-06-18 16:25:23+00:00
2020-06-22 08:09:25+00:00
2020-06-22 08:09:24+00:00
CONTRIBUTOR
null
null
null
null
#283
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/285/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/285/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/285.diff", "html_url": "https://github.com/huggingface/datasets/pull/285", "merged_at": "2020-06-22T08:09:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/285.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/285...
true
https://api.github.com/repos/huggingface/datasets/issues/284
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/284/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/284/comments
https://api.github.com/repos/huggingface/datasets/issues/284/events
https://github.com/huggingface/datasets/pull/284
641,337,217
MDExOlB1bGxSZXF1ZXN0NDM2NTgxODQ2
284
Fix manual download instructions
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
5
2020-06-18 15:59:57+00:00
2020-06-19 08:24:21+00:00
2020-06-19 08:24:19+00:00
CONTRIBUTOR
null
null
null
null
This PR replaces the static `DatasetBulider` variable `MANUAL_DOWNLOAD_INSTRUCTIONS` by a property function `manual_download_instructions()`. Some datasets like XTREME and all WMT need the manual data dir only for a small fraction of the possible configs. After some brainstorming with @mariamabarham and @lhoestq...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/284/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/284/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/284.diff", "html_url": "https://github.com/huggingface/datasets/pull/284", "merged_at": "2020-06-19T08:24:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/284.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/284...
true
https://api.github.com/repos/huggingface/datasets/issues/283
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/283/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/283/comments
https://api.github.com/repos/huggingface/datasets/issues/283/events
https://github.com/huggingface/datasets/issues/283
641,270,439
MDU6SXNzdWU2NDEyNzA0Mzk=
283
Consistent formatting of citations
{ "avatar_url": "https://avatars.githubusercontent.com/u/35882?v=4", "events_url": "https://api.github.com/users/srush/events{/privacy}", "followers_url": "https://api.github.com/users/srush/followers", "following_url": "https://api.github.com/users/srush/following{/other_user}", "gists_url": "https://api.git...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_...
null
0
2020-06-18 14:48:45+00:00
2020-06-22 17:30:46+00:00
2020-06-22 17:30:46+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
The citations are all of a different format, some have "```" and have text inside, others are proper bibtex. Can we make it so that they all are proper citations, i.e. parse by the bibtex spec: https://bibtexparser.readthedocs.io/en/master/
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/283/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/283/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/282
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/282/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/282/comments
https://api.github.com/repos/huggingface/datasets/issues/282/events
https://github.com/huggingface/datasets/pull/282
641,217,759
MDExOlB1bGxSZXF1ZXN0NDM2NDgxNzMy
282
Update dataset_info from gcs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-06-18 13:41:15+00:00
2020-06-18 16:24:52+00:00
2020-06-18 16:24:51+00:00
MEMBER
null
null
null
null
Some datasets are hosted on gcs (wikipedia for example). In this PR I make sure that, when a user loads such datasets, the file_instructions are built using the dataset_info.json from gcs and not from the info extracted from the local `dataset_infos.json` (the one that contain the info for each config). Indeed local fi...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/282/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/282/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/282.diff", "html_url": "https://github.com/huggingface/datasets/pull/282", "merged_at": "2020-06-18T16:24:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/282.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/282...
true
https://api.github.com/repos/huggingface/datasets/issues/281
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/281/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/281/comments
https://api.github.com/repos/huggingface/datasets/issues/281/events
https://github.com/huggingface/datasets/issues/281
641,067,856
MDU6SXNzdWU2NDEwNjc4NTY=
281
Private/sensitive data
{ "avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4", "events_url": "https://api.github.com/users/MFreidank/events{/privacy}", "followers_url": "https://api.github.com/users/MFreidank/followers", "following_url": "https://api.github.com/users/MFreidank/following{/other_user}", "gists_url": "h...
[]
closed
false
null
[]
null
3
2020-06-18 09:47:27+00:00
2020-06-20 13:15:12+00:00
2020-06-20 13:15:12+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi all, Thanks for this fantastic library, it makes it very easy to do prototyping for NLP projects interchangeably between TF/Pytorch. Unfortunately, there is data that cannot easily be shared publicly as it may contain sensitive information. Is there support/a plan to support such data with NLP, e.g. by readin...
{ "avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4", "events_url": "https://api.github.com/users/MFreidank/events{/privacy}", "followers_url": "https://api.github.com/users/MFreidank/followers", "following_url": "https://api.github.com/users/MFreidank/following{/other_user}", "gists_url": "h...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/281/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/281/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/280
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/280/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/280/comments
https://api.github.com/repos/huggingface/datasets/issues/280/events
https://github.com/huggingface/datasets/issues/280
640,677,615
MDU6SXNzdWU2NDA2Nzc2MTU=
280
Error with SquadV2 Metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/32203792?v=4", "events_url": "https://api.github.com/users/avinregmi/events{/privacy}", "followers_url": "https://api.github.com/users/avinregmi/followers", "following_url": "https://api.github.com/users/avinregmi/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
0
2020-06-17 19:10:54+00:00
2020-06-19 08:33:41+00:00
2020-06-19 08:33:41+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I can't seem to import squad v2 metrics. **squad_metric = nlp.load_metric('squad_v2')** **This throws me an error.:** ``` ImportError Traceback (most recent call last) <ipython-input-8-170b6a170555> in <module> ----> 1 squad_metric = nlp.load_metric('squad_v2') ~/env/lib6...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/280/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/280/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/279
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/279/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/279/comments
https://api.github.com/repos/huggingface/datasets/issues/279/events
https://github.com/huggingface/datasets/issues/279
640,611,692
MDU6SXNzdWU2NDA2MTE2OTI=
279
Dataset Preprocessing Cache with .map() function not working as expected
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
5
2020-06-17 17:17:21+00:00
2021-07-06 21:43:28+00:00
2021-04-18 23:43:49+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. Is there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/279/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/279/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/278
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/278/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/278/comments
https://api.github.com/repos/huggingface/datasets/issues/278/events
https://github.com/huggingface/datasets/issues/278
640,518,917
MDU6SXNzdWU2NDA1MTg5MTc=
278
MemoryError when loading German Wikipedia
{ "avatar_url": "https://avatars.githubusercontent.com/u/4698028?v=4", "events_url": "https://api.github.com/users/gregburman/events{/privacy}", "followers_url": "https://api.github.com/users/gregburman/followers", "following_url": "https://api.github.com/users/gregburman/following{/other_user}", "gists_url":...
[]
closed
false
null
[]
null
7
2020-06-17 15:06:21+00:00
2020-06-19 12:53:02+00:00
2020-06-19 12:53:02+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :) I'm trying to download the German Wikipedia dataset as follows: ``` wiki = nlp.load_dataset("wikip...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/278/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/278/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/277
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/277/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/277/comments
https://api.github.com/repos/huggingface/datasets/issues/277/events
https://github.com/huggingface/datasets/issues/277
640,163,053
MDU6SXNzdWU2NDAxNjMwNTM=
277
Empty samples in glue/qqp
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
null
[]
null
2
2020-06-17 05:54:52+00:00
2020-06-21 00:21:45+00:00
2020-06-21 00:21:45+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
``` qqp = nlp.load_dataset('glue', 'qqp') print(qqp['train'][310121]) print(qqp['train'][362225]) ``` ``` {'question1': 'How can I create an Android app?', 'question2': '', 'label': 0, 'idx': 310137} {'question1': 'How can I develop android app?', 'question2': '', 'label': 0, 'idx': 362246} ``` Notice that que...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/277/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/277/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/276
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/276/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/276/comments
https://api.github.com/repos/huggingface/datasets/issues/276/events
https://github.com/huggingface/datasets/pull/276
639,490,858
MDExOlB1bGxSZXF1ZXN0NDM1MDY5Nzg5
276
Fix metric compute (original_instructions missing)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
2
2020-06-16 08:52:01+00:00
2020-06-18 07:41:45+00:00
2020-06-18 07:41:44+00:00
MEMBER
null
null
null
null
When loading arrow data we added in cc8d250 a way to specify the instructions that were used to store them with the loaded dataset. However metrics load data the same way but don't need instructions (we use one single file). In this PR I just make `original_instructions` optional when reading files to load a `Datas...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/276/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/276/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/276.diff", "html_url": "https://github.com/huggingface/datasets/pull/276", "merged_at": "2020-06-18T07:41:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/276.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/276...
true
https://api.github.com/repos/huggingface/datasets/issues/275
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/275/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/275/comments
https://api.github.com/repos/huggingface/datasets/issues/275/events
https://github.com/huggingface/datasets/issues/275
639,439,052
MDU6SXNzdWU2Mzk0MzkwNTI=
275
NonMatchingChecksumError when loading pubmed dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/48441753?v=4", "events_url": "https://api.github.com/users/DavideStenner/events{/privacy}", "followers_url": "https://api.github.com/users/DavideStenner/followers", "following_url": "https://api.github.com/users/DavideStenner/following{/other_user}", "g...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
1
2020-06-16 07:31:51+00:00
2020-06-19 07:37:07+00:00
2020-06-19 07:37:07+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I get this error when i run `nlp.load_dataset('scientific_papers', 'pubmed', split = 'train[:50%]')`. The error is: ``` --------------------------------------------------------------------------- NonMatchingChecksumError Traceback (most recent call last) <ipython-input-2-7742dea167d0> in <module...
{ "avatar_url": "https://avatars.githubusercontent.com/u/48441753?v=4", "events_url": "https://api.github.com/users/DavideStenner/events{/privacy}", "followers_url": "https://api.github.com/users/DavideStenner/followers", "following_url": "https://api.github.com/users/DavideStenner/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/275/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/275/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/274
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/274/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/274/comments
https://api.github.com/repos/huggingface/datasets/issues/274/events
https://github.com/huggingface/datasets/issues/274
639,156,625
MDU6SXNzdWU2MzkxNTY2MjU=
274
PG-19
{ "avatar_url": "https://avatars.githubusercontent.com/u/108653?v=4", "events_url": "https://api.github.com/users/lucidrains/events{/privacy}", "followers_url": "https://api.github.com/users/lucidrains/followers", "following_url": "https://api.github.com/users/lucidrains/following{/other_user}", "gists_url": ...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
[]
null
4
2020-06-15 21:02:26+00:00
2020-07-06 15:35:02+00:00
2020-07-06 15:35:02+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, and thanks for all your open-sourced work, as always! I was wondering if you would be open to adding PG-19 to your collection of datasets. https://github.com/deepmind/pg19 It is often used for benchmarking long-range language modeling.
{ "avatar_url": "https://avatars.githubusercontent.com/u/108653?v=4", "events_url": "https://api.github.com/users/lucidrains/events{/privacy}", "followers_url": "https://api.github.com/users/lucidrains/followers", "following_url": "https://api.github.com/users/lucidrains/following{/other_user}", "gists_url": ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/274/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/274/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/273
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/273/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/273/comments
https://api.github.com/repos/huggingface/datasets/issues/273/events
https://github.com/huggingface/datasets/pull/273
638,968,054
MDExOlB1bGxSZXF1ZXN0NDM0NjM0MzU4
273
update cos_e to add cos_e v1.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-06-15 16:03:22+00:00
2020-06-16 08:25:54+00:00
2020-06-16 08:25:52+00:00
CONTRIBUTOR
null
null
null
null
This PR updates the cos_e dataset to add v1.0 as requested here #163 @nazneenrajani
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/273/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/273/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/273.diff", "html_url": "https://github.com/huggingface/datasets/pull/273", "merged_at": "2020-06-16T08:25:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/273.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/273...
true
https://api.github.com/repos/huggingface/datasets/issues/272
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/272/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/272/comments
https://api.github.com/repos/huggingface/datasets/issues/272/events
https://github.com/huggingface/datasets/pull/272
638,307,313
MDExOlB1bGxSZXF1ZXN0NDM0MTExOTQ3
272
asd
{ "avatar_url": "https://avatars.githubusercontent.com/u/66900970?v=4", "events_url": "https://api.github.com/users/sn696/events{/privacy}", "followers_url": "https://api.github.com/users/sn696/followers", "following_url": "https://api.github.com/users/sn696/following{/other_user}", "gists_url": "https://api....
[]
closed
false
null
[]
null
0
2020-06-14 08:20:38+00:00
2020-06-14 09:16:41+00:00
2020-06-14 09:16:41+00:00
NONE
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/272/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/272/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/272.diff", "html_url": "https://github.com/huggingface/datasets/pull/272", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/272.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/272" }
true
https://api.github.com/repos/huggingface/datasets/issues/271
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/271/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/271/comments
https://api.github.com/repos/huggingface/datasets/issues/271/events
https://github.com/huggingface/datasets/pull/271
638,135,754
MDExOlB1bGxSZXF1ZXN0NDMzOTg3NDkw
271
Fix allociné dataset configuration
{ "avatar_url": "https://avatars.githubusercontent.com/u/37028092?v=4", "events_url": "https://api.github.com/users/TheophileBlard/events{/privacy}", "followers_url": "https://api.github.com/users/TheophileBlard/followers", "following_url": "https://api.github.com/users/TheophileBlard/following{/other_user}", ...
[]
closed
false
null
[]
null
6
2020-06-13 10:12:10+00:00
2020-06-18 07:41:21+00:00
2020-06-18 07:41:20+00:00
CONTRIBUTOR
null
null
null
null
This is a patch for #244. According to the [live nlp viewer](url), the Allociné dataset must be loaded with : ```python dataset = load_dataset('allocine', 'allocine') ``` This is redundant, as there is only one "dataset configuration", and should only be: ```python dataset = load_dataset('allocine') ``` This ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/271/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/271/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/271.diff", "html_url": "https://github.com/huggingface/datasets/pull/271", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/271.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/271" }
true
https://api.github.com/repos/huggingface/datasets/issues/270
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/270/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/270/comments
https://api.github.com/repos/huggingface/datasets/issues/270/events
https://github.com/huggingface/datasets/issues/270
638,121,617
MDU6SXNzdWU2MzgxMjE2MTc=
270
c4 dataset is not viewable in nlpviewer demo
{ "avatar_url": "https://avatars.githubusercontent.com/u/6441313?v=4", "events_url": "https://api.github.com/users/rajarsheem/events{/privacy}", "followers_url": "https://api.github.com/users/rajarsheem/followers", "following_url": "https://api.github.com/users/rajarsheem/following{/other_user}", "gists_url":...
[ { "color": "94203D", "default": false, "description": "", "id": 2107841032, "name": "nlp-viewer", "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer" } ]
closed
false
null
[]
null
1
2020-06-13 08:26:16+00:00
2020-10-27 15:35:29+00:00
2020-10-27 15:35:13+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I get the following error when I try to view the c4 dataset in [nlpviewer](https://huggingface.co/nlp/viewer/) ```python ModuleNotFoundError: No module named 'langdetect' Traceback: File "/home/sasha/.local/lib/python3.7/site-packages/streamlit/ScriptRunner.py", line 322, in _run_script exec(code, module.__d...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/270/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/270/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/269
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/269/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/269/comments
https://api.github.com/repos/huggingface/datasets/issues/269/events
https://github.com/huggingface/datasets/issues/269
638,106,774
MDU6SXNzdWU2MzgxMDY3NzQ=
269
Error in metric.compute: missing `original_instructions` argument
{ "avatar_url": "https://avatars.githubusercontent.com/u/1668462?v=4", "events_url": "https://api.github.com/users/zphang/events{/privacy}", "followers_url": "https://api.github.com/users/zphang/followers", "following_url": "https://api.github.com/users/zphang/following{/other_user}", "gists_url": "https://ap...
[ { "color": "25b21e", "default": false, "description": "A bug in a metric script", "id": 2067393914, "name": "metric bug", "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
0
2020-06-13 06:26:54+00:00
2020-06-18 07:41:44+00:00
2020-06-18 07:41:44+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I'm running into an error using metrics for computation in the latest master as well as version 0.2.1. Here is a minimal example: ```python import nlp rte_metric = nlp.load_metric('glue', name="rte") rte_metric.compute( [0, 0, 1, 1], [0, 1, 0, 1], ) ``` ``` 181 # Read the predictio...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/269/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/269/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/268
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/268/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/268/comments
https://api.github.com/repos/huggingface/datasets/issues/268/events
https://github.com/huggingface/datasets/pull/268
637,848,056
MDExOlB1bGxSZXF1ZXN0NDMzNzU5NzQ1
268
add Rotten Tomatoes Movie Review sentences sentiment dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
1
2020-06-12 15:53:59+00:00
2020-06-18 07:46:24+00:00
2020-06-18 07:46:23+00:00
CONTRIBUTOR
null
null
null
null
Sentence-level movie reviews v1.0 from here: http://www.cs.cornell.edu/people/pabo/movie-review-data/
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/268/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/268/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/268.diff", "html_url": "https://github.com/huggingface/datasets/pull/268", "merged_at": "2020-06-18T07:46:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/268.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/268...
true
https://api.github.com/repos/huggingface/datasets/issues/267
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/267/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/267/comments
https://api.github.com/repos/huggingface/datasets/issues/267/events
https://github.com/huggingface/datasets/issues/267
637,415,545
MDU6SXNzdWU2Mzc0MTU1NDU=
267
How can I load/find WMT en-romanian?
{ "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "events_url": "https://api.github.com/users/sshleifer/events{/privacy}", "followers_url": "https://api.github.com/users/sshleifer/followers", "following_url": "https://api.github.com/users/sshleifer/following{/other_user}", "gists_url": "h...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/followin...
null
1
2020-06-12 01:09:37+00:00
2020-06-19 08:24:19+00:00
2020-06-19 08:24:19+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I believe it is from `wmt16` When I run ```python wmt = nlp.load_dataset('wmt16') ``` I get: ```python AssertionError: The dataset wmt16 with config cs-en requires manual data. Please follow the manual download instructions: Some of the wmt configs here, require a manual download. Please look into wm...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/267/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/267/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/266
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/266/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/266/comments
https://api.github.com/repos/huggingface/datasets/issues/266/events
https://github.com/huggingface/datasets/pull/266
637,156,392
MDExOlB1bGxSZXF1ZXN0NDMzMTk1NDgw
266
Add sort, shuffle, test_train_split and select methods
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
4
2020-06-11 16:22:20+00:00
2020-06-18 16:23:25+00:00
2020-06-18 16:23:24+00:00
MEMBER
null
null
null
null
Add a bunch of methods to reorder/split/select rows in a dataset: - `dataset.select(indices)`: Create a new dataset with rows selected following the list/array of indices (which can have a different size than the dataset and contain duplicated indices, the only constrain is that all the integers in the list must be sm...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/266/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/266/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/266.diff", "html_url": "https://github.com/huggingface/datasets/pull/266", "merged_at": "2020-06-18T16:23:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/266.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/266...
true
https://api.github.com/repos/huggingface/datasets/issues/265
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/265/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/265/comments
https://api.github.com/repos/huggingface/datasets/issues/265/events
https://github.com/huggingface/datasets/pull/265
637,139,220
MDExOlB1bGxSZXF1ZXN0NDMzMTgxNDMz
265
Add pyarrow warning colab
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-06-11 15:57:51+00:00
2020-08-02 18:14:36+00:00
2020-06-12 08:14:16+00:00
MEMBER
null
null
null
null
When a user installs `nlp` on google colab, then google colab doesn't update pyarrow, and the runtime needs to be restarted to use the updated version of pyarrow. This is an issue because `nlp` requires the updated version to work correctly. In this PR I added en error that is shown to the user in google colab if...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/265/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/265/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/265.diff", "html_url": "https://github.com/huggingface/datasets/pull/265", "merged_at": "2020-06-12T08:14:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/265.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/265...
true
https://api.github.com/repos/huggingface/datasets/issues/264
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/264/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/264/comments
https://api.github.com/repos/huggingface/datasets/issues/264/events
https://github.com/huggingface/datasets/pull/264
637,106,170
MDExOlB1bGxSZXF1ZXN0NDMzMTU0ODQ4
264
Fix small issues creating dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-06-11 15:20:16+00:00
2020-06-12 08:15:57+00:00
2020-06-12 08:15:56+00:00
MEMBER
null
null
null
null
Fix many small issues mentioned in #249: - don't force to install apache beam for commands - fix None cache dir when using `dl_manager.download_custom` - added new extras in `setup.py` named `dev` that contains tests and quality dependencies - mock dataset sizes when running tests with dummy data - add a note abou...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/264/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/264/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/264.diff", "html_url": "https://github.com/huggingface/datasets/pull/264", "merged_at": "2020-06-12T08:15:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/264.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/264...
true
https://api.github.com/repos/huggingface/datasets/issues/263
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/263/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/263/comments
https://api.github.com/repos/huggingface/datasets/issues/263/events
https://github.com/huggingface/datasets/issues/263
637,028,015
MDU6SXNzdWU2MzcwMjgwMTU=
263
[Feature request] Support for external modality for language datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/1479733?v=4", "events_url": "https://api.github.com/users/aleSuglia/events{/privacy}", "followers_url": "https://api.github.com/users/aleSuglia/followers", "following_url": "https://api.github.com/users/aleSuglia/following{/other_user}", "gists_url": "h...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "c5def5", "default": fals...
closed
false
null
[]
null
5
2020-06-11 13:42:18+00:00
2022-02-10 13:26:35+00:00
2022-02-10 13:26:35+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
# Background In recent years many researchers have advocated that learning meanings from text-based only datasets is just like asking a human to "learn to speak by listening to the radio" [[E. Bender and A. Koller,2020](https://openreview.net/forum?id=GKTvAcb12b), [Y. Bisk et. al, 2020](https://arxiv.org/abs/2004.10...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
{ "+1": 17, "-1": 0, "confused": 0, "eyes": 4, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 22, "url": "https://api.github.com/repos/huggingface/datasets/issues/263/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/263/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/262
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/262/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/262/comments
https://api.github.com/repos/huggingface/datasets/issues/262/events
https://github.com/huggingface/datasets/pull/262
636,702,849
MDExOlB1bGxSZXF1ZXN0NDMyODI3Mzcz
262
Add new dataset ANLI Round 1
{ "avatar_url": "https://avatars.githubusercontent.com/u/11016329?v=4", "events_url": "https://api.github.com/users/easonnie/events{/privacy}", "followers_url": "https://api.github.com/users/easonnie/followers", "following_url": "https://api.github.com/users/easonnie/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
1
2020-06-11 04:14:57+00:00
2020-06-12 22:03:03+00:00
2020-06-12 22:03:03+00:00
CONTRIBUTOR
null
null
null
null
Adding new dataset [ANLI](https://github.com/facebookresearch/anli/). I'm not familiar with how to add new dataset. Let me know if there is any issue. I only include round 1 data here. There will be round 2, round 3 and more in the future with potentially different format. I think it will be better to separate them.
{ "avatar_url": "https://avatars.githubusercontent.com/u/11016329?v=4", "events_url": "https://api.github.com/users/easonnie/events{/privacy}", "followers_url": "https://api.github.com/users/easonnie/followers", "following_url": "https://api.github.com/users/easonnie/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/262/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/262/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/262.diff", "html_url": "https://github.com/huggingface/datasets/pull/262", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/262.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/262" }
true
https://api.github.com/repos/huggingface/datasets/issues/261
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/261/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/261/comments
https://api.github.com/repos/huggingface/datasets/issues/261/events
https://github.com/huggingface/datasets/issues/261
636,372,380
MDU6SXNzdWU2MzYzNzIzODA=
261
Downloading dataset error with pyarrow.lib.RecordBatch
{ "avatar_url": "https://avatars.githubusercontent.com/u/5248968?v=4", "events_url": "https://api.github.com/users/cuent/events{/privacy}", "followers_url": "https://api.github.com/users/cuent/followers", "following_url": "https://api.github.com/users/cuent/following{/other_user}", "gists_url": "https://api.g...
[]
closed
false
null
[]
null
2
2020-06-10 16:04:19+00:00
2020-06-11 14:35:12+00:00
2020-06-11 14:35:12+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I am trying to download `sentiment140` and I have the following error ``` /usr/local/lib/python3.6/dist-packages/nlp/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs) 518 download_mode=...
{ "avatar_url": "https://avatars.githubusercontent.com/u/5248968?v=4", "events_url": "https://api.github.com/users/cuent/events{/privacy}", "followers_url": "https://api.github.com/users/cuent/followers", "following_url": "https://api.github.com/users/cuent/following{/other_user}", "gists_url": "https://api.g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/261/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/261/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/260
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/260/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/260/comments
https://api.github.com/repos/huggingface/datasets/issues/260/events
https://github.com/huggingface/datasets/pull/260
636,261,118
MDExOlB1bGxSZXF1ZXN0NDMyNDY3NDM5
260
Consistency fixes
{ "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "events_url": "https://api.github.com/users/julien-c/events{/privacy}", "followers_url": "https://api.github.com/users/julien-c/followers", "following_url": "https://api.github.com/users/julien-c/following{/other_user}", "gists_url": "https...
[]
closed
false
null
[]
null
0
2020-06-10 13:44:42+00:00
2020-06-11 10:34:37+00:00
2020-06-11 10:34:36+00:00
MEMBER
null
null
null
null
A few bugs I've found while hacking
{ "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "events_url": "https://api.github.com/users/julien-c/events{/privacy}", "followers_url": "https://api.github.com/users/julien-c/followers", "following_url": "https://api.github.com/users/julien-c/following{/other_user}", "gists_url": "https...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/260/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/260/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/260.diff", "html_url": "https://github.com/huggingface/datasets/pull/260", "merged_at": "2020-06-11T10:34:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/260.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/260...
true
https://api.github.com/repos/huggingface/datasets/issues/259
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/259/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/259/comments
https://api.github.com/repos/huggingface/datasets/issues/259/events
https://github.com/huggingface/datasets/issues/259
636,239,529
MDU6SXNzdWU2MzYyMzk1Mjk=
259
documentation missing how to split a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/2873355?v=4", "events_url": "https://api.github.com/users/fotisj/events{/privacy}", "followers_url": "https://api.github.com/users/fotisj/followers", "following_url": "https://api.github.com/users/fotisj/following{/other_user}", "gists_url": "https://ap...
[]
closed
false
null
[]
null
7
2020-06-10 13:18:13+00:00
2023-03-14 13:56:07+00:00
2020-06-18 22:20:24+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I am trying to understand how to split a dataset ( as arrow_dataset). I know I can do something like this to access a split which is already in the original dataset : `ds_test = nlp.load_dataset('imdb, split='test') ` But how can I split ds_test into a test and a validation set (without reading the data into m...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/259/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/259/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/258
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/258/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/258/comments
https://api.github.com/repos/huggingface/datasets/issues/258/events
https://github.com/huggingface/datasets/issues/258
635,859,525
MDU6SXNzdWU2MzU4NTk1MjU=
258
Why is dataset after tokenization far more larger than the orginal one ?
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
null
[]
null
4
2020-06-10 01:27:07+00:00
2020-06-10 12:46:34+00:00
2020-06-10 12:46:34+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I tokenize wiki dataset by `map` and cache the results. ``` def tokenize_tfm(example): example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text'])) return example wiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train'] wiki.map(token...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/258/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/258/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/257
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/257/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/257/comments
https://api.github.com/repos/huggingface/datasets/issues/257/events
https://github.com/huggingface/datasets/issues/257
635,620,979
MDU6SXNzdWU2MzU2MjA5Nzk=
257
Tokenizer pickling issue fix not landed in `nlp` yet?
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
2
2020-06-09 17:12:34+00:00
2020-06-10 21:45:32+00:00
2020-06-09 17:26:53+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Unless I recreate an arrow_dataset from my loaded nlp dataset myself (which I think does not use the cache by default), I get the following error when applying the map function: ``` dataset = nlp.load_dataset('cos_e') tokenizer = GPT2TokenizerFast.from_pretrained('gpt2', cache_dir=cache_dir) for split in datase...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/257/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/257/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/256
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/256/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/256/comments
https://api.github.com/repos/huggingface/datasets/issues/256/events
https://github.com/huggingface/datasets/issues/256
635,596,295
MDU6SXNzdWU2MzU1OTYyOTU=
256
[Feature request] Add a feature to dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
5
2020-06-09 16:38:12+00:00
2020-06-09 16:51:42+00:00
2020-06-09 16:51:42+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Is there a straightforward way to add a field to the arrow_dataset, prior to performing map?
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/256/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/256/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/255
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/255/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/255/comments
https://api.github.com/repos/huggingface/datasets/issues/255/events
https://github.com/huggingface/datasets/pull/255
635,300,822
MDExOlB1bGxSZXF1ZXN0NDMxNjg3MDM0
255
Add dataset/piaf
{ "avatar_url": "https://avatars.githubusercontent.com/u/36986299?v=4", "events_url": "https://api.github.com/users/RachelKer/events{/privacy}", "followers_url": "https://api.github.com/users/RachelKer/followers", "following_url": "https://api.github.com/users/RachelKer/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
1
2020-06-09 10:16:01+00:00
2020-06-12 08:31:27+00:00
2020-06-12 08:31:27+00:00
CONTRIBUTOR
null
null
null
null
Small SQuAD-like French QA dataset [PIAF](https://www.aclweb.org/anthology/2020.lrec-1.673.pdf)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/255/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/255/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/255.diff", "html_url": "https://github.com/huggingface/datasets/pull/255", "merged_at": "2020-06-12T08:31:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/255.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/255...
true
https://api.github.com/repos/huggingface/datasets/issues/254
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/254/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/254/comments
https://api.github.com/repos/huggingface/datasets/issues/254/events
https://github.com/huggingface/datasets/issues/254
635,057,568
MDU6SXNzdWU2MzUwNTc1Njg=
254
[Feature request] Be able to remove a specific sample of the dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
1
2020-06-09 02:22:13+00:00
2020-06-09 08:41:38+00:00
2020-06-09 08:41:38+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
As mentioned in #117, it's currently not possible to remove a sample of the dataset. But it is a important use case : After applying some preprocessing, some samples might be empty for example. We should be able to remove these samples from the dataset, or at least mark them as `removed` so when iterating the datase...
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/254/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/254/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/253
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/253/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/253/comments
https://api.github.com/repos/huggingface/datasets/issues/253/events
https://github.com/huggingface/datasets/pull/253
634,791,939
MDExOlB1bGxSZXF1ZXN0NDMxMjgwOTYz
253
add flue dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
10
2020-06-08 17:11:09+00:00
2023-09-24 09:46:03+00:00
2020-07-16 07:50:59+00:00
CONTRIBUTOR
null
null
null
null
This PR add the Flue dataset as requested in this issue #223 . @lbourdois made a detailed description in that issue.
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/253/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/253/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/253.diff", "html_url": "https://github.com/huggingface/datasets/pull/253", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/253.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/253" }
true
https://api.github.com/repos/huggingface/datasets/issues/252
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/252/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/252/comments
https://api.github.com/repos/huggingface/datasets/issues/252/events
https://github.com/huggingface/datasets/issues/252
634,563,239
MDU6SXNzdWU2MzQ1NjMyMzk=
252
NonMatchingSplitsSizesError error when reading the IMDB dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/17463361?v=4", "events_url": "https://api.github.com/users/antmarakis/events{/privacy}", "followers_url": "https://api.github.com/users/antmarakis/followers", "following_url": "https://api.github.com/users/antmarakis/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
4
2020-06-08 12:26:24+00:00
2021-08-27 15:20:58+00:00
2020-06-08 14:01:26+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi! I am trying to load the `imdb` dataset with this line: `dataset = nlp.load_dataset('imdb', data_dir='/A/PATH', cache_dir='/A/PATH')` but I am getting the following error: ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/mounts/Users/cisintern/antmarakis/anaconda3/...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17463361?v=4", "events_url": "https://api.github.com/users/antmarakis/events{/privacy}", "followers_url": "https://api.github.com/users/antmarakis/followers", "following_url": "https://api.github.com/users/antmarakis/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/252/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/252/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/251
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/251/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/251/comments
https://api.github.com/repos/huggingface/datasets/issues/251/events
https://github.com/huggingface/datasets/pull/251
634,544,977
MDExOlB1bGxSZXF1ZXN0NDMxMDgwMDkw
251
Better access to all dataset information
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
0
2020-06-08 11:56:50+00:00
2020-06-12 08:13:00+00:00
2020-06-12 08:12:58+00:00
MEMBER
null
null
null
null
Moves all the dataset info down one level from `dataset.info.XXX` to `dataset.XXX` This way it's easier to access `dataset.feature['label']` for instance Also, add the original split instructions used to create the dataset in `dataset.split` Ex: ``` from nlp import load_dataset stsb = load_dataset('glue', name=...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/251/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/251/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/251.diff", "html_url": "https://github.com/huggingface/datasets/pull/251", "merged_at": "2020-06-12T08:12:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/251.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/251...
true
https://api.github.com/repos/huggingface/datasets/issues/250
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/250/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/250/comments
https://api.github.com/repos/huggingface/datasets/issues/250/events
https://github.com/huggingface/datasets/pull/250
634,416,751
MDExOlB1bGxSZXF1ZXN0NDMwOTcyMzg4
250
Remove checksum download in c4
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
1
2020-06-08 09:13:00+00:00
2020-08-25 07:04:56+00:00
2020-06-08 09:16:59+00:00
MEMBER
null
null
null
null
There was a line from the original tfds script that was still there and causing issues when loading the c4 script. This one should fix #233 and allow anyone to load the c4 script to generate the dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/250/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/250/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/250.diff", "html_url": "https://github.com/huggingface/datasets/pull/250", "merged_at": "2020-06-08T09:16:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/250.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/250...
true
https://api.github.com/repos/huggingface/datasets/issues/249
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/249/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/249/comments
https://api.github.com/repos/huggingface/datasets/issues/249/events
https://github.com/huggingface/datasets/issues/249
633,393,443
MDU6SXNzdWU2MzMzOTM0NDM=
249
[Dataset created] some critical small issues when I was creating a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
2
2020-06-07 12:58:54+00:00
2020-06-12 08:28:51+00:00
2020-06-12 08:28:51+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, I successfully created a dataset and has made a pr #248. But I have encountered several problems when I was creating it, and those should be easy to fix. 1. Not found dataset_info.json should be fixed by #241 , eager to wait it be merged. 2. Forced to install `apach_beam` If we should install it, then it m...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/249/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/249/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/248
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/248/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/248/comments
https://api.github.com/repos/huggingface/datasets/issues/248/events
https://github.com/huggingface/datasets/pull/248
633,390,427
MDExOlB1bGxSZXF1ZXN0NDMwMDQ0MzU0
248
add Toronto BooksCorpus
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
null
[]
null
11
2020-06-07 12:54:56+00:00
2020-06-12 08:45:03+00:00
2020-06-12 08:45:02+00:00
CONTRIBUTOR
null
null
null
null
1. I knew there is a branch `toronto_books_corpus` - After I downloaded it, I found it is all non-english, and only have one row. - It seems that it cites the wrong paper - according to papar using it, it is called `BooksCorpus` but not `TornotoBooksCorpus` 2. It use a text mirror in google drive - `bookscorpu...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/248/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/248/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/248.diff", "html_url": "https://github.com/huggingface/datasets/pull/248", "merged_at": "2020-06-12T08:45:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/248.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/248...
true
https://api.github.com/repos/huggingface/datasets/issues/247
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/247/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/247/comments
https://api.github.com/repos/huggingface/datasets/issues/247/events
https://github.com/huggingface/datasets/pull/247
632,380,078
MDExOlB1bGxSZXF1ZXN0NDI5MTMwMzQ2
247
Make all dataset downloads deterministic by applying `sorted` to glob and os.listdir
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
3
2020-06-06 11:02:10+00:00
2020-06-08 09:18:16+00:00
2020-06-08 09:18:14+00:00
CONTRIBUTOR
null
null
null
null
This PR makes all datasets loading deterministic by applying `sorted()` to all `glob.glob` and `os.listdir` statements. Are there other "non-deterministic" functions apart from `glob.glob()` and `os.listdir()` that you can think of @thomwolf @lhoestq @mariamabarham @jplu ? **Important** It does break backward c...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/247/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/247/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/247.diff", "html_url": "https://github.com/huggingface/datasets/pull/247", "merged_at": "2020-06-08T09:18:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/247.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/247...
true
https://api.github.com/repos/huggingface/datasets/issues/246
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/246/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/246/comments
https://api.github.com/repos/huggingface/datasets/issues/246/events
https://github.com/huggingface/datasets/issues/246
632,380,054
MDU6SXNzdWU2MzIzODAwNTQ=
246
What is the best way to cache a dataset?
{ "avatar_url": "https://avatars.githubusercontent.com/u/112599?v=4", "events_url": "https://api.github.com/users/Mistobaan/events{/privacy}", "followers_url": "https://api.github.com/users/Mistobaan/followers", "following_url": "https://api.github.com/users/Mistobaan/following{/other_user}", "gists_url": "ht...
[]
closed
false
null
[]
null
2
2020-06-06 11:02:07+00:00
2020-07-09 09:15:07+00:00
2020-07-09 09:15:07+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
For example if I want to use streamlit with a nlp dataset: ``` @st.cache def load_data(): return nlp.load_dataset('squad') ``` This code raises the error "uncachable object" Right now I just fixed with a constant for my specific case: ``` @st.cache(hash_funcs={pyarrow.lib.Buffer: lambda b: 0}) ```...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/246/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/246/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/245
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/245/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/245/comments
https://api.github.com/repos/huggingface/datasets/issues/245/events
https://github.com/huggingface/datasets/issues/245
631,985,108
MDU6SXNzdWU2MzE5ODUxMDg=
245
SST-2 test labels are all -1
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
10
2020-06-05 21:41:42+00:00
2021-12-08 00:47:32+00:00
2020-06-06 16:56:41+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1. ``` >>> import nlp >>> glue = nlp.load_dataset('glue', 'sst2') >>> glue {'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'st...
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/245/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/245/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/244
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/244/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/244/comments
https://api.github.com/repos/huggingface/datasets/issues/244/events
https://github.com/huggingface/datasets/pull/244
631,869,155
MDExOlB1bGxSZXF1ZXN0NDI4NjgxMTcx
244
Add Allociné Dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/37028092?v=4", "events_url": "https://api.github.com/users/TheophileBlard/events{/privacy}", "followers_url": "https://api.github.com/users/TheophileBlard/followers", "following_url": "https://api.github.com/users/TheophileBlard/following{/other_user}", ...
[]
closed
false
null
[]
null
3
2020-06-05 19:19:26+00:00
2020-06-11 07:47:26+00:00
2020-06-11 07:47:26+00:00
CONTRIBUTOR
null
null
null
null
This is a french binary sentiment classification dataset, which was used to train this model: https://huggingface.co/tblard/tf-allocine. Basically, it's a french "IMDB" dataset, with more reviews. More info on [this repo](https://github.com/TheophileBlard/french-sentiment-analysis-with-bert).
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/244/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/244/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/244.diff", "html_url": "https://github.com/huggingface/datasets/pull/244", "merged_at": "2020-06-11T07:47:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/244.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/244...
true
https://api.github.com/repos/huggingface/datasets/issues/243
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/243/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/243/comments
https://api.github.com/repos/huggingface/datasets/issues/243/events
https://github.com/huggingface/datasets/pull/243
631,735,848
MDExOlB1bGxSZXF1ZXN0NDI4NTY2MTEy
243
Specify utf-8 encoding for GLUE
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
1
2020-06-05 16:33:00+00:00
2020-06-17 21:16:06+00:00
2020-06-08 08:42:01+00:00
CONTRIBUTOR
null
null
null
null
#242 This makes the GLUE-MNLI dataset readable on my machine, not sure if it's a Windows-only bug.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/243/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/243/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/243.diff", "html_url": "https://github.com/huggingface/datasets/pull/243", "merged_at": "2020-06-08T08:42:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/243.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/243...
true
https://api.github.com/repos/huggingface/datasets/issues/242
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/242/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/242/comments
https://api.github.com/repos/huggingface/datasets/issues/242/events
https://github.com/huggingface/datasets/issues/242
631,733,683
MDU6SXNzdWU2MzE3MzM2ODM=
242
UnicodeDecodeError when downloading GLUE-MNLI
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
2
2020-06-05 16:30:01+00:00
2020-06-09 16:06:47+00:00
2020-06-08 08:45:03+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
When I run ```python dataset = nlp.load_dataset('glue', 'mnli') ``` I get an encoding error (could it be because I'm using Windows?) : ```python # Lots of error log lines later... ~\Miniconda3\envs\nlp\lib\site-packages\tqdm\std.py in __iter__(self) 1128 try: -> 1129 for obj in iterable:...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/242/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/242/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/241
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/241/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/241/comments
https://api.github.com/repos/huggingface/datasets/issues/241/events
https://github.com/huggingface/datasets/pull/241
631,703,079
MDExOlB1bGxSZXF1ZXN0NDI4NTQwMDM0
241
Fix empty cache dir
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
2
2020-06-05 15:45:22+00:00
2020-06-08 08:35:33+00:00
2020-06-08 08:35:31+00:00
MEMBER
null
null
null
null
If the cache dir of a dataset is empty, the dataset fails to load and throws a FileNotFounfError. We could end up with empty cache dir because there was a line in the code that created the cache dir without using a temp dir. Using a temp dir is useful as it gets renamed to the real cache dir only if the full process is...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/241/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/241/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/241.diff", "html_url": "https://github.com/huggingface/datasets/pull/241", "merged_at": "2020-06-08T08:35:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/241.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/241...
true
https://api.github.com/repos/huggingface/datasets/issues/240
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/240/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/240/comments
https://api.github.com/repos/huggingface/datasets/issues/240/events
https://github.com/huggingface/datasets/issues/240
631,434,677
MDU6SXNzdWU2MzE0MzQ2Nzc=
240
Deterministic dataset loading
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
4
2020-06-05 09:03:26+00:00
2020-06-08 09:18:14+00:00
2020-06-08 09:18:14+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
When calling: ```python import nlp dataset = nlp.load_dataset("trivia_qa", split="validation[:1%]") ``` the resulting dataset is not deterministic over different google colabs. After talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line: https://github.com/huggingface/nlp/blob/2e0...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/240/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/240/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/239
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/239/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/239/comments
https://api.github.com/repos/huggingface/datasets/issues/239/events
https://github.com/huggingface/datasets/issues/239
631,340,440
MDU6SXNzdWU2MzEzNDA0NDA=
239
[Creating new dataset] Not found dataset_info.json
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
5
2020-06-05 06:15:04+00:00
2020-06-07 13:01:04+00:00
2020-06-07 13:01:04+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, I am trying to create Toronto Book Corpus. #131 I ran `~/nlp % python nlp-cli test datasets/bookcorpus --save_infos --all_configs` but this doesn't create `dataset_info.json` and try to use it ``` INFO:nlp.load:Checking datasets/bookcorpus/bookcorpus.py for additional imports. INFO:filelock:Lock 1397953257...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/239/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/239/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/238
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/238/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/238/comments
https://api.github.com/repos/huggingface/datasets/issues/238/events
https://github.com/huggingface/datasets/issues/238
631,260,143
MDU6SXNzdWU2MzEyNjAxNDM=
238
[Metric] Bertscore : Warning : Empty candidate sentence; Setting recall to be 0.
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[ { "color": "25b21e", "default": false, "description": "A bug in a metric script", "id": 2067393914, "name": "metric bug", "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug" } ]
closed
false
null
[]
null
1
2020-06-05 02:14:47+00:00
2020-06-29 17:10:19+00:00
2020-06-29 17:10:19+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
When running BERT-Score, I'm meeting this warning : > Warning: Empty candidate sentence; Setting recall to be 0. Code : ``` import nlp metric = nlp.load_metric("bertscore") scores = metric.compute(["swag", "swags"], ["swags", "totally something different"], lang="en", device=0) ``` --- **What am I do...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/238/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/238/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/237
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/237/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/237/comments
https://api.github.com/repos/huggingface/datasets/issues/237/events
https://github.com/huggingface/datasets/issues/237
631,199,940
MDU6SXNzdWU2MzExOTk5NDA=
237
Can't download MultiNLI
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
3
2020-06-04 23:05:21+00:00
2020-06-06 10:51:34+00:00
2020-06-06 10:51:34+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
When I try to download MultiNLI with ```python dataset = load_dataset('multi_nli') ``` I get this long error: ```python --------------------------------------------------------------------------- OSError Traceback (most recent call last) <ipython-input-13-3b11f6be4cb9> in <m...
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/237/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/237/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/236
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/236/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/236/comments
https://api.github.com/repos/huggingface/datasets/issues/236/events
https://github.com/huggingface/datasets/pull/236
631,099,875
MDExOlB1bGxSZXF1ZXN0NDI4MDUwNzI4
236
CompGuessWhat?! dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/1479733?v=4", "events_url": "https://api.github.com/users/aleSuglia/events{/privacy}", "followers_url": "https://api.github.com/users/aleSuglia/followers", "following_url": "https://api.github.com/users/aleSuglia/following{/other_user}", "gists_url": "h...
[]
closed
false
null
[]
null
9
2020-06-04 19:45:50+00:00
2020-06-11 09:43:42+00:00
2020-06-11 07:45:21+00:00
CONTRIBUTOR
null
null
null
null
Hello, Thanks for the amazing library that you put together. I'm Alessandro Suglia, the first author of CompGuessWhat?!, a recently released dataset for grounded language learning accepted to ACL 2020 ([https://compguesswhat.github.io](https://compguesswhat.github.io)). This pull-request adds the CompGuessWhat?! ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/236/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/236/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/236.diff", "html_url": "https://github.com/huggingface/datasets/pull/236", "merged_at": "2020-06-11T07:45:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/236.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/236...
true
https://api.github.com/repos/huggingface/datasets/issues/235
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/235/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/235/comments
https://api.github.com/repos/huggingface/datasets/issues/235/events
https://github.com/huggingface/datasets/pull/235
630,952,297
MDExOlB1bGxSZXF1ZXN0NDI3OTM1MjQ0
235
Add experimental datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
6
2020-06-04 15:54:56+00:00
2020-06-12 15:38:55+00:00
2020-06-12 15:38:55+00:00
MEMBER
null
null
null
null
## Adding an *experimental datasets* folder After using the 🤗nlp library for some time, I find that while it makes it super easy to create new memory-mapped datasets with lots of cool utilities, a lot of what I want to do doesn't work well with the current `MockDownloader` based testing paradigm, making it hard to ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/235/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/235/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/235.diff", "html_url": "https://github.com/huggingface/datasets/pull/235", "merged_at": "2020-06-12T15:38:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/235.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/235...
true
https://api.github.com/repos/huggingface/datasets/issues/234
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/234/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/234/comments
https://api.github.com/repos/huggingface/datasets/issues/234/events
https://github.com/huggingface/datasets/issues/234
630,534,427
MDU6SXNzdWU2MzA1MzQ0Mjc=
234
Huggingface NLP, Uploading custom dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42269506?v=4", "events_url": "https://api.github.com/users/Nouman97/events{/privacy}", "followers_url": "https://api.github.com/users/Nouman97/followers", "following_url": "https://api.github.com/users/Nouman97/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
4
2020-06-04 05:59:06+00:00
2020-07-06 09:33:26+00:00
2020-07-06 09:33:26+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hello, Does anyone know how we can call our custom dataset using the nlp.load command? Let's say that I have a dataset based on the same format as that of squad-v1.1, how am I supposed to load it using huggingface nlp. Thank you!
{ "avatar_url": "https://avatars.githubusercontent.com/u/42269506?v=4", "events_url": "https://api.github.com/users/Nouman97/events{/privacy}", "followers_url": "https://api.github.com/users/Nouman97/followers", "following_url": "https://api.github.com/users/Nouman97/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/234/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/234/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/233
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/233/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/233/comments
https://api.github.com/repos/huggingface/datasets/issues/233/events
https://github.com/huggingface/datasets/issues/233
630,432,132
MDU6SXNzdWU2MzA0MzIxMzI=
233
Fail to download c4 english corpus
{ "avatar_url": "https://avatars.githubusercontent.com/u/16605764?v=4", "events_url": "https://api.github.com/users/donggyukimc/events{/privacy}", "followers_url": "https://api.github.com/users/donggyukimc/followers", "following_url": "https://api.github.com/users/donggyukimc/following{/other_user}", "gists_u...
[]
closed
false
null
[]
null
5
2020-06-04 01:06:38+00:00
2021-01-08 07:17:32+00:00
2020-06-08 09:16:59+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
i run following code to download c4 English corpus. ``` dataset = nlp.load_dataset('c4', 'en', beam_runner='DirectRunner' , data_dir='/mypath') ``` and i met failure as follows ``` Downloading and preparing dataset c4/en (download: Unknown size, generated: Unknown size, total: Unknown size) to /home/adam/....
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 3, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/233/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/233/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/232
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/232/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/232/comments
https://api.github.com/repos/huggingface/datasets/issues/232/events
https://github.com/huggingface/datasets/pull/232
630,029,568
MDExOlB1bGxSZXF1ZXN0NDI3MjI5NDcy
232
Nlp cli fix endpoints
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
1
2020-06-03 14:10:39+00:00
2020-06-08 09:02:58+00:00
2020-06-08 09:02:57+00:00
MEMBER
null
null
null
null
With this PR users will be able to upload their own datasets and metrics. As mentioned in #181, I had to use the new endpoints and revert the use of dataclasses (just in case we have changes in the API in the future). We now distinguish commands for datasets and commands for metrics: ```bash nlp-cli upload_data...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/232/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/232/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/232.diff", "html_url": "https://github.com/huggingface/datasets/pull/232", "merged_at": "2020-06-08T09:02:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/232.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/232...
true
https://api.github.com/repos/huggingface/datasets/issues/231
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/231/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/231/comments
https://api.github.com/repos/huggingface/datasets/issues/231/events
https://github.com/huggingface/datasets/pull/231
629,988,694
MDExOlB1bGxSZXF1ZXN0NDI3MTk3MTcz
231
Add .download to MockDownloadManager
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-06-03 13:20:00+00:00
2020-06-03 14:25:56+00:00
2020-06-03 14:25:55+00:00
MEMBER
null
null
null
null
One method from the DownloadManager was missing and some users couldn't run the tests because of that. @yjernite
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/231/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/231/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/231.diff", "html_url": "https://github.com/huggingface/datasets/pull/231", "merged_at": "2020-06-03T14:25:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/231.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/231...
true
https://api.github.com/repos/huggingface/datasets/issues/230
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/230/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/230/comments
https://api.github.com/repos/huggingface/datasets/issues/230/events
https://github.com/huggingface/datasets/pull/230
629,983,684
MDExOlB1bGxSZXF1ZXN0NDI3MTkzMTQ0
230
Don't force to install apache beam for wikipedia dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-06-03 13:13:07+00:00
2020-06-03 14:34:09+00:00
2020-06-03 14:34:07+00:00
MEMBER
null
null
null
null
As pointed out in #227, we shouldn't force users to install apache beam if the processed dataset can be downloaded. I moved the imports of some datasets to avoid this problem
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/230/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/230/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/230.diff", "html_url": "https://github.com/huggingface/datasets/pull/230", "merged_at": "2020-06-03T14:34:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/230.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/230...
true
https://api.github.com/repos/huggingface/datasets/issues/227
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/227/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/227/comments
https://api.github.com/repos/huggingface/datasets/issues/227/events
https://github.com/huggingface/datasets/issues/227
629,845,704
MDU6SXNzdWU2Mjk4NDU3MDQ=
227
Should we still have to force to install apache_beam to download wikipedia ?
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
3
2020-06-03 09:33:20+00:00
2020-06-03 15:25:41+00:00
2020-06-03 15:25:41+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, first thanks to @lhoestq 's revolutionary work, I successfully downloaded processed wikipedia according to the doc. 😍😍😍 But at the first try, it tell me to install `apache_beam` and `mwparserfromhell`, which I thought wouldn't be used according to #204 , it was kind of confusing me at that time. Maybe we s...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/227/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/227/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/226
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/226/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/226/comments
https://api.github.com/repos/huggingface/datasets/issues/226/events
https://github.com/huggingface/datasets/pull/226
628,344,520
MDExOlB1bGxSZXF1ZXN0NDI1OTA0MjEz
226
add BlendedSkillTalk dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
1
2020-06-01 10:54:45+00:00
2020-06-03 14:37:23+00:00
2020-06-03 14:37:22+00:00
CONTRIBUTOR
null
null
null
null
This PR add the BlendedSkillTalk dataset, which is used to fine tune the blenderbot.
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/226/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/226/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/226.diff", "html_url": "https://github.com/huggingface/datasets/pull/226", "merged_at": "2020-06-03T14:37:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/226.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/226...
true
https://api.github.com/repos/huggingface/datasets/issues/225
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/225/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/225/comments
https://api.github.com/repos/huggingface/datasets/issues/225/events
https://github.com/huggingface/datasets/issues/225
628,083,366
MDU6SXNzdWU2MjgwODMzNjY=
225
[ROUGE] Different scores with `files2rouge`
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[ { "color": "d722e8", "default": false, "description": "Discussions on the metrics", "id": 2067400959, "name": "Metric discussion", "node_id": "MDU6TGFiZWwyMDY3NDAwOTU5", "url": "https://api.github.com/repos/huggingface/datasets/labels/Metric%20discussion" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gi...
null
3
2020-06-01 00:50:36+00:00
2020-06-03 15:27:18+00:00
2020-06-03 15:27:18+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
It seems that the ROUGE score of `nlp` is lower than the one of `files2rouge`. Here is a self-contained notebook to reproduce both scores : https://colab.research.google.com/drive/14EyAXValB6UzKY9x4rs_T3pyL7alpw_F?usp=sharing --- `nlp` : (Only mid F-scores) >rouge1 0.33508031962733364 rouge2 0.145743337761...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/225/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/225/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/224
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/224/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/224/comments
https://api.github.com/repos/huggingface/datasets/issues/224/events
https://github.com/huggingface/datasets/issues/224
627,791,693
MDU6SXNzdWU2Mjc3OTE2OTM=
224
[Feature Request/Help] BLEURT model -> PyTorch
{ "avatar_url": "https://avatars.githubusercontent.com/u/6889910?v=4", "events_url": "https://api.github.com/users/adamwlev/events{/privacy}", "followers_url": "https://api.github.com/users/adamwlev/followers", "following_url": "https://api.github.com/users/adamwlev/following{/other_user}", "gists_url": "http...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gi...
null
6
2020-05-30 18:30:40+00:00
2023-08-26 17:38:48+00:00
2021-01-04 09:53:32+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, I am interested in porting google research's new BLEURT learned metric to PyTorch (because I wish to do something experimental with language generation and backpropping through BLEURT). I noticed that you guys don't have it yet so I am partly just asking if you plan to add it (@thomwolf said you want to do so on Tw...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/224/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/224/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/223
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/223/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/223/comments
https://api.github.com/repos/huggingface/datasets/issues/223/events
https://github.com/huggingface/datasets/issues/223
627,683,386
MDU6SXNzdWU2Mjc2ODMzODY=
223
[Feature request] Add FLUE dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/58078086?v=4", "events_url": "https://api.github.com/users/lbourdois/events{/privacy}", "followers_url": "https://api.github.com/users/lbourdois/followers", "following_url": "https://api.github.com/users/lbourdois/following{/other_user}", "gists_url": "...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
[]
null
3
2020-05-30 08:52:15+00:00
2020-12-03 13:39:33+00:00
2020-12-03 13:39:33+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, I think it would be interesting to add the FLUE dataset for francophones or anyone wishing to work on French. In other requests, I read that you are already working on some datasets, and I was wondering if FLUE was planned. If it is not the case, I can provide each of the cleaned FLUE datasets (in the form...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/223/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/223/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/222
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/222/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/222/comments
https://api.github.com/repos/huggingface/datasets/issues/222/events
https://github.com/huggingface/datasets/issues/222
627,586,690
MDU6SXNzdWU2Mjc1ODY2OTA=
222
Colab Notebook breaks when downloading the squad dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/338917?v=4", "events_url": "https://api.github.com/users/carlos-aguayo/events{/privacy}", "followers_url": "https://api.github.com/users/carlos-aguayo/followers", "following_url": "https://api.github.com/users/carlos-aguayo/following{/other_user}", "gis...
[]
closed
false
null
[]
null
6
2020-05-29 22:55:59+00:00
2020-06-04 00:21:05+00:00
2020-06-04 00:21:05+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
When I run the notebook in Colab https://colab.research.google.com/github/huggingface/nlp/blob/master/notebooks/Overview.ipynb breaks when running this cell: ![image](https://user-images.githubusercontent.com/338917/83311709-ffd1b800-a1dd-11ea-8394-3a87df0d7f8b.png)
{ "avatar_url": "https://avatars.githubusercontent.com/u/338917?v=4", "events_url": "https://api.github.com/users/carlos-aguayo/events{/privacy}", "followers_url": "https://api.github.com/users/carlos-aguayo/followers", "following_url": "https://api.github.com/users/carlos-aguayo/following{/other_user}", "gis...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/222/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/222/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/221
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/221/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/221/comments
https://api.github.com/repos/huggingface/datasets/issues/221/events
https://github.com/huggingface/datasets/pull/221
627,300,648
MDExOlB1bGxSZXF1ZXN0NDI1MTI5OTc0
221
Fix tests/test_dataset_common.py
{ "avatar_url": "https://avatars.githubusercontent.com/u/13635495?v=4", "events_url": "https://api.github.com/users/tayciryahmed/events{/privacy}", "followers_url": "https://api.github.com/users/tayciryahmed/followers", "following_url": "https://api.github.com/users/tayciryahmed/following{/other_user}", "gist...
[]
closed
false
null
[]
null
1
2020-05-29 14:12:15+00:00
2020-06-01 12:20:42+00:00
2020-05-29 15:02:23+00:00
CONTRIBUTOR
null
null
null
null
When I run the command `RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_arcd` while working on #220. I get the error ` unexpected keyword argument "'download_and_prepare_kwargs'"` at the level of `load_dataset`. Indeed, this [function](https://github.com/huggingface/nlp/blob/ma...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/221/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/221/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/221.diff", "html_url": "https://github.com/huggingface/datasets/pull/221", "merged_at": "2020-05-29T15:02:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/221.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/221...
true
https://api.github.com/repos/huggingface/datasets/issues/220
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/220/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/220/comments
https://api.github.com/repos/huggingface/datasets/issues/220/events
https://github.com/huggingface/datasets/pull/220
627,280,683
MDExOlB1bGxSZXF1ZXN0NDI1MTEzMzEy
220
dataset_arcd
{ "avatar_url": "https://avatars.githubusercontent.com/u/13635495?v=4", "events_url": "https://api.github.com/users/tayciryahmed/events{/privacy}", "followers_url": "https://api.github.com/users/tayciryahmed/followers", "following_url": "https://api.github.com/users/tayciryahmed/following{/other_user}", "gist...
[]
closed
false
null
[]
null
2
2020-05-29 13:46:50+00:00
2020-05-29 14:58:40+00:00
2020-05-29 14:57:21+00:00
CONTRIBUTOR
null
null
null
null
Added Arabic Reading Comprehension Dataset (ARCD): https://arxiv.org/abs/1906.05394
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/220/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/220/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/220.diff", "html_url": "https://github.com/huggingface/datasets/pull/220", "merged_at": "2020-05-29T14:57:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/220.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/220...
true
https://api.github.com/repos/huggingface/datasets/issues/219
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/219/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/219/comments
https://api.github.com/repos/huggingface/datasets/issues/219/events
https://github.com/huggingface/datasets/pull/219
627,235,893
MDExOlB1bGxSZXF1ZXN0NDI1MDc2NjQx
219
force mwparserfromhell as third party
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-05-29 12:33:17+00:00
2020-05-29 13:30:13+00:00
2020-05-29 13:30:12+00:00
MEMBER
null
null
null
null
This should fix your env because you had `mwparserfromhell ` as a first party for `isort` @patrickvonplaten
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/219/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/219/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/219.diff", "html_url": "https://github.com/huggingface/datasets/pull/219", "merged_at": "2020-05-29T13:30:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/219.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/219...
true
https://api.github.com/repos/huggingface/datasets/issues/218
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/218/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/218/comments
https://api.github.com/repos/huggingface/datasets/issues/218/events
https://github.com/huggingface/datasets/pull/218
627,173,407
MDExOlB1bGxSZXF1ZXN0NDI1MDI2NzEz
218
Add Natual Questions and C4 scripts
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-05-29 10:40:30+00:00
2020-05-29 12:31:01+00:00
2020-05-29 12:31:00+00:00
MEMBER
null
null
null
null
Scripts are ready ! However they are not processed nor directly available from gcp yet.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/218/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/218/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/218.diff", "html_url": "https://github.com/huggingface/datasets/pull/218", "merged_at": "2020-05-29T12:31:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/218.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/218...
true
https://api.github.com/repos/huggingface/datasets/issues/217
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/217/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/217/comments
https://api.github.com/repos/huggingface/datasets/issues/217/events
https://github.com/huggingface/datasets/issues/217
627,128,403
MDU6SXNzdWU2MjcxMjg0MDM=
217
Multi-task dataset mixing
{ "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gist...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "c5def5", "default": fals...
open
false
null
[]
null
28
2020-05-29 09:22:26+00:00
2025-09-24 08:59:38+00:00
NaT
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks). The [T5 paper](https://arxiv.org/pdf/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning: - **Examples-proportional mixing** - sam...
null
{ "+1": 12, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 12, "url": "https://api.github.com/repos/huggingface/datasets/issues/217/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/217/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/216
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/216/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/216/comments
https://api.github.com/repos/huggingface/datasets/issues/216/events
https://github.com/huggingface/datasets/issues/216
626,896,890
MDU6SXNzdWU2MjY4OTY4OTA=
216
❓ How to get ROUGE-2 with the ROUGE metric ?
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
3
2020-05-28 23:47:32+00:00
2020-06-01 00:04:35+00:00
2020-06-01 00:04:35+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I'm trying to use ROUGE metric, but I don't know how to get the ROUGE-2 metric. --- I compute scores with : ```python import nlp rouge = nlp.load_metric('rouge') with open("pred.txt") as p, open("ref.txt") as g: for lp, lg in zip(p, g): rouge.add([lp], [lg]) score = rouge.compute() ``` ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/216/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/216/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/215
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/215/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/215/comments
https://api.github.com/repos/huggingface/datasets/issues/215/events
https://github.com/huggingface/datasets/issues/215
626,867,879
MDU6SXNzdWU2MjY4Njc4Nzk=
215
NonMatchingSplitsSizesError when loading blog_authorship_corpus
{ "avatar_url": "https://avatars.githubusercontent.com/u/52105365?v=4", "events_url": "https://api.github.com/users/cedricconol/events{/privacy}", "followers_url": "https://api.github.com/users/cedricconol/followers", "following_url": "https://api.github.com/users/cedricconol/following{/other_user}", "gists_u...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
12
2020-05-28 22:55:19+00:00
2025-01-04 00:03:12+00:00
2022-02-10 13:05:45+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Getting this error when i run `nlp.load_dataset('blog_authorship_corpus')`. ``` raise NonMatchingSplitsSizesError(str(bad_splits)) nlp.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/215/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/215/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/214
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/214/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/214/comments
https://api.github.com/repos/huggingface/datasets/issues/214/events
https://github.com/huggingface/datasets/pull/214
626,641,549
MDExOlB1bGxSZXF1ZXN0NDI0NTk1NjIx
214
[arrow_dataset.py] add new filter function
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
13
2020-05-28 16:21:40+00:00
2020-05-29 11:43:29+00:00
2020-05-29 11:32:20+00:00
CONTRIBUTOR
null
null
null
null
The `.map()` function is super useful, but can IMO a bit tedious when filtering certain examples. I think, filtering out examples is also a very common operation people would like to perform on datasets. This PR is a proposal to add a `.filter()` function in the same spirit than the `.map()` function. Here is a ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/214/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/214/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/214.diff", "html_url": "https://github.com/huggingface/datasets/pull/214", "merged_at": "2020-05-29T11:32:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/214.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/214...
true
https://api.github.com/repos/huggingface/datasets/issues/213
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/213/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/213/comments
https://api.github.com/repos/huggingface/datasets/issues/213/events
https://github.com/huggingface/datasets/pull/213
626,587,995
MDExOlB1bGxSZXF1ZXN0NDI0NTUxODE3
213
better message if missing beam options
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-05-28 15:06:57+00:00
2020-05-29 09:51:17+00:00
2020-05-29 09:51:16+00:00
MEMBER
null
null
null
null
WDYT @yjernite ? For example: ```python dataset = nlp.load_dataset('wikipedia', '20200501.aa') ``` Raises: ``` MissingBeamOptions: Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided in `load_dataset` or in the builder arguments. For big datasets it has to ru...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/213/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/213/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/213.diff", "html_url": "https://github.com/huggingface/datasets/pull/213", "merged_at": "2020-05-29T09:51:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/213.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/213...
true
https://api.github.com/repos/huggingface/datasets/issues/212
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/212/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/212/comments
https://api.github.com/repos/huggingface/datasets/issues/212/events
https://github.com/huggingface/datasets/pull/212
626,580,198
MDExOlB1bGxSZXF1ZXN0NDI0NTQ1NjAy
212
have 'add' and 'add_batch' for metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-05-28 14:56:47+00:00
2020-05-29 10:41:05+00:00
2020-05-29 10:41:04+00:00
MEMBER
null
null
null
null
This should fix #116 Previously the `.add` method of metrics expected a batch of examples. Now `.add` expects one prediction/reference and `.add_batch` expects a batch. I think it is more coherent with the way the ArrowWriter works.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/212/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/212/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/212.diff", "html_url": "https://github.com/huggingface/datasets/pull/212", "merged_at": "2020-05-29T10:41:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/212.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/212...
true
https://api.github.com/repos/huggingface/datasets/issues/211
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/211/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/211/comments
https://api.github.com/repos/huggingface/datasets/issues/211/events
https://github.com/huggingface/datasets/issues/211
626,565,994
MDU6SXNzdWU2MjY1NjU5OTQ=
211
[Arrow writer, Trivia_qa] Could not convert TagMe with type str: converting to null type
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gis...
null
7
2020-05-28 14:38:14+00:00
2020-07-23 10:15:16+00:00
2020-07-23 10:15:16+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Running the following code ``` import nlp ds = nlp.load_dataset("trivia_qa", "rc", split="validation[:1%]") # this might take 2.3 min to download but it's cached afterwards... ds.map(lambda x: x, load_from_cache_file=False) ``` triggers a `ArrowInvalid: Could not convert TagMe with type str: converting to n...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/211/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/211/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/210
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/210/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/210/comments
https://api.github.com/repos/huggingface/datasets/issues/210/events
https://github.com/huggingface/datasets/pull/210
626,504,243
MDExOlB1bGxSZXF1ZXN0NDI0NDgyNDgz
210
fix xnli metric kwargs description
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-05-28 13:21:44+00:00
2020-05-28 13:22:11+00:00
2020-05-28 13:22:10+00:00
MEMBER
null
null
null
null
The text was wrong as noticed in #202
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/210/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/210/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/210.diff", "html_url": "https://github.com/huggingface/datasets/pull/210", "merged_at": "2020-05-28T13:22:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/210.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/210...
true
https://api.github.com/repos/huggingface/datasets/issues/209
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/209/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/209/comments
https://api.github.com/repos/huggingface/datasets/issues/209/events
https://github.com/huggingface/datasets/pull/209
626,405,849
MDExOlB1bGxSZXF1ZXN0NDI0NDAwOTc4
209
Add a Google Drive exception for small files
{ "avatar_url": "https://avatars.githubusercontent.com/u/25703835?v=4", "events_url": "https://api.github.com/users/airKlizz/events{/privacy}", "followers_url": "https://api.github.com/users/airKlizz/followers", "following_url": "https://api.github.com/users/airKlizz/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
3
2020-05-28 10:40:17+00:00
2020-05-28 15:15:04+00:00
2020-05-28 15:15:04+00:00
CONTRIBUTOR
null
null
null
null
I tried to use the ``nlp`` library to load personnal datasets. I mainly copy-paste the code for ``multi-news`` dataset because my files are stored on Google Drive. One of my dataset is small (< 25Mo) so it can be verified by Drive without asking the authorization to the user. This makes the download starts directly...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/209/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/209/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/209.diff", "html_url": "https://github.com/huggingface/datasets/pull/209", "merged_at": "2020-05-28T15:15:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/209.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/209...
true
https://api.github.com/repos/huggingface/datasets/issues/208
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/208/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/208/comments
https://api.github.com/repos/huggingface/datasets/issues/208/events
https://github.com/huggingface/datasets/pull/208
626,398,519
MDExOlB1bGxSZXF1ZXN0NDI0Mzk0ODIx
208
[Dummy data] insert config name instead of config
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
0
2020-05-28 10:28:19+00:00
2020-05-28 12:48:01+00:00
2020-05-28 12:48:00+00:00
CONTRIBUTOR
null
null
null
null
Thanks @yjernite for letting me know. in the dummy data command the config name shuold be passed to the dataset builder and not the config itself. Also, @lhoestq fixed small import bug introduced by beam command I think.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/208/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/208/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/208.diff", "html_url": "https://github.com/huggingface/datasets/pull/208", "merged_at": "2020-05-28T12:48:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/208.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/208...
true
https://api.github.com/repos/huggingface/datasets/issues/207
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/207/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/207/comments
https://api.github.com/repos/huggingface/datasets/issues/207/events
https://github.com/huggingface/datasets/issues/207
625,932,200
MDU6SXNzdWU2MjU5MzIyMDA=
207
Remove test set from NLP viewer
{ "avatar_url": "https://avatars.githubusercontent.com/u/748399?v=4", "events_url": "https://api.github.com/users/chrisdonahue/events{/privacy}", "followers_url": "https://api.github.com/users/chrisdonahue/followers", "following_url": "https://api.github.com/users/chrisdonahue/following{/other_user}", "gists_...
[ { "color": "94203D", "default": false, "description": "", "id": 2107841032, "name": "nlp-viewer", "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer" } ]
closed
false
null
[]
null
3
2020-05-27 18:32:07+00:00
2022-02-10 13:17:45+00:00
2022-02-10 13:17:45+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
While the new [NLP viewer](https://huggingface.co/nlp/viewer/) is a great tool, I think it would be best to outright remove the option of looking at the test sets. At the very least, a warning should be displayed to users before showing the test set. Newcomers to the field might not be aware of best practices, and smal...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
{ "+1": 3, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/207/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/207/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/206
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/206/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/206/comments
https://api.github.com/repos/huggingface/datasets/issues/206/events
https://github.com/huggingface/datasets/issues/206
625,842,989
MDU6SXNzdWU2MjU4NDI5ODk=
206
[Question] Combine 2 datasets which have the same columns
{ "avatar_url": "https://avatars.githubusercontent.com/u/25703835?v=4", "events_url": "https://api.github.com/users/airKlizz/events{/privacy}", "followers_url": "https://api.github.com/users/airKlizz/followers", "following_url": "https://api.github.com/users/airKlizz/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
2
2020-05-27 16:25:52+00:00
2020-06-10 09:11:14+00:00
2020-06-10 09:11:14+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, I am using ``nlp`` to load personal datasets. I created summarization datasets in multi-languages based on wikinews. I have one dataset for english and one for german (french is getting to be ready as well). I want to keep these datasets independent because they need different pre-processing (add different task-...
{ "avatar_url": "https://avatars.githubusercontent.com/u/25703835?v=4", "events_url": "https://api.github.com/users/airKlizz/events{/privacy}", "followers_url": "https://api.github.com/users/airKlizz/followers", "following_url": "https://api.github.com/users/airKlizz/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/206/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/206/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/205
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/205/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/205/comments
https://api.github.com/repos/huggingface/datasets/issues/205/events
https://github.com/huggingface/datasets/pull/205
625,839,335
MDExOlB1bGxSZXF1ZXN0NDIzOTY2ODE1
205
Better arrow dataset iter
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-05-27 16:20:21+00:00
2020-05-27 16:39:58+00:00
2020-05-27 16:39:56+00:00
MEMBER
null
null
null
null
I tried to play around with `tf.data.Dataset.from_generator` and I found out that the `__iter__` that we have for `nlp.arrow_dataset.Dataset` ignores the format that has been set (torch or tensorflow). With these changes I should be able to come up with a `tf.data.Dataset` that uses lazy loading, as asked in #193.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/205/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/205/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/205.diff", "html_url": "https://github.com/huggingface/datasets/pull/205", "merged_at": "2020-05-27T16:39:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/205.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/205...
true
https://api.github.com/repos/huggingface/datasets/issues/204
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/204/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/204/comments
https://api.github.com/repos/huggingface/datasets/issues/204/events
https://github.com/huggingface/datasets/pull/204
625,655,849
MDExOlB1bGxSZXF1ZXN0NDIzODE5MTQw
204
Add Dataflow support + Wikipedia + Wiki40b
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-05-27 12:32:49+00:00
2020-05-28 08:10:35+00:00
2020-05-28 08:10:34+00:00
MEMBER
null
null
null
null
# Add Dataflow support + Wikipedia + Wiki40b ## Support datasets processing with Apache Beam Some datasets are too big to be processed on a single machine, for example: wikipedia, wiki40b, etc. Apache Beam allows to process datasets on many execution engines like Dataflow, Spark, Flink, etc. To process such da...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/204/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/204/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/204.diff", "html_url": "https://github.com/huggingface/datasets/pull/204", "merged_at": "2020-05-28T08:10:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/204.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/204...
true
https://api.github.com/repos/huggingface/datasets/issues/203
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/203/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/203/comments
https://api.github.com/repos/huggingface/datasets/issues/203/events
https://github.com/huggingface/datasets/pull/203
625,515,488
MDExOlB1bGxSZXF1ZXN0NDIzNzEyMTQ3
203
Raise an error if no config name for datasets like glue
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-05-27 09:03:58+00:00
2020-05-27 16:40:39+00:00
2020-05-27 16:40:38+00:00
MEMBER
null
null
null
null
Some datasets like glue (see #130) and scientific_papers (see #197) have many configs. For example for glue there are cola, sst2, mrpc etc. Currently if a user does `load_dataset('glue')`, then Cola is loaded by default and it can be confusing. Instead, we should raise an error to let the user know that he has to p...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/203/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/203/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/203.diff", "html_url": "https://github.com/huggingface/datasets/pull/203", "merged_at": "2020-05-27T16:40:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/203.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/203...
true
https://api.github.com/repos/huggingface/datasets/issues/202
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/202/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/202/comments
https://api.github.com/repos/huggingface/datasets/issues/202/events
https://github.com/huggingface/datasets/issues/202
625,493,983
MDU6SXNzdWU2MjU0OTM5ODM=
202
Mistaken `_KWARGS_DESCRIPTION` for XNLI metric
{ "avatar_url": "https://avatars.githubusercontent.com/u/33572125?v=4", "events_url": "https://api.github.com/users/phiyodr/events{/privacy}", "followers_url": "https://api.github.com/users/phiyodr/followers", "following_url": "https://api.github.com/users/phiyodr/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
1
2020-05-27 08:34:42+00:00
2020-05-28 13:22:36+00:00
2020-05-28 13:22:36+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi! The [`_KWARGS_DESCRIPTION`](https://github.com/huggingface/nlp/blob/7d0fa58641f3f462fb2861dcdd6ce7f0da3f6a56/metrics/xnli/xnli.py#L45) for the XNLI metric uses `Args` and `Returns` text from [BLEU](https://github.com/huggingface/nlp/blob/7d0fa58641f3f462fb2861dcdd6ce7f0da3f6a56/metrics/bleu/bleu.py#L58) metric: ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/202/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/202/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/201
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/201/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/201/comments
https://api.github.com/repos/huggingface/datasets/issues/201/events
https://github.com/huggingface/datasets/pull/201
625,235,430
MDExOlB1bGxSZXF1ZXN0NDIzNDkzNTMw
201
Fix typo in README
{ "avatar_url": "https://avatars.githubusercontent.com/u/30755778?v=4", "events_url": "https://api.github.com/users/LysandreJik/events{/privacy}", "followers_url": "https://api.github.com/users/LysandreJik/followers", "following_url": "https://api.github.com/users/LysandreJik/following{/other_user}", "gists_u...
[]
closed
false
null
[]
null
2
2020-05-26 22:18:21+00:00
2020-05-26 23:40:31+00:00
2020-05-26 23:00:56+00:00
MEMBER
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/201/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/201/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/201.diff", "html_url": "https://github.com/huggingface/datasets/pull/201", "merged_at": "2020-05-26T23:00:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/201.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/201...
true
https://api.github.com/repos/huggingface/datasets/issues/200
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/200/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/200/comments
https://api.github.com/repos/huggingface/datasets/issues/200/events
https://github.com/huggingface/datasets/pull/200
625,226,638
MDExOlB1bGxSZXF1ZXN0NDIzNDg2NTM0
200
[ArrowWriter] Set schema at first write example
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
1
2020-05-26 21:59:48+00:00
2020-05-27 09:07:54+00:00
2020-05-27 09:07:53+00:00
MEMBER
null
null
null
null
Right now if the schema was not specified when instantiating `ArrowWriter`, then it could be set with the first `write_table` for example (it calls `self._build_writer()` to do so). I noticed that it was not done if the first example is added via `.write`, so I added it for coherence.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/200/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/200/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/200.diff", "html_url": "https://github.com/huggingface/datasets/pull/200", "merged_at": "2020-05-27T09:07:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/200.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/200...
true
https://api.github.com/repos/huggingface/datasets/issues/199
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/199/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/199/comments
https://api.github.com/repos/huggingface/datasets/issues/199/events
https://github.com/huggingface/datasets/pull/199
625,217,440
MDExOlB1bGxSZXF1ZXN0NDIzNDc4ODIx
199
Fix GermEval 2014 dataset infos
{ "avatar_url": "https://avatars.githubusercontent.com/u/20651387?v=4", "events_url": "https://api.github.com/users/stefan-it/events{/privacy}", "followers_url": "https://api.github.com/users/stefan-it/followers", "following_url": "https://api.github.com/users/stefan-it/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
2
2020-05-26 21:41:44+00:00
2020-05-26 21:50:24+00:00
2020-05-26 21:50:24+00:00
CONTRIBUTOR
null
null
null
null
Hi, this PR just removes the `dataset_info.json` file and adds a newly generated `dataset_infos.json` file.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/199/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/199/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/199.diff", "html_url": "https://github.com/huggingface/datasets/pull/199", "merged_at": "2020-05-26T21:50:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/199.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/199...
true
https://api.github.com/repos/huggingface/datasets/issues/198
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/198/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/198/comments
https://api.github.com/repos/huggingface/datasets/issues/198/events
https://github.com/huggingface/datasets/issues/198
625,200,627
MDU6SXNzdWU2MjUyMDA2Mjc=
198
Index outside of table length
{ "avatar_url": "https://avatars.githubusercontent.com/u/305717?v=4", "events_url": "https://api.github.com/users/casajarm/events{/privacy}", "followers_url": "https://api.github.com/users/casajarm/followers", "following_url": "https://api.github.com/users/casajarm/following{/other_user}", "gists_url": "https...
[]
closed
false
null
[]
null
2
2020-05-26 21:09:40+00:00
2020-05-26 22:43:49+00:00
2020-05-26 22:43:49+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
The offset input box warns of numbers larger than a limit (like 2000) but then the errors start at a smaller value than that limit (like 1955). > ValueError: Index (2000) outside of table length (2000). > Traceback: > File "/home/sasha/.local/lib/python3.7/site-packages/streamlit/ScriptRunner.py", line 322, in _ru...
{ "avatar_url": "https://avatars.githubusercontent.com/u/35882?v=4", "events_url": "https://api.github.com/users/srush/events{/privacy}", "followers_url": "https://api.github.com/users/srush/followers", "following_url": "https://api.github.com/users/srush/following{/other_user}", "gists_url": "https://api.git...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/198/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/198/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/197
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/197/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/197/comments
https://api.github.com/repos/huggingface/datasets/issues/197/events
https://github.com/huggingface/datasets/issues/197
624,966,904
MDU6SXNzdWU2MjQ5NjY5MDQ=
197
Scientific Papers only downloading Pubmed
{ "avatar_url": "https://avatars.githubusercontent.com/u/17463361?v=4", "events_url": "https://api.github.com/users/antmarakis/events{/privacy}", "followers_url": "https://api.github.com/users/antmarakis/followers", "following_url": "https://api.github.com/users/antmarakis/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
3
2020-05-26 15:18:47+00:00
2020-05-28 08:19:28+00:00
2020-05-28 08:19:28+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi! I have been playing around with this module, and I am a bit confused about the `scientific_papers` dataset. I thought that it would download two separate datasets, arxiv and pubmed. But when I run the following: ``` dataset = nlp.load_dataset('scientific_papers', data_dir='.', cache_dir='.') Downloading: 10...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/197/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/197/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/196
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/196/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/196/comments
https://api.github.com/repos/huggingface/datasets/issues/196/events
https://github.com/huggingface/datasets/pull/196
624,901,266
MDExOlB1bGxSZXF1ZXN0NDIzMjIwMjIw
196
Check invalid config name
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
13
2020-05-26 13:52:51+00:00
2020-05-26 21:04:56+00:00
2020-05-26 21:04:55+00:00
MEMBER
null
null
null
null
As said in #194, we should raise an error if the config name has bad characters. Bad characters are those that are not allowed for directory names on windows.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/196/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/196/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/196.diff", "html_url": "https://github.com/huggingface/datasets/pull/196", "merged_at": "2020-05-26T21:04:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/196.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/196...
true
https://api.github.com/repos/huggingface/datasets/issues/195
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/195/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/195/comments
https://api.github.com/repos/huggingface/datasets/issues/195/events
https://github.com/huggingface/datasets/pull/195
624,858,686
MDExOlB1bGxSZXF1ZXN0NDIzMTg1NTAy
195
[Dummy data command] add new case to command
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
1
2020-05-26 12:50:47+00:00
2020-05-26 14:38:28+00:00
2020-05-26 14:38:27+00:00
CONTRIBUTOR
null
null
null
null
Qanta: #194 introduces a case that was not noticed before. This change in code helps community users to have an easier time creating the dummy data.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/195/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/195/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/195.diff", "html_url": "https://github.com/huggingface/datasets/pull/195", "merged_at": "2020-05-26T14:38:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/195.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/195...
true
https://api.github.com/repos/huggingface/datasets/issues/194
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/194/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/194/comments
https://api.github.com/repos/huggingface/datasets/issues/194/events
https://github.com/huggingface/datasets/pull/194
624,854,897
MDExOlB1bGxSZXF1ZXN0NDIzMTgyNDM5
194
Add Dataset: Qanta
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
3
2020-05-26 12:44:35+00:00
2020-05-26 16:58:17+00:00
2020-05-26 13:16:20+00:00
CONTRIBUTOR
null
null
null
null
Fixes dummy data for #169 @EntilZha
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/194/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/194/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/194.diff", "html_url": "https://github.com/huggingface/datasets/pull/194", "merged_at": "2020-05-26T13:16:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/194.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/194...
true
https://api.github.com/repos/huggingface/datasets/issues/193
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/193/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/193/comments
https://api.github.com/repos/huggingface/datasets/issues/193/events
https://github.com/huggingface/datasets/issues/193
624,655,558
MDU6SXNzdWU2MjQ2NTU1NTg=
193
[Tensorflow] Use something else than `from_tensor_slices()`
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
7
2020-05-26 07:19:14+00:00
2020-10-27 15:28:11+00:00
2020-10-27 15:28:11+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
In the example notebook, the TF Dataset is built using `from_tensor_slices()` : ```python columns = ['input_ids', 'token_type_ids', 'attention_mask', 'start_positions', 'end_positions'] train_tf_dataset.set_format(type='tensorflow', columns=columns) features = {x: train_tf_dataset[x] for x in columns[:3]} label...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/193/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/193/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/192
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/192/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/192/comments
https://api.github.com/repos/huggingface/datasets/issues/192/events
https://github.com/huggingface/datasets/issues/192
624,397,592
MDU6SXNzdWU2MjQzOTc1OTI=
192
[Question] Create Apache Arrow dataset from raw text file
{ "avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4", "events_url": "https://api.github.com/users/mrm8488/events{/privacy}", "followers_url": "https://api.github.com/users/mrm8488/followers", "following_url": "https://api.github.com/users/mrm8488/following{/other_user}", "gists_url": "https:/...
[]
closed
false
null
[]
null
4
2020-05-25 16:42:47+00:00
2021-12-18 01:45:34+00:00
2020-10-27 15:20:22+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi guys, I have gathered and preprocessed about 2GB of COVID papers from CORD dataset @ Kggle. I have seen you have a text dataset as "Crime and punishment" in Apache arrow format. Do you have any script to do it from a raw txt file (preprocessed as for BERT like) or any guide? Is the worth of send it to you and add i...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/192/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/192/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/191
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/191/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/191/comments
https://api.github.com/repos/huggingface/datasets/issues/191/events
https://github.com/huggingface/datasets/pull/191
624,394,936
MDExOlB1bGxSZXF1ZXN0NDIyODI3MDMy
191
[Squad es] add dataset_infos
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
0
2020-05-25 16:35:52+00:00
2020-05-25 16:39:59+00:00
2020-05-25 16:39:58+00:00
CONTRIBUTOR
null
null
null
null
@mariamabarham - was still about to upload this. Should have waited with my comment a bit more :D
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/191/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/191/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/191.diff", "html_url": "https://github.com/huggingface/datasets/pull/191", "merged_at": "2020-05-25T16:39:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/191.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/191...
true
https://api.github.com/repos/huggingface/datasets/issues/190
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/190/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/190/comments
https://api.github.com/repos/huggingface/datasets/issues/190/events
https://github.com/huggingface/datasets/pull/190
624,124,600
MDExOlB1bGxSZXF1ZXN0NDIyNjA4NzAw
190
add squad Spanish v1 and v2
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
5
2020-05-25 08:08:40+00:00
2020-05-25 16:28:46+00:00
2020-05-25 16:28:45+00:00
CONTRIBUTOR
null
null
null
null
This PR add the Spanish Squad versions 1 and 2 datasets. Fixes #164
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/190/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/190/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/190.diff", "html_url": "https://github.com/huggingface/datasets/pull/190", "merged_at": "2020-05-25T16:28:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/190.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/190...
true
https://api.github.com/repos/huggingface/datasets/issues/189
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/189/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/189/comments
https://api.github.com/repos/huggingface/datasets/issues/189/events
https://github.com/huggingface/datasets/issues/189
624,048,881
MDU6SXNzdWU2MjQwNDg4ODE=
189
[Question] BERT-style multiple choice formatting
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
2
2020-05-25 05:11:05+00:00
2020-05-25 18:38:28+00:00
2020-05-25 18:38:28+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hello, I am wondering what the equivalent formatting of a dataset should be to allow for multiple-choice answering prediction, BERT-style. Previously, this was done by passing a list of `InputFeatures` to the dataloader instead of a list of `InputFeature`, where `InputFeatures` contained lists of length equal to the nu...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/189/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/189/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/188
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/188/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/188/comments
https://api.github.com/repos/huggingface/datasets/issues/188/events
https://github.com/huggingface/datasets/issues/188
623,890,430
MDU6SXNzdWU2MjM4OTA0MzA=
188
When will the remaining math_dataset modules be added as dataset objects
{ "avatar_url": "https://avatars.githubusercontent.com/u/31251196?v=4", "events_url": "https://api.github.com/users/tylerroost/events{/privacy}", "followers_url": "https://api.github.com/users/tylerroost/followers", "following_url": "https://api.github.com/users/tylerroost/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
3
2020-05-24 15:46:52+00:00
2020-05-24 18:53:48+00:00
2020-05-24 18:53:48+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Currently only the algebra_linear_1d is supported. Is there a timeline for making the other modules supported. If no timeline is established, how can I help?
{ "avatar_url": "https://avatars.githubusercontent.com/u/31251196?v=4", "events_url": "https://api.github.com/users/tylerroost/events{/privacy}", "followers_url": "https://api.github.com/users/tylerroost/followers", "following_url": "https://api.github.com/users/tylerroost/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/188/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/188/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/187
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/187/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/187/comments
https://api.github.com/repos/huggingface/datasets/issues/187/events
https://github.com/huggingface/datasets/issues/187
623,627,800
MDU6SXNzdWU2MjM2Mjc4MDA=
187
[Question] How to load wikipedia ? Beam runner ?
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
null
[]
null
2
2020-05-23 10:18:52+00:00
2020-05-25 00:12:02+00:00
2020-05-25 00:12:02+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
When `nlp.load_dataset('wikipedia')`, I got * `WARNING:nlp.builder:Trying to generate a dataset using Apache Beam, yet no Beam Runner or PipelineOptions() has been provided. Please pass a nlp.DownloadConfig(beam_runner=...) object to the builder.download_and_prepare(download_config=...) method. Default values will be ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/187/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/187/timeline
null
completed
null
null
false