Datasets:
Update datasets task tags to align tags with models (#4067)
Browse files* update tasks list
* update tags in dataset cards
* more cards updates
* update dataset tags parser
* fix multi-choice-qa
* style
* small improvements in some dataset cards
* allow certain tag fields to be empty
* update vision datasets tags
* use multi-class-image-classification and remove other tags
Commit from https://github.com/huggingface/datasets/commit/edb4411d4e884690b8b328dba4360dbda6b3cbc8
README.md
CHANGED
@@ -14,10 +14,10 @@ size_categories:
|
|
14 |
source_datasets:
|
15 |
- original
|
16 |
task_categories:
|
17 |
-
-
|
18 |
task_ids:
|
19 |
- abstractive-qa
|
20 |
-
- open-domain-qa
|
21 |
paperswithcode_id: eli5
|
22 |
pretty_name: ELI5
|
23 |
---
|
@@ -61,7 +61,7 @@ The ELI5 dataset is an English-language dataset of questions and answers gathere
|
|
61 |
|
62 |
### Supported Tasks and Leaderboards
|
63 |
|
64 |
-
- `abstractive-qa`, `open-domain-qa`: The dataset can be used to train a model for Open Domain Long Form Question Answering. An LFQA model is presented with a non-factoid and asked to retrieve relevant information from a knowledge source (such as [Wikipedia](https://www.wikipedia.org/)), then use it to generate a multi-sentence answer. The model performance is measured by how high its [ROUGE](https://huggingface.co/metrics/rouge) score to the reference is. A [BART-based model](https://huggingface.co/yjernite/bart_eli5) with a [dense retriever](https://huggingface.co/yjernite/retribert-base-uncased) trained to draw information from [Wikipedia passages](https://huggingface.co/datasets/wiki_snippets) achieves a [ROUGE-L of 0.149](https://yjernite.github.io/lfqa.html#generation).
|
65 |
|
66 |
### Languages
|
67 |
|
|
|
14 |
source_datasets:
|
15 |
- original
|
16 |
task_categories:
|
17 |
+
- text2text-generation
|
18 |
task_ids:
|
19 |
- abstractive-qa
|
20 |
+
- open-domain-abstrative-qa
|
21 |
paperswithcode_id: eli5
|
22 |
pretty_name: ELI5
|
23 |
---
|
|
|
61 |
|
62 |
### Supported Tasks and Leaderboards
|
63 |
|
64 |
+
- `abstractive-qa`, `open-domain-abstractive-qa`: The dataset can be used to train a model for Open Domain Long Form Question Answering. An LFQA model is presented with a non-factoid and asked to retrieve relevant information from a knowledge source (such as [Wikipedia](https://www.wikipedia.org/)), then use it to generate a multi-sentence answer. The model performance is measured by how high its [ROUGE](https://huggingface.co/metrics/rouge) score to the reference is. A [BART-based model](https://huggingface.co/yjernite/bart_eli5) with a [dense retriever](https://huggingface.co/yjernite/retribert-base-uncased) trained to draw information from [Wikipedia passages](https://huggingface.co/datasets/wiki_snippets) achieves a [ROUGE-L of 0.149](https://yjernite.github.io/lfqa.html#generation).
|
65 |
|
66 |
### Languages
|
67 |
|