Attend the Women in ML Symposium on December 7 Register now

multi_re_qa

References:

SearchQA

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/SearchQA')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'train' 3163801
'validation' 454836
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}

TriviaQA

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/TriviaQA')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'train' 1893674
'validation' 238339
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}

HotpotQA

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/HotpotQA')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'train' 508879
'validation' 52191
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}

SQuAD

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/SQuAD')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'train' 95659
'validation' 10642
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}

NaturalQuestions

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/NaturalQuestions')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'train' 448355
'validation' 22118
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}

BioASQ

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/BioASQ')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'test' 14158
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}

RelationExtraction

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/RelationExtraction')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'test' 3301
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}

TextbookQA

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/TextbookQA')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'test' 71147
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}

DuoRC

Use the following command to load this dataset in TFDS:

ds = tfds.load('huggingface:multi_re_qa/DuoRC')
  • Description:
MultiReQA contains the sentence boundary annotation from eight publicly available QA datasets including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, BioASQ, RelationExtraction, and TextbookQA. Five of these datasets, including SearchQA, TriviaQA, HotpotQA, NaturalQuestions, SQuAD, contain both training and test data, and three, including BioASQ, RelationExtraction, TextbookQA, contain only the test data
  • License: No known license
  • Version: 1.0.0
  • Splits:
Split Examples
'test' 5525
  • Features:
{
    "candidate_id": {
        "dtype": "string",
        "id": null,
        "_type": "Value"
    },
    "response_start": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    },
    "response_end": {
        "dtype": "int32",
        "id": null,
        "_type": "Value"
    }
}