ir_datasets
: MSMARCO (document)"Based the questions in the [MS-MARCO] Question Answering Dataset and the documents which answered the questions a document ranking task was formulated. There are 3.2 million documents and the goal is to rank based on their relevance. Relevance labels are derived from what passages was marked as having the answer in the QnA dataset."
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Official dev set. All queries have exactly 1 (positive) relevance judgment.
scoreddocs are the top 100 results from Indri QL. These are used for the "re-ranking" setting.
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/dev')
for query in dataset.queries_iter():
query # namedtuple<query_id, text>
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/dev')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Relevance levels
Rel. | Definition |
---|---|
1 | Labeled by crowd worker as relevant |
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/dev')
for qrel in dataset.qrels_iter():
qrel # namedtuple<query_id, doc_id, relevance, iteration>
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/dev')
for scoreddoc in dataset.scoreddocs_iter():
scoreddoc # namedtuple<query_id, doc_id, score>
Official eval set for submission to MS MARCO leaderboard. Relevance judgments are hidden.
scoreddocs are the top 100 results from Indri QL. These are used for the "re-ranking" setting.
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/eval')
for query in dataset.queries_iter():
query # namedtuple<query_id, text>
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/eval')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/eval')
for scoreddoc in dataset.scoreddocs_iter():
scoreddoc # namedtuple<query_id, doc_id, score>
"ORCAS is a click-based dataset associated with the TREC Deep Learning Track. It covers 1.4 million of the TREC DL documents, providing 18 million connections to 10 million distinct queries."
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/orcas')
for query in dataset.queries_iter():
query # namedtuple<query_id, text>
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/orcas')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Relevance levels
Rel. | Definition |
---|---|
1 | User click |
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/orcas')
for qrel in dataset.qrels_iter():
qrel # namedtuple<query_id, doc_id, relevance, iteration>
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/orcas')
for scoreddoc in dataset.scoreddocs_iter():
scoreddoc # namedtuple<query_id, doc_id, score>
Official train set. All queries have exactly 1 (positive) relevance judgment.
scoreddocs are the top 100 results from Indri QL. These are used for the "re-ranking" setting.
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/train')
for query in dataset.queries_iter():
query # namedtuple<query_id, text>
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/train')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Relevance levels
Rel. | Definition |
---|---|
1 | Labeled by crowd worker as relevant |
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/train')
for qrel in dataset.qrels_iter():
qrel # namedtuple<query_id, doc_id, relevance, iteration>
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/train')
for scoreddoc in dataset.scoreddocs_iter():
scoreddoc # namedtuple<query_id, doc_id, score>
Queries from the TREC Deep Learning (DL) 2019 shared task, which were sampled from msmarco-document/eval. A subset of these queries were judged by NIST assessors, (filtered list available in msmarco-document/trec-dl-2019/judged).
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2019')
for query in dataset.queries_iter():
query # namedtuple<query_id, text>
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2019')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Relevance levels
Rel. | Definition |
---|---|
0 | Irrelevant: Document does not provide any useful information about the query |
1 | Relevant: Document provides some information relevant to the query, which may be minimal. |
2 | Highly relevant: The content of this document provides substantial information on the query. |
3 | Perfectly relevant: Document is dedicated to the query, it is worthy of being a top result in a search engine. |
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2019')
for qrel in dataset.qrels_iter():
qrel # namedtuple<query_id, doc_id, relevance, iteration>
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2019')
for scoreddoc in dataset.scoreddocs_iter():
scoreddoc # namedtuple<query_id, doc_id, score>
Subset of msmarco-document/trec-dl-2019, only including queries with qrels.
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2019/judged')
for query in dataset.queries_iter():
query # namedtuple<query_id, text>
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2019/judged')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Relevance levels
Rel. | Definition |
---|---|
0 | Irrelevant: Document does not provide any useful information about the query |
1 | Relevant: Document provides some information relevant to the query, which may be minimal. |
2 | Highly relevant: The content of this document provides substantial information on the query. |
3 | Perfectly relevant: Document is dedicated to the query, it is worthy of being a top result in a search engine. |
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2019/judged')
for qrel in dataset.qrels_iter():
qrel # namedtuple<query_id, doc_id, relevance, iteration>
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2019/judged')
for scoreddoc in dataset.scoreddocs_iter():
scoreddoc # namedtuple<query_id, doc_id, score>
Queries from the TREC Deep Learning (DL) 2020 shared task, which were sampled from msmarco-document/eval. A subset of these queries were judged by NIST assessors, (filtered list available in msmarco-document/trec-dl-2020/judged).
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2020')
for query in dataset.queries_iter():
query # namedtuple<query_id, text>
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2020')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Relevance levels
Rel. | Definition |
---|---|
0 | Irrelevant: Document does not provide any useful information about the query |
1 | Relevant: Document provides some information relevant to the query, which may be minimal. |
2 | Highly relevant: The content of this document provides substantial information on the query. |
3 | Perfectly relevant: Document is dedicated to the query, it is worthy of being a top result in a search engine. |
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2020')
for qrel in dataset.qrels_iter():
qrel # namedtuple<query_id, doc_id, relevance, iteration>
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2020')
for scoreddoc in dataset.scoreddocs_iter():
scoreddoc # namedtuple<query_id, doc_id, score>
Subset of msmarco-document/trec-dl-2020, only including queries with qrels.
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2020/judged')
for query in dataset.queries_iter():
query # namedtuple<query_id, text>
Language: en
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2020/judged')
for doc in dataset.docs_iter():
doc # namedtuple<doc_id, url, title, body>
Relevance levels
Rel. | Definition |
---|---|
0 | Irrelevant: Document does not provide any useful information about the query |
1 | Relevant: Document provides some information relevant to the query, which may be minimal. |
2 | Highly relevant: The content of this document provides substantial information on the query. |
3 | Perfectly relevant: Document is dedicated to the query, it is worthy of being a top result in a search engine. |
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2020/judged')
for qrel in dataset.qrels_iter():
qrel # namedtuple<query_id, doc_id, relevance, iteration>
Example
import ir_datasets
dataset = ir_datasets.load('msmarco-document/trec-dl-2020/judged')
for scoreddoc in dataset.scoreddocs_iter():
scoreddoc # namedtuple<query_id, doc_id, score>