|
[1] — (1a) |
BM25 (k1=0.9, b=0.4) |
0.3013 |
0.5058 |
0.7501 |
|
0.2856 |
0.4796 |
0.7863 |
|
0.1840 |
0.8526 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-slim \
--output run.msmarco-v1-passage.bm25-default.dl19.txt \
--bm25 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-default.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-default.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-default.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-slim \
--output run.msmarco-v1-passage.bm25-default.dl20.txt \
--bm25 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-default.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-default.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-default.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-slim \
--output run.msmarco-v1-passage.bm25-default.dev.txt \
--bm25 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-default.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-default.dev.txt
|
|
[1] — (1b) |
BM25+RM3 (k1=0.9, b=0.4) |
0.3390 |
0.5180 |
0.7998 |
|
0.3019 |
0.4821 |
0.8217 |
|
0.1564 |
0.8606 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rm3-default.dl19.txt \
--bm25 --k1 0.9 --b 0.4 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-rm3-default.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-rm3-default.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-rm3-default.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rm3-default.dl20.txt \
--bm25 --k1 0.9 --b 0.4 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-rm3-default.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-rm3-default.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-rm3-default.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rm3-default.dev.txt \
--bm25 --k1 0.9 --b 0.4 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rm3-default.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rm3-default.dev.txt
|
|
|
BM25+Rocchio (k1=0.9, b=0.4) |
0.3474 |
0.5275 |
0.8007 |
|
0.3102 |
0.4893 |
0.8156 |
|
0.1597 |
0.8620 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rocchio-default.dl19.txt \
--bm25 --k1 0.9 --b 0.4 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-rocchio-default.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-rocchio-default.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-rocchio-default.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rocchio-default.dl20.txt \
--bm25 --k1 0.9 --b 0.4 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-rocchio-default.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-rocchio-default.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-rocchio-default.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rocchio-default.dev.txt \
--bm25 --k1 0.9 --b 0.4 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rocchio-default.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rocchio-default.dev.txt
|
|
|
|
BM25 (k1=0.82, b=0.68) |
0.2903 |
0.4973 |
0.7450 |
|
0.2876 |
0.4876 |
0.8031 |
|
0.1875 |
0.8573 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-slim \
--output run.msmarco-v1-passage.bm25-tuned.dl19.txt \
--bm25
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-tuned.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-slim \
--output run.msmarco-v1-passage.bm25-tuned.dl20.txt \
--bm25
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-tuned.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-slim \
--output run.msmarco-v1-passage.bm25-tuned.dev.txt \
--bm25
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-tuned.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-tuned.dev.txt
|
|
|
BM25+RM3 (k1=0.82, b=0.68) |
0.3377 |
0.5231 |
0.7792 |
|
0.3056 |
0.4808 |
0.8286 |
|
0.1668 |
0.8687 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rm3-tuned.dl19.txt \
--bm25 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-rm3-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-rm3-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-rm3-tuned.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rm3-tuned.dl20.txt \
--bm25 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-rm3-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-rm3-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-rm3-tuned.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rm3-tuned.dev.txt \
--bm25 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rm3-tuned.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rm3-tuned.dev.txt
|
|
|
BM25+Rocchio (k1=0.82, b=0.68) |
0.3394 |
0.5271 |
0.7969 |
|
0.3110 |
0.4901 |
0.8327 |
|
0.1685 |
0.8726 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rocchio-tuned.dl19.txt \
--bm25 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-rocchio-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-rocchio-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-rocchio-tuned.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rocchio-tuned.dl20.txt \
--bm25 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-rocchio-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-rocchio-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-rocchio-tuned.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-full \
--output run.msmarco-v1-passage.bm25-rocchio-tuned.dev.txt \
--bm25 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rocchio-tuned.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rocchio-tuned.dev.txt
|
|
|
[1] — (2a) |
BM25 w/ doc2query-T5 (k1=0.9, b=0.4) |
0.4034 |
0.6417 |
0.8310 |
|
0.4074 |
0.6187 |
0.8452 |
|
0.2723 |
0.9470 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-d2q-t5 \
--output run.msmarco-v1-passage.bm25-d2q-t5-default.dl19.txt \
--bm25 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-d2q-t5-default.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-d2q-t5-default.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-d2q-t5-default.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-d2q-t5 \
--output run.msmarco-v1-passage.bm25-d2q-t5-default.dl20.txt \
--bm25 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-d2q-t5-default.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-d2q-t5-default.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-d2q-t5-default.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-d2q-t5 \
--output run.msmarco-v1-passage.bm25-d2q-t5-default.dev.txt \
--bm25 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-d2q-t5-default.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-d2q-t5-default.dev.txt
|
|
[1] — (2b) |
BM25+RM3 w/ doc2query-T5 (k1=0.9, b=0.4) |
0.4485 |
0.6548 |
0.8861 |
|
0.4295 |
0.6172 |
0.8699 |
|
0.2140 |
0.9463 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dl19.txt \
--bm25 --rm3 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dl20.txt \
--bm25 --rm3 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dev.txt \
--bm25 --rm3 --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rm3-d2q-t5-default.dev.txt
|
|
|
BM25+Rocchio w/ doc2query-T5 (k1=0.9, b=0.4) |
0.4469 |
0.6538 |
0.8855 |
|
0.4246 |
0.6102 |
0.8675 |
|
0.2158 |
0.9467 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dl19.txt \
--bm25 --rocchio --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dl20.txt \
--bm25 --rocchio --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dev.txt \
--bm25 --rocchio --k1 0.9 --b 0.4
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rocchio-d2q-t5-default.dev.txt
|
|
|
|
BM25 w/ doc2query-T5 (k1=2.18, b=0.86) |
0.4046 |
0.6336 |
0.8134 |
|
0.4171 |
0.6265 |
0.8393 |
|
0.2816 |
0.9506 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-d2q-t5 \
--output run.msmarco-v1-passage.bm25-d2q-t5-tuned.dl19.txt \
--bm25
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-d2q-t5-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-d2q-t5-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-d2q-t5-tuned.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-d2q-t5 \
--output run.msmarco-v1-passage.bm25-d2q-t5-tuned.dl20.txt \
--bm25
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-d2q-t5-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-d2q-t5-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-d2q-t5-tuned.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-d2q-t5 \
--output run.msmarco-v1-passage.bm25-d2q-t5-tuned.dev.txt \
--bm25
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-d2q-t5-tuned.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-d2q-t5-tuned.dev.txt
|
|
|
BM25+RM3 w/ doc2query-T5 (k1=2.18, b=0.86) |
0.4360 |
0.6528 |
0.8424 |
|
0.4347 |
0.6232 |
0.8609 |
|
0.2374 |
0.9528 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dl19.txt \
--bm25 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dl20.txt \
--bm25 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dev.txt \
--bm25 --rm3
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rm3-d2q-t5-tuned.dev.txt
|
|
|
BM25+Rocchio w/ doc2query-T5 (k1=2.18, b=0.86) |
0.4339 |
0.6559 |
0.8465 |
|
0.4376 |
0.6224 |
0.8641 |
|
0.2396 |
0.9535 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--topics dl19-passage \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dl19.txt \
--bm25 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--topics dl20 \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dl20.txt \
--bm25 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--topics msmarco-passage-dev-subset \
--index msmarco-v1-passage-d2q-t5-docvectors \
--output run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dev.txt \
--bm25 --rocchio
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.bm25-rocchio-d2q-t5-tuned.dev.txt
|
|
|
[1] — (3a) |
uniCOIL (noexp): pre-encoded queries |
0.4033 |
0.6433 |
0.7752 |
|
0.4021 |
0.6523 |
0.7861 |
|
0.3153 |
0.9239 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil-noexp \
--topics dl19-passage-unicoil-noexp \
--output run.msmarco-v1-passage.unicoil-noexp.dl19.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.unicoil-noexp.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.unicoil-noexp.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.unicoil-noexp.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil-noexp \
--topics dl20-unicoil-noexp \
--output run.msmarco-v1-passage.unicoil-noexp.dl20.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.unicoil-noexp.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.unicoil-noexp.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.unicoil-noexp.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil-noexp \
--topics msmarco-passage-dev-subset-unicoil-noexp \
--output run.msmarco-v1-passage.unicoil-noexp.dev.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.unicoil-noexp.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.unicoil-noexp.dev.txt
|
|
[1] — (3b) |
uniCOIL (w/ doc2query-T5): pre-encoded queries |
0.4612 |
0.7024 |
0.8292 |
|
0.4430 |
0.6745 |
0.8430 |
|
0.3516 |
0.9582 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil \
--topics dl19-passage-unicoil \
--output run.msmarco-v1-passage.unicoil.dl19.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.unicoil.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.unicoil.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.unicoil.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil \
--topics dl20-unicoil \
--output run.msmarco-v1-passage.unicoil.dl20.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.unicoil.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.unicoil.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.unicoil.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil \
--topics msmarco-passage-dev-subset-unicoil \
--output run.msmarco-v1-passage.unicoil.dev.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.unicoil.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.unicoil.dev.txt
|
|
|
|
uniCOIL (noexp): on-the-fly query inference |
0.4033 |
0.6434 |
0.7752 |
|
0.4022 |
0.6524 |
0.7861 |
|
0.3153 |
0.9239 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil-noexp \
--topics dl19-passage --encoder castorini/unicoil-noexp-msmarco-passage \
--output run.msmarco-v1-passage.unicoil-noexp-otf.dl19.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.unicoil-noexp-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.unicoil-noexp-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.unicoil-noexp-otf.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil-noexp \
--topics dl20 --encoder castorini/unicoil-noexp-msmarco-passage \
--output run.msmarco-v1-passage.unicoil-noexp-otf.dl20.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.unicoil-noexp-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.unicoil-noexp-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.unicoil-noexp-otf.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil-noexp \
--topics msmarco-passage-dev-subset --encoder castorini/unicoil-noexp-msmarco-passage \
--output run.msmarco-v1-passage.unicoil-noexp-otf.dev.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.unicoil-noexp-otf.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.unicoil-noexp-otf.dev.txt
|
|
|
uniCOIL (w/ doc2query-T5): on-the-fly query inference |
0.4617 |
0.7027 |
0.8291 |
|
0.4429 |
0.6745 |
0.8433 |
|
0.3509 |
0.9581 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil \
--topics dl19-passage --encoder castorini/unicoil-msmarco-passage \
--output run.msmarco-v1-passage.unicoil-otf.dl19.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.unicoil-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.unicoil-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.unicoil-otf.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil \
--topics dl20 --encoder castorini/unicoil-msmarco-passage \
--output run.msmarco-v1-passage.unicoil-otf.dl20.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.unicoil-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.unicoil-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.unicoil-otf.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.lucene \
--index msmarco-v1-passage-unicoil \
--topics msmarco-passage-dev-subset --encoder castorini/unicoil-msmarco-passage \
--output run.msmarco-v1-passage.unicoil-otf.dev.txt \
--batch 36 --threads 12 --hits 1000 --impact
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.unicoil-otf.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.unicoil-otf.dev.txt
|
|
|
[2] |
ANCE: pre-encoded queries |
- |
- |
- |
|
- |
- |
- |
|
0.3302 |
0.9584 |
|
Not available.
Not available.
Command to generate run on dev queries:
python -m pyserini.search.faiss \
--index msmarco-passage-ance-bf \
--topics msmarco-passage-dev-subset --encoded-queries ance-msmarco-passage-dev-subset \
--output run.msmarco-v1-passage.ance.dev.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.ance.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.ance.dev.txt
|
|
[3] |
DistilBERT KD: pre-encoded queries |
- |
- |
- |
|
- |
- |
- |
|
0.3251 |
0.9553 |
|
Not available.
Not available.
Command to generate run on dev queries:
python -m pyserini.search.faiss \
--index msmarco-passage-distilbert-dot-margin_mse-T2-bf \
--topics msmarco-passage-dev-subset --encoded-queries distilbert_kd-msmarco-passage-dev-subset \
--output run.msmarco-v1-passage.distilbert-kd.dev.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.distilbert-kd.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.distilbert-kd.dev.txt
|
|
[4] |
DistilBERT KD TASB: pre-encoded queries |
- |
- |
- |
|
- |
- |
- |
|
0.3444 |
0.9771 |
|
Not available.
Not available.
Command to generate run on dev queries:
python -m pyserini.search.faiss \
--index msmarco-passage-distilbert-dot-tas_b-b256-bf \
--topics msmarco-passage-dev-subset --encoded-queries distilbert_tas_b-msmarco-passage-dev-subset \
--output run.msmarco-v1-passage.distilbert-kd-tasb.dev.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.distilbert-kd-tasb.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.distilbert-kd-tasb.dev.txt
|
|
[5] |
TCT_ColBERT-V2-HN+: pre-encoded queries |
- |
- |
- |
|
- |
- |
- |
|
0.3584 |
0.9695 |
|
Not available.
Not available.
Command to generate run on dev queries:
python -m pyserini.search.faiss \
--index msmarco-passage-tct_colbert-v2-hnp-bf \
--topics msmarco-passage-dev-subset --encoded-queries tct_colbert-v2-hnp-msmarco-passage-dev-subset \
--output run.msmarco-v1-passage.tct_colbert-v2-hnp.dev.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.tct_colbert-v2-hnp.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.tct_colbert-v2-hnp.dev.txt
|
|
|
[2] |
ANCE: on-the-fly query inference |
0.3710 |
0.6452 |
0.7554 |
|
0.4076 |
0.6458 |
0.7764 |
|
0.3302 |
0.9587 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.faiss \
--index msmarco-passage-ance-bf \
--topics dl19-passage --encoder castorini/ance-msmarco-passage \
--output run.msmarco-v1-passage.ance-otf.dl19.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.ance-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.ance-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.ance-otf.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.faiss \
--index msmarco-passage-ance-bf \
--topics dl20 --encoder castorini/ance-msmarco-passage \
--output run.msmarco-v1-passage.ance-otf.dl20.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.ance-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.ance-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.ance-otf.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.faiss \
--index msmarco-passage-ance-bf \
--topics msmarco-passage-dev-subset --encoder castorini/ance-msmarco-passage \
--output run.msmarco-v1-passage.ance-otf.dev.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.ance-otf.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.ance-otf.dev.txt
|
|
[3] |
DistilBERT KD: on-the-fly query inference |
0.4053 |
0.6994 |
0.7653 |
|
0.4159 |
0.6447 |
0.7953 |
|
0.3251 |
0.9553 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.faiss \
--index msmarco-passage-distilbert-dot-margin_mse-T2-bf \
--topics dl19-passage --encoder sebastian-hofstaetter/distilbert-dot-margin_mse-T2-msmarco \
--output run.msmarco-v1-passage.distilbert-kd-otf.dl19.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.distilbert-kd-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.distilbert-kd-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.distilbert-kd-otf.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.faiss \
--index msmarco-passage-distilbert-dot-margin_mse-T2-bf \
--topics dl20 --encoder sebastian-hofstaetter/distilbert-dot-margin_mse-T2-msmarco \
--output run.msmarco-v1-passage.distilbert-kd-otf.dl20.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.distilbert-kd-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.distilbert-kd-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.distilbert-kd-otf.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.faiss \
--index msmarco-passage-distilbert-dot-margin_mse-T2-bf \
--topics msmarco-passage-dev-subset --encoder sebastian-hofstaetter/distilbert-dot-margin_mse-T2-msmarco \
--output run.msmarco-v1-passage.distilbert-kd-otf.dev.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.distilbert-kd-otf.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.distilbert-kd-otf.dev.txt
|
|
[4] |
DistilBERT KD TASB: on-the-fly query inference |
0.4590 |
0.7210 |
0.8406 |
|
0.4698 |
0.6854 |
0.8727 |
|
0.3444 |
0.9771 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.faiss \
--index msmarco-passage-distilbert-dot-tas_b-b256-bf \
--topics dl19-passage --encoder sebastian-hofstaetter/distilbert-dot-tas_b-b256-msmarco \
--output run.msmarco-v1-passage.distilbert-kd-tasb-otf.dl19.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.distilbert-kd-tasb-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.distilbert-kd-tasb-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.distilbert-kd-tasb-otf.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.faiss \
--index msmarco-passage-distilbert-dot-tas_b-b256-bf \
--topics dl20 --encoder sebastian-hofstaetter/distilbert-dot-tas_b-b256-msmarco \
--output run.msmarco-v1-passage.distilbert-kd-tasb-otf.dl20.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.distilbert-kd-tasb-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.distilbert-kd-tasb-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.distilbert-kd-tasb-otf.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.faiss \
--index msmarco-passage-distilbert-dot-tas_b-b256-bf \
--topics msmarco-passage-dev-subset --encoder sebastian-hofstaetter/distilbert-dot-tas_b-b256-msmarco \
--output run.msmarco-v1-passage.distilbert-kd-tasb-otf.dev.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.distilbert-kd-tasb-otf.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.distilbert-kd-tasb-otf.dev.txt
|
|
[5] |
TCT_ColBERT-V2-HN+: on-the-fly query inference |
0.4469 |
0.7204 |
0.8261 |
|
0.4754 |
0.6882 |
0.8429 |
|
0.3584 |
0.9695 |
|
Command to generate run on TREC 2019 queries:
python -m pyserini.search.faiss \
--index msmarco-passage-tct_colbert-v2-hnp-bf \
--topics dl19-passage --encoder castorini/tct_colbert-v2-hnp-msmarco \
--output run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dl19.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl19-passage run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl19-passage run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dl19.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl19-passage run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dl19.txt
Command to generate run on TREC 2020 queries:
python -m pyserini.search.faiss \
--index msmarco-passage-tct_colbert-v2-hnp-bf \
--topics dl20 --encoder castorini/tct_colbert-v2-hnp-msmarco \
--output run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dl20.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -l 2 -m map dl20-passage run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -m ndcg_cut.10 dl20-passage run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dl20.txt
python -m pyserini.eval.trec_eval -c -l 2 -m recall.1000 dl20-passage run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dl20.txt
Command to generate run on dev queries:
python -m pyserini.search.faiss \
--index msmarco-passage-tct_colbert-v2-hnp-bf \
--topics msmarco-passage-dev-subset --encoder castorini/tct_colbert-v2-hnp-msmarco \
--output run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dev.txt \
--batch-size 36 --threads 12
Evaluation commands:
python -m pyserini.eval.trec_eval -c -M 10 -m recip_rank msmarco-passage-dev-subset run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dev.txt
python -m pyserini.eval.trec_eval -c -m recall.1000 msmarco-passage-dev-subset run.msmarco-v1-passage.tct_colbert-v2-hnp-otf.dev.txt
|