mirror of
https://github.com/deepset-ai/haystack.git
synced 2025-12-31 09:10:15 +00:00
* quick fix benchmark runs to make them work with current haystack version * fix minor typo * update readme. fix minor things to make benchmarks run again * Update Documentation & Code Style * fix typo in readme * update result files for reader and retriever querying * reduce batch size for update embeddings to prevent xlarge bulk_update requests that exceed elastic's limits (happening in dense 500k runs) * change default memory allocation back to normal. add note to readme * add first indexing results * add memory to docker cmd * full benchmarks results on commit c5a2651fcbbeffca06ffa9036b10e62669bcc1b0 Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
75 lines
1.5 KiB
JSON
75 lines
1.5 KiB
JSON
{
|
|
"params": {
|
|
"full": {
|
|
"retriever_doc_stores": [
|
|
[
|
|
"elastic",
|
|
"elasticsearch"
|
|
],
|
|
[
|
|
"elastic",
|
|
"opensearch_flat"
|
|
],
|
|
[
|
|
"dpr",
|
|
"opensearch_flat"
|
|
],
|
|
[
|
|
"dpr",
|
|
"opensearch_hnsw"
|
|
],
|
|
[
|
|
"dpr",
|
|
"elasticsearch"
|
|
],
|
|
[
|
|
"dpr",
|
|
"faiss_flat"
|
|
],
|
|
[
|
|
"dpr",
|
|
"faiss_hnsw"
|
|
],
|
|
[
|
|
"sentence_transformers",
|
|
"elasticsearch"
|
|
]
|
|
],
|
|
"n_docs_options": [
|
|
1000,
|
|
10000,
|
|
100000,
|
|
500000
|
|
],
|
|
"n_queries": null
|
|
},
|
|
"ci": {
|
|
"retriever_doc_stores": [
|
|
[
|
|
"elastic",
|
|
"elasticsearch"
|
|
]
|
|
],
|
|
"n_docs_options": [
|
|
1000
|
|
],
|
|
"n_queries": 100
|
|
}
|
|
},
|
|
"filenames": {
|
|
"data_s3_url": "https://ext-haystack-retriever-eval.s3-eu-west-1.amazonaws.com/",
|
|
"data_dir": "../../data/retriever/",
|
|
"filename_gold": "nq2squad-dev.json",
|
|
"filenames_negative": {
|
|
"10000": "psgs_w100_minus_gold_10k.tsv",
|
|
"100000": "psgs_w100_minus_gold_100k.tsv",
|
|
"1000000": "psgs_w100_minus_gold_1m.tsv"
|
|
},
|
|
"embeddings_dir": "embeddings/",
|
|
"embeddings_filenames": {
|
|
"10000": "wikipedia_passages_10k.pkl",
|
|
"100000": "wikipedia_passages_100k.pkl",
|
|
"1000000": "wikipedia_passages_1m.pkl"}
|
|
}
|
|
}
|