haystack/docs/_src/benchmarks/reader_performance.json
Malte Pietsch 7e79a48540
bug: reactivate benchmarks with quick fixes (#2766)
* quick fix benchmark runs to make them work with current haystack version

* fix minor typo

* update readme. fix minor things to make benchmarks run again

* Update Documentation & Code Style

* fix typo in readme

* update result files for reader and retriever querying

* reduce batch size for update embeddings to prevent xlarge bulk_update requests that exceed elastic's limits (happening in dense 500k runs)

* change default memory allocation back to normal. add note to readme

* add first indexing results

* add memory to docker cmd

* full benchmarks results on commit  c5a2651fcbbeffca06ffa9036b10e62669bcc1b0

Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
2022-09-20 10:22:08 +02:00

44 lines
2.0 KiB
JSON

{
"chart_type": "BarChart",
"title": "Reader Performance",
"subtitle": "Time and Accuracy Benchmarks",
"description": "Performance benchmarks of different Readers that can be used off-the-shelf in Haystack. Some models are geared towards speed, while others are more performance-focused. Accuracy is measured as F1 score and speed as passages/sec (with passages of 384 tokens). Each Reader is benchmarked using the SQuAD v2.0 development set, which contains 11866 question answer pairs. When tokenized using the BERT tokenizer and split using a sliding window approach, these become 12350 passages that are passed into the model. We set <i>max_seq_len=384</i> and <i>doc_stride=128</i>. These benchmarking tests are run using an AWS p3.2xlarge instance with a Nvidia V100 GPU with this <a href='https://github.com/deepset-ai/haystack/blob/master/test/benchmarks/reader.py'>script</a>. Please note that we are using the FARMReader class rather than the TransformersReader class. Also, the F1 measure that is reported here is in fact calculated on token level, rather than word level as is done in the official SQuAD script.",
"bars": "horizontal",
"columns": [
"Model",
"F1",
"Speed (passages/sec)"
],
"data": [
{
"F1": 78.21542223321491,
"Speed": 118.41880460779026,
"Model": "RoBERTa"
},
{
"F1": 79.6479775678566,
"Speed": 235.0263836060577,
"Model": "MiniLM"
},
{
"F1": 73.49806294434092,
"Speed": 116.99833085428466,
"Model": "BERT base"
},
{
"F1": 87.35038229973254,
"Speed": 43.45399164221017,
"Model": "RoBERTa large"
},
{
"F1": 82.29388851439239,
"Speed": 118.88131118898794,
"Model": "RoBERTa distilled"
},
{
"F1": 43.34028303155876,
"Speed": 198.4500591997352,
"Model": "distilBERT"
}
]
}