language:
- en
Model Card for passage-ranker-v1-XS-en
This model is a passage ranker developed by Sinequa. It produces a relevance score given a query-passage pair and is used to order search results.
Model name: passage-ranker-v1-XS-en
Supported Languages
The model was trained and tested in the following languages:
- English
Scores
Metric | Value |
---|---|
Relevance (NDCG@10) | 0.438 |
Note that the relevance score is computed as an average over 14 retrieval datasets (see details below).
Inference Times
GPU | Batch size 32 |
---|---|
NVIDIA A10 | 8 ms |
NVIDIA T4 | 20 ms |
The inference times only measure the time the model takes to process a single batch, it does not include pre- or post-processing steps like the tokenization.
Requirements
- Minimal Sinequa version: 11.10.0
- GPU memory usage: 170 MiB
Note that GPU memory usage only includes how much GPU memory the actual model consumes on an NVIDIA T4 GPU with a batch size of 32. It does not include the fix amount of memory that is consumed by the ONNX Runtime upon initialization which can be around 0.5 to 1 GiB depending on the used GPU.
Model Details
Overview
- Number of parameters: 11 million
- Base language model: English BERT-Mini
- Insensitive to casing and accents
- Training procedure: MonoBERT
Training Data
- MS MARCO Passage Ranking (Paper, Official Page, English dataset on the HF dataset hub)
Evaluation Metrics
To determine the relevance score, we averaged the results that we obtained when evaluating on the datasets of the BEIR benchmark. Note that all these datasets are in English.
Dataset | NDCG@10 |
---|---|
Average | 0.438 |
Arguana | 0.524 |
CLIMATE-FEVER | 0.150 |
DBPedia Entity | 0.338 |
FEVER | 0.706 |
FiQA-2018 | 0.269 |
HotpotQA | 0.630 |
MS MARCO | 0.328 |
NFCorpus | 0.340 |
NQ | 0.429 |
Quora | 0.722 |
SCIDOCS | 0.141 |
SciFact | 0.627 |
TREC-COVID | 0.628 |
Webis-Touche-2020 | 0.306 |