Initial model card

#2
by youval - opened
Files changed (1) hide show
  1. README.md +88 -0
README.md ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - feature-extraction
5
+ - sentence-similarity
6
+ language:
7
+ - en
8
+ ---
9
+
10
+ # Model Card for `vectorizer.vanilla`
11
+
12
+ This model is a vectorizer developed by Sinequa. It produces an embedding vector given a passage or a query. The
13
+ passage vectors are stored in our vector index and the query vector is used at query time to look up relevant passages
14
+ in the index.
15
+
16
+ Model name: `vectorizer.vanilla`
17
+
18
+ ## Supported Languages
19
+
20
+ The model was trained and tested in the following languages:
21
+
22
+ - English
23
+
24
+ ## Scores
25
+
26
+ | Metric | Value |
27
+ |:-----------------------|------:|
28
+ | Relevance (Recall@100) | 0.639 |
29
+
30
+ Note that the relevance score is computed as an average over 14 retrieval datasets (see
31
+ [details below](#evaluation-metrics)).
32
+
33
+ ## Inference Times
34
+
35
+ | GPU | Batch size 1 (at query time) | Batch size 32 (at indexing) |
36
+ |:-----------|-----------------------------:|----------------------------:|
37
+ | NVIDIA A10 | 2 ms | 19 ms |
38
+ | NVIDIA T4 | 4 ms | 53 ms |
39
+
40
+ The inference times only measure the time the model takes to process a single batch, it does not include pre- or
41
+ post-processing steps like the tokenization.
42
+
43
+ ## Requirements
44
+
45
+ - Minimal Sinequa version: 11.10.0
46
+ - GPU memory usage: 330 MiB
47
+
48
+ Note that GPU memory usage only includes how much GPU memory the actual model consumes on an NVIDIA T4 GPU with a batch
49
+ size of 32. It does not include the fix amount of memory that is consumed by the ONNX Runtime upon initialization which
50
+ can be around 0.5 to 1 GiB depending on the used GPU.
51
+
52
+ ## Model Details
53
+
54
+ ### Overview
55
+
56
+ - Number of parameters: 23 million
57
+ - Base language model: [English MiniLM-L6-H384](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased)
58
+ - Insensitive to casing and accents
59
+ - Output dimensions: 256 (reduced with an additional dense layer)
60
+ - Training procedure: Query-passage-negative triplets for datasets that have mined hard negative data, Query-passage pairs for the rest. Number of negatives is augmented with in-batch negative strategy.
61
+
62
+ ### Training Data
63
+
64
+ The model have been trained using all datasets that are cited in the [all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) model.
65
+
66
+ ### Evaluation Metrics
67
+
68
+ To determine the relevance score, we averaged the results that we obtained when evaluating on the datasets of the
69
+ [BEIR benchmark](https://github.com/beir-cellar/beir). Note that all these datasets are in English.
70
+
71
+ | Dataset | Recall@100 |
72
+ |:------------------|-----------:|
73
+ | Average | 0.639 |
74
+ | | |
75
+ | Arguana | 0.969 |
76
+ | CLIMATE-FEVER | 0.509 |
77
+ | DBPedia Entity | 0.409 |
78
+ | FEVER | 0.839 |
79
+ | FiQA-2018 | 0.702 |
80
+ | HotpotQA | 0.609 |
81
+ | MS MARCO | 0.849 |
82
+ | NFCorpus | 0.315 |
83
+ | NQ | 0.786 |
84
+ | Quora | 0.995 |
85
+ | SCIDOCS | 0.497 |
86
+ | SciFact | 0.911 |
87
+ | TREC-COVID | 0.129 |
88
+ | Webis-Touche-2020 | 0.427 |