robinsmits commited on
Commit
5afafd5
1 Parent(s): 9959dd7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -14
README.md CHANGED
@@ -127,6 +127,35 @@ DPO Finetuning was performed on the Dutch [BramVanroy/ultra_feedback_dutch_clean
127
  See [Qwen/Qwen1.5-7B-Chat](https://huggingface.co/Qwen/Qwen1.5-7B-Chat) for all information about the base model.
128
 
129
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
130
  ## Model usage
131
 
132
  A basic example of how to use the finetuned model.
@@ -231,7 +260,7 @@ The following hyperparameters were used during training:
231
  - Tokenizers 0.15.2
232
 
233
  ## Citation
234
- Thanks to the creators of Qwen1.5 for there great work!
235
  ```
236
  @article{qwen,
237
  title={Qwen Technical Report},
@@ -240,16 +269,3 @@ Thanks to the creators of Qwen1.5 for there great work!
240
  year={2023}
241
  }
242
  ```
243
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
244
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_robinsmits__Qwen1.5-7B-Dutch-Chat)
245
-
246
- | Metric |Value|
247
- |---------------------------------|----:|
248
- |Avg. |53.66|
249
- |AI2 Reasoning Challenge (25-Shot)|53.92|
250
- |HellaSwag (10-Shot) |76.03|
251
- |MMLU (5-Shot) |62.38|
252
- |TruthfulQA (0-shot) |45.34|
253
- |Winogrande (5-shot) |68.82|
254
- |GSM8k (5-shot) |15.47|
255
-
 
127
  See [Qwen/Qwen1.5-7B-Chat](https://huggingface.co/Qwen/Qwen1.5-7B-Chat) for all information about the base model.
128
 
129
 
130
+
131
+ ## ScandEval Dutch Leaderboard Evaluation Results
132
+
133
+ For evaluation results based on the Dutch language you can take a look at the site of ScandEval.
134
+
135
+ This model achieves a score which is very close to the performance of GPT-3.5.
136
+
137
+ [Dutch Natural Language Understanding](https://scandeval.com/dutch-nlu/)
138
+
139
+ [Dutch Natural Language Generation](https://scandeval.com/dutch-nlg/)
140
+
141
+
142
+
143
+ ## [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
144
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_robinsmits__Qwen1.5-7B-Dutch-Chat)
145
+
146
+ Note that these Evaluation Results are for the English language.
147
+
148
+ | Metric |Value|
149
+ |---------------------------------|----:|
150
+ |Avg. |53.66|
151
+ |AI2 Reasoning Challenge (25-Shot)|53.92|
152
+ |HellaSwag (10-Shot) |76.03|
153
+ |MMLU (5-Shot) |62.38|
154
+ |TruthfulQA (0-shot) |45.34|
155
+ |Winogrande (5-shot) |68.82|
156
+ |GSM8k (5-shot) |15.47|
157
+
158
+
159
  ## Model usage
160
 
161
  A basic example of how to use the finetuned model.
 
260
  - Tokenizers 0.15.2
261
 
262
  ## Citation
263
+ Thanks to the creators of Qwen1.5 for their great work!
264
  ```
265
  @article{qwen,
266
  title={Qwen Technical Report},
 
269
  year={2023}
270
  }
271
  ```