mzboito commited on
Commit
afb4a28
1 Parent(s): de73e7f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md CHANGED
@@ -1,3 +1,36 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ datasets:
4
+ - mozilla-foundation/common_voice_13_0
5
+ language:
6
+ - ca
7
+ - bg
8
+ - cs
9
+ - fi
10
+ - gl
11
+ - hi
12
+ - hu
13
+ - pl
14
+ - ro
15
+ - sk
16
+ - ta
17
  ---
18
+
19
+ ## About
20
+
21
+ Multilingual Distilwhisper allows for better inference in target languages by adding lightweight CLSR modules on top of whisper-small.
22
+ These modules are trained on a mix of cross-entropy (ASR) and knowledge distillation losses, where whisper-large-v2 is used as teacher.
23
+
24
+ ## Inference
25
+
26
+ Loader will be made available soon at https://github.com/naver
27
+
28
+ ## Citation (submitted to ICASSP 2024)
29
+ ```
30
+ @article{ferraz2023distilwhisper,
31
+ title={DistilWhisper: Efficient Distillation of Multi-task Speech Models via Language-Specific Experts},
32
+ author={Ferraz, Thomas Palmeira and Boito, Marcely Zanon and Brun, Caroline and Nikoulina, Vassilina},
33
+ journal={arXiv preprint arXiv:2311.01070},
34
+ year={2023}
35
+ }
36
+ ```