waveletdeboshir
commited on
Commit
•
c21debc
1
Parent(s):
bb8b24a
Update numbers
Browse files
README.md
CHANGED
@@ -22,15 +22,15 @@ This is a pruned version of [openai/whisper-small](https://huggingface.co/openai
|
|
22 |
Pruning was made without any fine-tuning. Method from [this post](https://medium.com/m/global-identity-2?redirectUrl=https%3A%2F%2Ftowardsdatascience.com%2Fhow-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90) was used.
|
23 |
|
24 |
## Size
|
25 |
-
Only 10% tokens was left including special whisper tokens
|
26 |
|
27 |
Model size is 15% less then original whisper-small:
|
28 |
| | openai/whisper-small | waveletdeboshir/whisper-small-ru-pruned |
|
29 |
| :------ | :------ | :------ |
|
30 |
| n of parameters | 242 M | 205 M |
|
31 |
-
| n of parameters (with proj_out layer) | 281 M |
|
32 |
-
| model file size | 967 Mb |
|
33 |
-
| vocab_size | 51865 |
|
34 |
|
35 |
## Usage
|
36 |
Model can be used as an original whisper:
|
|
|
22 |
Pruning was made without any fine-tuning. Method from [this post](https://medium.com/m/global-identity-2?redirectUrl=https%3A%2F%2Ftowardsdatascience.com%2Fhow-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90) was used.
|
23 |
|
24 |
## Size
|
25 |
+
Only 10% tokens was left including special whisper tokens (no language tokens except \<|ru|\> and \<|en|\>, no timestamp tokens), 200 most popular tokens from tokenizer and 4000 most popular Russian tokens computed by tokenization of russian text corpus.
|
26 |
|
27 |
Model size is 15% less then original whisper-small:
|
28 |
| | openai/whisper-small | waveletdeboshir/whisper-small-ru-pruned |
|
29 |
| :------ | :------ | :------ |
|
30 |
| n of parameters | 242 M | 205 M |
|
31 |
+
| n of parameters (with proj_out layer) | 281 M | 208 M |
|
32 |
+
| model file size | 967 Mb | 834 Mb |
|
33 |
+
| vocab_size | 51865 | 4207 |
|
34 |
|
35 |
## Usage
|
36 |
Model can be used as an original whisper:
|