Spaces:
Running
Running
jhj0517
commited on
Commit
•
556fa8d
1
Parent(s):
4589c43
Update README.md
Browse files
README.md
CHANGED
@@ -14,8 +14,10 @@ If you wish to try this on Colab, you can do it in [here](https://colab.research
|
|
14 |
- Currently supported subtitle formats :
|
15 |
- SRT
|
16 |
- WebVTT
|
17 |
-
- Speech to Text Translation
|
18 |
-
- From other languages to English.
|
|
|
|
|
19 |
|
20 |
# Installation and Running
|
21 |
## Prerequisite
|
@@ -39,9 +41,19 @@ If you have satisfied the prerequisites listed above, you are now ready to start
|
|
39 |
|
40 |
And you can also run the project with command line arguments if you like by running `user-start-webui.bat`, see [wiki](https://github.com/jhj0517/Whisper-WebUI/wiki/Command-Line-Arguments) for a guide to arguments.
|
41 |
|
42 |
-
|
|
|
43 |
|
44 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
|
46 |
| Size | Parameters | English-only model | Multilingual model | Required VRAM | Relative speed |
|
47 |
|:------:|:----------:|:------------------:|:------------------:|:-------------:|:--------------:|
|
|
|
14 |
- Currently supported subtitle formats :
|
15 |
- SRT
|
16 |
- WebVTT
|
17 |
+
- Speech to Text Translation
|
18 |
+
- From other languages to English. ( This is Whisper's end-to-end speech-to-text translation feature )
|
19 |
+
- Text to Text Translation
|
20 |
+
- Translate subtitle files using Facebook NLLB models
|
21 |
|
22 |
# Installation and Running
|
23 |
## Prerequisite
|
|
|
41 |
|
42 |
And you can also run the project with command line arguments if you like by running `user-start-webui.bat`, see [wiki](https://github.com/jhj0517/Whisper-WebUI/wiki/Command-Line-Arguments) for a guide to arguments.
|
43 |
|
44 |
+
## VRAM Usages
|
45 |
+
This project is integrated with [faster-whisper](https://github.com/guillaumekln/faster-whisper) by default for better VRAM usage and transcription speed.
|
46 |
|
47 |
+
According to faster-whisper, the efficiency of the optimized whisper model is as follows:
|
48 |
+
| Implementation | Precision | Beam size | Time | Max. GPU memory | Max. CPU memory |
|
49 |
+
|-------------------|-----------|-----------|-------|-----------------|-----------------|
|
50 |
+
| openai/whisper | fp16 | 5 | 4m30s | 11325MB | 9439MB |
|
51 |
+
| faster-whisper | fp16 | 5 | 54s | 4755MB | 3244MB |
|
52 |
+
|
53 |
+
If you want to use the original Open AI whisper implementation instead of optimized whisper, you can set the command line argument `DISABLE_FASTER_WHISPER` to `True`. See the [wiki](https://github.com/jhj0517/Whisper-WebUI/wiki/Command-Line-Arguments) for more information.
|
54 |
+
|
55 |
+
## Available models
|
56 |
+
This is Whisper's original VRAM usage table for models.
|
57 |
|
58 |
| Size | Parameters | English-only model | Multilingual model | Required VRAM | Relative speed |
|
59 |
|:------:|:----------:|:------------------:|:------------------:|:-------------:|:--------------:|
|