jhj0517 commited on
Commit
556fa8d
1 Parent(s): 4589c43

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -4
README.md CHANGED
@@ -14,8 +14,10 @@ If you wish to try this on Colab, you can do it in [here](https://colab.research
14
  - Currently supported subtitle formats :
15
  - SRT
16
  - WebVTT
17
- - Speech to Text Translation
18
- - From other languages to English.
 
 
19
 
20
  # Installation and Running
21
  ## Prerequisite
@@ -39,9 +41,19 @@ If you have satisfied the prerequisites listed above, you are now ready to start
39
 
40
  And you can also run the project with command line arguments if you like by running `user-start-webui.bat`, see [wiki](https://github.com/jhj0517/Whisper-WebUI/wiki/Command-Line-Arguments) for a guide to arguments.
41
 
42
- # Available models
 
43
 
44
- The WebUI uses the Open AI Whisper model
 
 
 
 
 
 
 
 
 
45
 
46
  | Size | Parameters | English-only model | Multilingual model | Required VRAM | Relative speed |
47
  |:------:|:----------:|:------------------:|:------------------:|:-------------:|:--------------:|
 
14
  - Currently supported subtitle formats :
15
  - SRT
16
  - WebVTT
17
+ - Speech to Text Translation
18
+ - From other languages to English. ( This is Whisper's end-to-end speech-to-text translation feature )
19
+ - Text to Text Translation
20
+ - Translate subtitle files using Facebook NLLB models
21
 
22
  # Installation and Running
23
  ## Prerequisite
 
41
 
42
  And you can also run the project with command line arguments if you like by running `user-start-webui.bat`, see [wiki](https://github.com/jhj0517/Whisper-WebUI/wiki/Command-Line-Arguments) for a guide to arguments.
43
 
44
+ ## VRAM Usages
45
+ This project is integrated with [faster-whisper](https://github.com/guillaumekln/faster-whisper) by default for better VRAM usage and transcription speed.
46
 
47
+ According to faster-whisper, the efficiency of the optimized whisper model is as follows:
48
+ | Implementation | Precision | Beam size | Time | Max. GPU memory | Max. CPU memory |
49
+ |-------------------|-----------|-----------|-------|-----------------|-----------------|
50
+ | openai/whisper | fp16 | 5 | 4m30s | 11325MB | 9439MB |
51
+ | faster-whisper | fp16 | 5 | 54s | 4755MB | 3244MB |
52
+
53
+ If you want to use the original Open AI whisper implementation instead of optimized whisper, you can set the command line argument `DISABLE_FASTER_WHISPER` to `True`. See the [wiki](https://github.com/jhj0517/Whisper-WebUI/wiki/Command-Line-Arguments) for more information.
54
+
55
+ ## Available models
56
+ This is Whisper's original VRAM usage table for models.
57
 
58
  | Size | Parameters | English-only model | Multilingual model | Required VRAM | Relative speed |
59
  |:------:|:----------:|:------------------:|:------------------:|:-------------:|:--------------:|