whisper-asr-webservice
Whisper is a general-purpose speech recognition model. It is trained on a large dataset of diverse audio and is also a multi-task model that can perform multilingual speech recognition as well as speech translation and language identification. For more details: github.com/openai/whisper Use ASR_MODEL to specify which model to run. Available models are tiny, base, small, medium, large, large-v1 and large-v2. Please note that large and large-v2 are the same model. For English-only applications, the .en models tend to perform better, especially for the tiny.en and base.en models. We observed that the difference becomes less significant for the small.en and medium.en models. The ASR model is downloaded each time you start the container, using the large model this can take some time. If you want to decrease the time it takes to start your container by skipping the download, you can provide a persistent storage path. Next time you start your container the ASR Model will be loaded from there instead of being downloaded again. For GPU support, add '--gpus all' to 'Extra Parameters'.