Skip to content

Transcribe audio and video with Whisper-WebUI

Whisper-WebUI is a browser-based speech-to-text tool for generating transcripts and subtitle files from audio, video, YouTube links, and microphone recordings. It also includes standalone tools for subtitle translation and vocal/background music separation.

Use this guide to transcribe media files on Olares, improve transcription results with optional filters, translate existing subtitle files, and separate vocals from background music when needed.

Learning objectives

In this guide, you will learn how to:

  • Install Whisper-WebUI on Olares.
  • Transcribe local files, YouTube videos, and microphone recordings.
  • Improve transcription results with background music removal, VAD, and speaker diarization.
  • Translate subtitles and separate vocals from background music.
  • Find generated files and handle model downloads if automatic downloads fail.

Install Whisper-WebUI

  1. Open Market and search for "Whisper-WebUI". Install Whisper-WebUI

  2. Click Get, then Install, and wait for installation to complete.

After installation, you will see two icons on Launchpad:

  • Whisper-WebUI: The main interface for transcription, subtitle translation, and background music separation.
  • Whisper-WebUI Terminal: A command-line terminal for managing models.

Understand the basics

Main workflows

Whisper-WebUI includes five main tabs, which fall into two categories: transcription and standalone tools.

TabTypeBest for
FileTranscriptionGenerating subtitles for local media files, up to 500 MB.
YoutubeTranscriptionTranscribing online videos via URL without downloading them manually.
MicTranscriptionRecording and transcribing audio directly in the browser.
T2T TranslationStandalone toolTranslating existing subtitle files.
BGM SeparationStandalone toolExporting separate vocal and instrumental tracks.

Whisper-WebUI interface

Choose an output format

When using the transcription tabs, choose the output format based on how you plan to use the result.

FormatBest for
SRTStandard subtitle files for video players and editors.
WebVTTWeb video subtitles and browser-based playback.
TXTPlain text transcripts without timestamps.
LRCSynchronized lyrics for music players and audio applications.

Choose a transcription model

For most tasks, start with large-v2. It is pre-installed and works well for general transcription.

Change the model when you need a different balance between speed, accuracy, and resource usage:

NeedRecommended modelNotes
Faster processingsmall or mediumUse when large models are too slow. Accuracy may be lower.
Lowest resource usagetiny or baseUse only for quick tests or simple audio. Accuracy is limited.
Higher accuracylarge-v3Better for complex, noisy, or non-English audio, but uses more resources.
Faster large-model transcriptionlarge-v3-turboFaster than large-v3, with some accuracy tradeoff.
English-only audioModels ending in .enUse only when the source audio is English.

First-time downloads

Only large-v2 is pre-installed. Other models are downloaded automatically the first time you select them. The download may take some time, depending on your network and the model size.

Transcribe audio and video

The File, Youtube, and Mic tabs follow the same transcription workflow and share core settings such as model, language, output format, and advanced settings.

After each transcription task finishes, Whisper-WebUI shows the transcript in the output area and provides a downloadable subtitle or text file. Generated files are also saved in Files under /External/olares/ai/output/whisperwebui/.

Transcribe local files

  1. Click the File tab.
  2. Click the upload area and select an audio or video file. The file size limit is 500 MB.
  3. Under Model, select a transcription model.
  4. Under Language, specify the source language or use Automatic Detection.

    TIP

    Specifying the language can improve accuracy, especially for short audio or non-English content.

  5. Under File Format, choose your preferred output format.
  6. Optional: Expand panels below to remove background music, detect speech with VAD, or identify speakers.
  7. Click GENERATE SUBTITLE FILE.

File transcription result

Transcribe YouTube videos

YouTube access limitation

YouTube transcription may fail if YouTube blocks automated access or download requests, or if the network environment cannot access the video.

  1. Click the Youtube tab.

  2. Paste the YouTube video URL into the input field. Whisper-WebUI detects the video's thumbnail, title, and description when available.

    YouTube URL input

  3. Under Model, select a transcription model.

  4. Under Language, specify the video's language.

  5. Under File Format, choose your preferred output format.

  6. Optional: Expand panels below to remove background music, detect speech with VAD, or identify speakers.

  7. Click GENERATE SUBTITLE FILE.

YouTube transcription result

Record and transcribe with microphone

Microphone access requirement

Microphone recording requires browser microphone permission and HTTPS or localhost access. If recording does not work, check browser permissions and how you opened Whisper-WebUI.

  1. Click the Mic tab.

  2. Click the record button to start recording. You can pause at any time.

    Mic recording

  3. Click Stop to end the recording. You can preview and trim the audio.

    Mic recorded

  4. Select the Model, Language, and File Format for transcription.

  5. Optional: Expand panels below to remove background music, detect speech with VAD, or identify speakers.

  6. Click GENERATE SUBTITLE FILE.

Optional transcription filters

The File, YouTube, and Mic tabs include optional filters that can improve results for specific audio types. Configure these filters before clicking GENERATE SUBTITLE FILE.

Remove background music before transcription

Use this feature when speech is mixed with music or background audio.

To enable it:

  1. Expand Background Music Remover Filter.
  2. Check Enable Background Music Remover Filter.
  3. Keep the default model and segment size unless you need a custom setup.

Whisper-WebUI separates the vocal track first, then transcribes the processed audio.

BGM Separation vs. background music removal

The BGM Separation tab only separates audio into vocal and instrumental tracks. It does not transcribe the result.

Background music removal is part of the transcription workflow. It separates vocals first, then transcribes the vocal track.

Detect speech segments with VAD

Use VAD for long recordings, meetings, podcasts, or audio with long silent sections. VAD can skip silence, speed up transcription, and reduce hallucinated text from silent audio.

To enable it:

  1. Expand Voice Detection Filter.
  2. Check Enable Silero VAD Filter.

Identify speakers in multi-speaker audio

Speaker diarization labels different speakers in the transcript, such as SPEAKER_00 and SPEAKER_01.

Before using it for the first time, complete the Hugging Face setup so Whisper-WebUI can download the required models:

  1. Expand the Diarization panel.
  2. Under the HuggingFace Token field, click the two provided pyannote model links.
  3. On Hugging Face, log in or create a free account, then accept the conditions to access both models.
  4. In your Hugging Face account settings, create an access token with Read permissions.
  5. Back in Whisper-WebUI, check Enable Diarization.
  6. Paste your Hugging Face token into the HuggingFace Token input field.
  7. Run transcription as usual.

First-time diarization download

The first time you enable speaker diarization, Whisper-WebUI uses your token to download the required models. This may take some time. After the models are downloaded, they can be reused for future transcriptions.

Use standalone tools

Besides transcription, Whisper-WebUI provides dedicated tabs for subtitle translation and audio separation.

Translate subtitles

Use the T2T Translation tab to translate existing subtitle files.

Whisper-WebUI provides two translation methods:

MethodRequirementBest for
NLLBDownloads a local translation model on first use.Local translation without an external API key.
DeepL APIRequires a DeepL API key.Online translation using DeepL.

Separate vocals from background music

Use the BGM Separation tab to split an audio file into separate vocal and instrumental tracks. This standalone tool does not transcribe the result.

  1. Click the BGM Separation tab.
  2. Upload the audio file you want to process.
  3. Under Device, select a processing device based on your hardware.
  4. Under Model, select a separation model.
  5. Click SEPARATE BACKGROUND MUSIC.

BGM separation result

Once complete, you can preview the result, download the generated file, or check it in Files:

  • Instrumental: /External/olares/ai/output/whisperwebui/UVR/instrumental
  • Vocals: /External/olares/ai/output/whisperwebui/UVR/vocals

Advanced: Manage model downloads from Terminal

Most users do not need to manage models manually. Use Whisper-WebUI Terminal only when automatic model downloads fail, time out, or you need to check whether a model has already been downloaded.

Click the Whisper-WebUI Terminal icon on the Launchpad to open the web terminal.

Check downloaded models

Open Whisper-WebUI Terminal, then run:

bash
find /Whisper-WebUI/models -maxdepth 3 -type d

To check specific model folders:

bash
# Whisper transcription models
ls -la /Whisper-WebUI/models/Whisper/faster-whisper/

# NLLB translation models
ls -la /Whisper-WebUI/models/NLLB/

# UVR background music separation models
ls -la /Whisper-WebUI/models/UVR/MDX_Net_Models/

# Speaker diarization models
ls -la /Whisper-WebUI/models/Diarization/

Manually download models

While Whisper-WebUI downloads models automatically, you can trigger downloads manually if the UI download times out:

For example, to download a Whisper transcription model, replace the repository name with the model you need:

bash
hf download Systran/faster-whisper-large-v3 \
  --cache-dir /Whisper-WebUI/models/Whisper/faster-whisper

After the download completes, refresh Whisper-WebUI and select the model from the model list.

FAQs

Why does T2T Translation fail with NLLB?

NLLB translation may fail if the model download was interrupted or the model folder is incomplete.

To reset the NLLB download:

  1. Open External/olares/ai/whisperwebui/NLLB/ in Files.
  2. Delete the contents inside the folder, but keep the folder itself.
  3. Return to Whisper-WebUI, and download the model again.

Why does speaker diarization fail?

Speaker diarization may fail if:

  • The Hugging Face token is missing or invalid.
  • The required pyannote model terms were not accepted.
  • The model download failed because of network issues.

Check that:

  • The Hugging Face token has Read permissions.
  • You accepted the terms for both pyannote models using the same Hugging Face account.
  • Your network is stable while Whisper-WebUI downloads the models.

Why do tasks fail after switching to a larger model?

A task may fail after you switch models for one of these reasons:

  • The selected model has not finished downloading.
  • The model requires more GPU memory than Whisper-WebUI currently has.

To fix this issue:

  • Wait for the first-time model download to complete, then retry the task.
  • Assign more VRAM to Whisper-WebUI in Memory slicing mode.
  • Switch the GPU to another suitable mode.
  • Choose a smaller model.

Learn more

  • Open WebUI: Use Whisper-WebUI as a speech-to-text backend for chat input.