jojojaeger / whisper-streamlit

this master thesis project is based on OpenAI Whisper with the goal to transcibe interviews
https://openai.com/research/whisper
Other
41 stars 16 forks source link
asr clinical-research speaker-identification whisper

Whisper with Streamlit UI

This project is a tool that was developed as part of a Master's thesis in cooperation with the University Clinic of Psychoanalysis and Psychotherapy of Vienna. It is based on the Whisper automatic speech recogniton system and is embedded into a Streamlit Web App.

Features

Data Privacy

User Interface

Start Screen

Start Screen

Results

Results

Getting Started

To use this tool, you will need to install the required dependencies and run the Streamlit app. You can do this by following these steps:

  1. Clone the repository: git clone https://github.com/jojojaeger/whisper-streamlit
  2. Install prerequisites: Python, Pip, Git, PyTorch (pip install torch torchvision torchaudio)
  3. Install dependencies: pip install -r requirements.txt
  4. Run the Streamlit app: streamlit run Transcribe.py (you can also launch it from a desktop shortcut following these instructions: https://discuss.streamlit.io/t/launching-streamlit-webapp-from-desktop-shortcut/26297)

How to Use

  1. Upload one or multiple audio files
  2. Select a model (large for the best result) and set additional parameters
  3. Download the resulting transcript (also saved to local transcripts-folder)

Fixes to common errors

Contact

If you have any questions or feedback about this project, please feel free to contact us by email at johanna.jaeger89@icloud.com.

Sources

This project includes code from multiple different sources, each licensed under the MIT License:

See the LICENSE file for the full text of the licenses.