TUM-Dev / gocast

TUMs lecture streaming service.
https://tum.live
MIT License
191 stars 42 forks source link

Show transcript for lectures with subtitles #1369

Open carlobortolan opened 2 months ago

carlobortolan commented 2 months ago

Is your feature request related to a problem? Please describe. -

Describe the solution you'd like

Describe alternatives you've considered Download option for transcripts as a text file.

Additional context

  1. I'd work on this issue as part of the Open Source Lab
2. YouTube transcript function ![image](https://github.com/user-attachments/assets/527c35e4-9655-48fb-aa32-bacffb410b9b)
joschahenningsen commented 2 months ago

probably duplicate of #1364 :) Functionality is implemented (see https://github.com/TUM-Dev/TUM-Live-Voice-Service), we never got enough GPU capacity to enable the feature. We're actively investigating buying hardware privately.

This lecture already has subtitles: grafik https://tum.live/w/cit-gr/26188

carlobortolan commented 2 months ago

probably duplicate of #1364 :) Functionality is implemented (see https://github.com/TUM-Dev/TUM-Live-Voice-Service)

Ah sorry for the confusion, besides searching for certain keywords, I was thinking more of a feature that lets you read and scroll through the entire transcript of a lecture. And while it probably won't be a major feature, I'd guess that it would still be an interesting functionality (especially for those who might be in a hurry and just want to quickly recap / scroll through the content of a certain lecture).

we never got enough GPU capacity to enable the feature. We're actively investigating buying hardware privately.

iirc, in a recent tumlive meeting, there was the idea of using some unused resources (e.g., the Worker VMs at night or during the semester breaks) to transcribe past VODs - even if the VMs don't have powerful GPUs, it still should be enough to transcribe VODs. So it might even be that next semester there will be more lecture VODs with subtitles :eyes:

CommanderStorm commented 2 months ago

we never got enough GPU capacity

This actually can't happen that we don't communicate such stuff. I have put a 3090 in "our the usual place". That should be enough and a big upgrade to the 1660 Super.

With algorithmic improvements like https://github.com/SYSTRAN/faster-whisper I don't see a reason why we could not have semi-realrime transcription if we are smart about this.

@joschahenningsen could you please install the GPU?