Open tensorfoo opened 2 years ago
hi, I have the same problem, have you slove it?
Hi, If you do an inference with a
Transcriber
objectt.transcribe(..)
, after returning the result, it should release any resources related to that inference. But it stays on VRAM and after a few calls tot.transcribe()
, I get CUDA related out of memory errors. Looking atnvidia-smi
it shows the memory is still occupied even after the transcript has been returned.It would be nice to have a long lived
Transcriber
object which can be reused avoiding the lengthy creation time. If you're busy please give me a hint on how it may be done so I can give it a shot and submit a PR. Thanks for your project.
HI I have the same problem, have you slove it?
Hi, If you do an inference with a
Transcriber
objectt.transcribe(..)
, after returning the result, it should release any resources related to that inference. But it stays on VRAM and after a few calls tot.transcribe()
, I get CUDA related out of memory errors. Looking atnvidia-smi
it shows the memory is still occupied even after the transcript has been returned.It would be nice to have a long lived
Transcriber
object which can be reused avoiding the lengthy creation time. If you're busy please give me a hint on how it may be done so I can give it a shot and submit a PR. Thanks for your project.