-
my arguments:
> python server.py --enable-modules=caption,summarize,classify,sd,silero-tts,edge-tts,coqui-tts,chromadb,talkinghead --cuda --talkinghead-gpu --coqui-gpu
but I get 'Error: Torch n…
-
Hi, I try the recaption in eval.sh. In the line 220, single_test, the model accurately output the caption, we show in the figure. But in captioning the Inter4K, line 134, the model's output just repea…
-
I am learning the Transformer_Captioning.ipynb in assignment3. After I run the cell of testing MultiHeadAttention, I get some incorrect results:
```
self_attn_output error: 0.449382070034207
maske…
-
When I run these codes on my computer, running is all okay.
But I confused about the epoch.
After training, I got 20 model files.
When I test with that model files, I got the best performance on **…
-
## タイトル: 思い通りに編集:複数段階のユーザー制御による動画キャプション編集
## リンク: https://arxiv.org/abs/2305.08389
## 概要:
ユーザーの要求に沿った自然言語による動画の自動ナレーション、すなわち制御可能なビデオキャプションタスクは、人々が膨大な動画を希望する意図に沿って管理するのに役立ちます。しかし、既存の研究には、2つの欠点があります…
-
Hi thanks for your work and public release of the code.
I have checked your code and I could not find the generate function of your model while using the VQA model. I want to be able to input new q…
-
The BLEU metric is a numerical metric used in image captioning.
They need to be implemented inside the `src/models/metrics.py` file (not sure if it is the correct place by the way).
I think this …
-
The enclosed pdf is a rough of an addition to the Overview Examples.
There are two thumbnails, I hope can be done by scaling the web code and removing all the links, or just a graphic is fine.
…
-
Share observations (pros and cons) between OBS and Zoom beta for slides or live coding in the background.
Recently I had an interesting teaching demo experience recently where a demonstrator didn't…
-
Zach Suggested that I write this up so I am doing so.
We talk about parity between HLS and DASH and one aspect of this is what information gets presented to the user agent ( e.g. information that is …