-
Hello Author! I would like to ask, the source code both visual.py and audio-visual.py call sal_unet in saliency_decoder, which doesn't match the structure proposed in the paper, ah? What's going on he…
-
- https://arxiv.org/abs/2104.12099
- 2021
最近では、CNNベースのアーキテクチャに依存した大規模な顕著性検出手法が有望な結果を得ています。
本研究では、この課題を畳み込みなしの配列対配列の観点から再考し、畳み込みでは実現できない長距離依存性をモデル化することで、顕著性を予測する。
具体的には、RGBとRGB-Dの両方の顕著なオブジェクト検出(S…
e4exp updated
3 years ago
-
### Problem Description
I've realized that SHAP doesn't have an implementation for Saliency Maps. Even though image_plots lets us compute the Shapley values for an image, it doesn't provide an attent…
-
## 論文概要
映像中の音声が人間の注意力に影響を与えることを明らかにした研究。大規模な音声付アイトラッキング映像データベース(34名被験者が300の動画を視聴)を取得して傾向を分析、基本的に人の顔に注意が行くことを確認。映像中のSaliencyを予測するためのマルチモーダルNNを提案。
![bib_20200920 00](https://user-images.githubuserco…
tkuri updated
3 years ago
-
Hi @ByungKwanLee and thanks for your work on this repo. I was trying to reproduce some experiments, and if I understood correctly, the MAD pipeline works as follows:
1-Clean pre-training of the mode…
-
非常感谢您的工作!请问可以提供一下数据集的融合MASK的生成代码吗?我想在别的数据集上做一下尝试。非常感谢!
-
https://arxiv.org/abs/1711.10795
-
Thank you for the insightful paper! To my knowledge, there are no publicly available datasets with head/eye tracking for audio-visual data so I would be very interested in further analyzing your datas…
-
You should identify and read some (3-5) scientific papers or works where similar to your research.
I think that in the world of egocentric they use a lot the term "attention" as a similar concept to …
-
Hi, I am trying to implement the DeepNet architecture in pytorch. The code seems to work fine but the result are not as expected. I have done as per the protext files which are provided in the issue 3…