-
- https://arxiv.org/abs/2104.12099
- 2021
最近では、CNNベースのアーキテクチャに依存した大規模な顕著性検出手法が有望な結果を得ています。
本研究では、この課題を畳み込みなしの配列対配列の観点から再考し、畳み込みでは実現できない長距離依存性をモデル化することで、顕著性を予測する。
具体的には、RGBとRGB-Dの両方の顕著なオブジェクト検出(S…
e4exp updated
3 years ago
-
## 論文概要
映像中の音声が人間の注意力に影響を与えることを明らかにした研究。大規模な音声付アイトラッキング映像データベース(34名被験者が300の動画を視聴)を取得して傾向を分析、基本的に人の顔に注意が行くことを確認。映像中のSaliencyを予測するためのマルチモーダルNNを提案。
![bib_20200920 00](https://user-images.githubuserco…
tkuri updated
3 years ago
-
Thank you for the insightful paper! To my knowledge, there are no publicly available datasets with head/eye tracking for audio-visual data so I would be very interested in further analyzing your datas…
-
非常感谢您的工作!请问可以提供一下数据集的融合MASK的生成代码吗?我想在别的数据集上做一下尝试。非常感谢!
-
https://arxiv.org/abs/1711.10795
-
You should identify and read some (3-5) scientific papers or works where similar to your research.
I think that in the world of egocentric they use a lot the term "attention" as a similar concept to …
-
paper link : https://www.arxiv.org/abs/2406.17815
github: https://github.com/Arhosseini77/SUM
google colab: https://colab.research.google.com/drive/1G6bZ_knpNDo105q4rLQUPIMhP-_5bpZR?usp=sharing
-
Hi, I am trying to implement the DeepNet architecture in pytorch. The code seems to work fine but the result are not as expected. I have done as per the protext files which are provided in the issue 3…
-
Dear all,
I am testing ObjectnessBING in matlab.
I am using a mexopencv wrapper:
https://github.com/kyamagu/mexopencv
When I tested the sample code:
mexopencv/opencv_contrib/samples/compute…
-
Hi, Lai, nice work.
Based on my understanding, the gt should be a 2D map instead of such groups of fixation pixels in Data.fixdata. However, how do you generate such fixation data into the ground tr…