batra-mlp-lab / visdial

[CVPR 2017] Torch code for Visual Dialog
https://arxiv.org/abs/1611.08669
Other
227 stars 69 forks source link

Pytorch starter code #28

Closed shubhamagarwal92 closed 6 years ago

shubhamagarwal92 commented 6 years ago

Hi,

Do you guyz plan to release starter code in pytorch for the challenge? visdial-rl does provide some insights but is tailored more for Visual Dialog Agents as described in the paper "Learning Cooperative Visual Dialog Agents with Deep Reinforcement Learning"

nirbhayjm commented 6 years ago

Hi, visdial-rl is tailored more towards the two-agent setup but does support Visual Dialog training and evaluation i.e. just the Answering agent (A-Bot). The training and evaluation can be used with A-Bot alone to do supervised training on the VisDial 0.5 and 0.9 datasets with evaluation on the standard VisDial metrics.

Though the codebase supports just HRE (Hierarchical recurrent encoder and decoder) in hre.py, it can still serve as a starting point for building your own models for visual dialog.

kdexd commented 6 years ago

Hi @shubhamagarwal92, visdial-rl will have some enhancements - we would add a section in _README_describing the usage of that codebase as a starter code for Visual Dialog Challenge. Please watch out for announcements on Discord about this soon. Thanks!