codertimo / paper-log

읽어야 하는 논문들을 관리하고, 읽은 논문들의 기록을 남기는 공간
31 stars 5 forks source link

Yoshua Bengio: From System 1 Deep Learning to System 2 Deep Learning (NeurIPS 2019) #17

Open codertimo opened 4 years ago

codertimo commented 4 years ago

어떤 내용의 레퍼런스인가요? 👋

Yoshua Bengio 교수님이 NeurIPS 2019 에서 발표하신 키노트 자료입니다. 앞으로 딥러닝이 나아가야할 방향에 대한 컨텐츠를 담고 있습니다.

이 글을 읽어서 무엇을 배울 수 있는지 알려주세요! 🤔

같이 읽어보면 좋을 만한 글이나 이슈가 있을까요?

Past progress in deep learning has concentrated mostly on learning from a static dataset, mostly for perception tasks and other System 1 tasks which are done intuitively and unconsciously by humans. However, in recent years, a shift in research direction and new tools such as soft-attention and progress in deep reinforcement learning are opening the door to the development of novel deep architectures and training frameworks for addressing System 2 tasks (which are done consciously), such as reasoning, planning, capturing causality and obtaining systematic generalization in natural language processing and other applications. Such an expansion of deep learning from System 1 tasks to System 2 tasks is important to achieve the old deep learning goal of discovering high-level abstract representations because we argue that System 2 requirements will put pressure on representation learning to discover the kind of high-level concepts which humans manipulate with language. We argue that towards this objective, soft attention mechanisms constitute a key ingredient to focus computation on a few concepts at a time (a "conscious thought") as per the consciousness prior and its associated assumption that many high-level dependencies can be approximately captured by a sparse factor graph. We also argue how the agent perspective in deep learning can help put more constraints on the learned representations to capture affordances, causal variables, and model transitions in the environment. Finally, we propose that meta-learning, the modularization aspect of the consciousness prior and the agent perspective on representation learning should facilitate re-use of learned components in novel ways (even if statistically improbable, as in counterfactuals), enabling more powerful forms of compositional generalization, i.e., out-of-distribution generalization based on the hypothesis of localized (in time, space, and concept space) changes in the environment due to interventions of agents.

레퍼런스의 URL을 알려주세요! 🔗

https://www.youtube.com/watch?v=T3sxeTgT4qc