OpenThought - System 2 Research Links
Here you find a collection of material (books, papers, blog-posts etc.) related to reasoning and cognition in AI systems. Specifically we want to cover agents, cognitive architectures, general problem solving strategies and self-improvement.
The term "System 2" in the page title refers to the slower, more deliberative, and more logical mode of thought as described by Daniel Kahneman in his book Thinking, Fast and Slow.
You know a great resource we should add? Please see How to contribute.
Cognitive Architectures
(looking for additional links & articles and summaries)
- SOAR (State, Operator, And Result) by John Laird, Allen Newell, and Paul Rosenbloom
- ACT-R (Adaptive Control of Thought-Rational) by John Anderson at CMU
- SPAUN (Semantic Pointer Architecture Unified Network) by Chris Eliasmith at Waterloo, SPAUN 2.0 by Feng-Xuan Choo
- ART (Adaptive resonance theory) by Stephen Grossberg and Gail Carpenter
- CLARION (Connectionist Learning with Adaptive Rule Induction ON-line) by Ron Sun
- EPIC (Executive Process/Interactive Control) by David Kieras and David Meyer
- LIDA (Learning Intelligent Distribution Agent) by Stan Franklin, 2015 Paper
- Sigma by Paul Rosenbloom
- OpenCog by Ben Goertzel
- NARS (Non-Axiomatic Reasoning System) by Pei Wang
- Icarus by Pat Langley
- MicroPsi by Joscha Bach
- Thousand Brains Theory & HTM (Hierarchical Temporal Memory) by Jeff Hawkins
- SPH (Sparse Predictive Hierarchie) by Eric Laukien
- Leabra (Local, Error-driven and Associative, Biologically Realistic Algorithm), 2016 Paper by Randall O'Reilly
- CogNGen (COGnitive Neural GENerative system) by Alexander Ororbia and Mary Alexandria Kelly, see also here and here
- KIX (KIX: A Metacognitive Generalization Framework) by A. Kumar and Paul Schrater
- ACE (Autonomous Cognitive Entity) by David Shapiro et al., gh: daveshap/ACE_Framework
- Iterative Updating of Working Memory by Jared Reser, website, Video
Agent Papers
LLM Based
LLM Reasoning Improvements / Training on Synthetic Data
Direct o1 Replication Efforts
Reward Models (ORM/PRM)
RL
MCTS
Minecraft Agents
Massive Sampling / Generate-and-Test
World Models
Neuro-Symbolic Approaches
Math
Active Inference
Prompting Techniques
- Surveys:
- Tools:
- Chain-of-Thoughts (COT): Paper
- Tree-of-Thoughts (ToT): Paper, impl: Strategic Debate
- Graph-of-Thoughts (GoT): Paper, code
- Algorithm of Thoughts (AoT): Paper
- Chain-of-Verification (CoVe/CoV): Paper
- Mixture-of-Agents (MoA): Paper
- Tool-Integrated Reasoning (ToRA / TIR): Paper
- Program of Thoughts (PoT): Paper
- Buffer of Thoughts (BoT): Paper
- Chain of Code (CoC): Paper
- Thought of Search (ToS): Paper
- Re-Reading the question as input (RE2): Paper
- Self-Harmonized Chain of Thought (ECHO): Paper, code
- Divergent CoT (DCoT), Paper
- Iteration of Thought (IoT), Paper
- Logic-of-Thought (LoT) Paper
Negative results
Mechanistic Interpretability
Blog Posts / Presentations
Graph Neural Networks
Complex Logical Query Answering (CQLA)
Answering logical queries over Incomplete Knowledge Graphs. Aspirationally this requires combining sparse symbolic index collation (SQL, SPARQL, etc) and dense vector search, preferably in a differentiable manner.
Inductive Reasoning over Heterogeneous Graphs
Similar to the regular CQLA, but with the emphasis on the "Inductive Setting" - i.e. querying over new, unseen during training nodes, edge types or even entire graphs. The latter part is interesting as it relies on the higher order "relations between relations" structure, connecting KG inference to Category Theory.
Neural Algorithmic Reasoning (NAR)
Initially attempted back in 2014 with general-purpose but unstable Neural Turing Machines, modern NAR approaches limit their scope to making GNN-based "Algorithmic Processor Networks" which learn to mimic classical algorithms on synthetic data and can be deployed on noisy real-world problems by sandwiching their frozen instances inside Encoder-Processor-Decoder architecture.
Grokking
Open-Source Agents & Agent Frameworks
- gpt-researcher, docs
- open-interpreter, docs
- ADAS (Automated Design of Agentic Systems)
- AI-Scientist
- Ollama_Agents
- AgentK
- Storm, Paper
- crewAI, docs
- AutoGPT, docs
- AutoGen, docs, AutoGen Studio Paper
- Trace, docs, Paper
- motleycrew, docs
- langflow, docs
- show-me: A Visual and Transparent Reasoning Agent
Algorithms
Weak Search Methods
Weak methods are general but don't use knowledge (heuristics) to guide the search process.
Strong Search Methods
Books
- The Soar Cognitive Architecture, John E. Laird, MIT Press, 2019
- How to Build a Brain: A Neural Architecture for Biological Cognition Chris Eliasmith, Oxford Series on Cognitive Models and Architectures, 2013
- Active Inference: The Free Energy Principle in Mind, Brain, and Behavior, Thomas Parr, Giovanni Pezzulo, Karl J. Friston, MIT Press, 2022, MLST Interview with Thomas Parr
- Principles of Synthetic Intelligence PSI: An Architecture of Motivated Cognition, Joscha Bach, Oxford Series on Cognitive Models and Architectures Book 4, 2009
- Conscious Mind, Resonant Brain: How Each Brain Makes a Mind, Stephen Grossberg, Oxford University Press, 2021
- The Society of Mind, Marvin Minsky, Simon & Schuster, 1986
- Reinforcement Learning: An Introduction 2nd Edition, Sutton & Barto, MIT Press, 2018
- Mathematical Foundations of Reinforcement Learning, Shiyu Zhao, open course on github + video lectures
- Natural Language Cognitive Architecture, David Shapiro, 2022, open source copy
- An Introduction to Universal Artificial Intelligence, Marcus Hutter, David Quarel, Elliot Catt, CRC Press, 2024 - AIXI, Slides, Video
Biologically Inspired Approaches
Diverse approaches some of which tap into classical PDE systems of biological NNs, some concentrate on Distibuted Sparse Representations (by default non-differentiable), others draw inspiration from Hippocampal Grid Cells, Place Cells, etc. Biological systems surpass most ML methods for Continual and Online Learning, but are hard to implement efficienly on GPU.
Dense Associative Memory
Dense Associative Memory is mainly represented by Modern Hopfield Networks (MHN), which can be viewed as a generalized Transformers capable of storing queries, keys and values explicitly (as in Vector Databases) and running recurrent retrival by energy minimization (relating them to Diffusion models). Application for Continual Learning is possible when combined with uncertainty quantification and differentiable top-k selection.
Continual Learning
Software Tools & Libraries
Commercial Offerings
Competitions & Benchmarks
- DevAI: Agent-as-a-Judge: Evaluate Agents with Agents
- AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agents, web: project page, gh: stonybrooknlp/appworld, Leaderboard
- CRAB: Cross-environment Agent Benchmark for Multimodal Language Model Agents, gh: camel-ai/crab
- WebArena: A Realistic Web Environment for Building Autonomous Agents, web: project page, Leaderboard
- ARC-AGI: Leaderboard, On the Measure of Intelligence
- PlanBench: Paper, gh: karthikv792/LLMs-Planning
- GAIA: a benchmark for General AI Assistants: Leaderboard
- StreamBench: Towards Benchmarking Continuous Improvement of Language Agents, gh: stream-bench/stream-bench
- VisualAgentBench: Towards Large Multimodal Models as Visual Foundation Agents
- ZebraLogic, Leaderboard
- Omni-MATH, gh: KbsdJames/Omni-MATH
- BatsResearch/planetarium - Dataset and benchmark for assessing LLMs in translating natural language descriptions of planning problems into PDDL
Code
Related Projects
Youtube Content
Joscha Bach
Best LLM APIs
Novel model architectures
Philosophy: Nature of Intelligence & Consciousness
Biology / Neuroscience
Workshops
https://s2r-at-scale-workshop.github.io (NeurIPS 2024)
How to contribute
To share a link related to reasoning in AI systems that is missing here please create a pull request for this file. See editing files in the github documentation.