Workshop
Memory in Artificial and Real Intelligence (MemARI)
Mariya Toneva · Javier Turek · Vy Vo · Shailee Jain · Kenneth Norman · Alexander Huth · Uri Hasson · Mihai Capotă
Room 397
Fri 2 Dec, 6:30 a.m. PST
One of the key challenges for AI is to understand, predict, and model data over time. Pretrained networks should be able to temporally generalize, or adapt to shifts in data distributions that occur over time. Our current state-of-the-art (SOTA) still struggles to model and understand data over long temporal durations – for example, SOTA models are limited to processing several seconds of video, and powerful transformer models are still fundamentally limited by their attention spans. On the other hand, humans and other biological systems are able to flexibly store and update information in memory to comprehend and manipulate multimodal streams of input. Cognitive neuroscientists propose that they do so via the interaction of multiple memory systems with different neural mechanisms. What types of memory systems and mechanisms already exist in our current AI models? First, there are extensions of the classic proposal that memories are formed via synaptic plasticity mechanisms – information can be stored in the static weights of a pre-trained network, or in fast weights that more closely resemble short-term plasticity mechanisms. Then there are persistent memory states, such as those in LSTMs or in external differentiable memory banks, which store information as neural activations that can change over time. Finally, there are models augmented with static databases of knowledge, akin to a high-precision long-term memory or semantic memory in humans. When is it useful to store information in each one of these mechanisms, and how should models retrieve from them or modify the information therein? How should we design models that may combine multiple memory mechanisms to address a problem? Furthermore, do the shortcomings of current models require some novel memory systems that retain information over different timescales, or with different capacity or precision? Finally, what can we learn from memory processes in biological systems that may advance our models in AI? We aim to explore how a deeper understanding of memory mechanisms can improve task performance in many different application domains, such as lifelong / continual learning, reinforcement learning, computer vision, and natural language processing.
Schedule
Fri 6:30 a.m. - 6:45 a.m.
|
Opening remarks
(
Talk
)
>
SlidesLive Video |
Vy Vo 🔗 |
Fri 6:45 a.m. - 7:30 a.m.
|
Sepp Hochreiter: "Modern Hopfield Networks"
(
Keynote
)
>
SlidesLive Video |
Sepp Hochreiter 🔗 |
Fri 7:30 a.m. - 8:15 a.m.
|
Sainbayar Sukhbaatar: "Brain-inspired memory models"
(
Keynote
)
>
SlidesLive Video |
Sainbayar Sukhbaatar 🔗 |
Fri 8:15 a.m. - 8:20 a.m.
|
The Emergence of Abstract and Episodic Neurons in Episodic Meta-RL
(
Spotlight
)
>
link
SlidesLive Video |
Badr AlKhamissi · Muhammad ElNokrashy · Michael Spranger 🔗 |
Fri 8:20 a.m. - 8:25 a.m.
|
Learning to Reason and Memorize with Self-Questioning
(
Spotlight
)
>
link
SlidesLive Video |
Jack Lanchantin · Shubham Toshniwal · Jason E Weston · arthur szlam · Sainbayar Sukhbaatar 🔗 |
Fri 8:25 a.m. - 8:30 a.m.
|
Recall-gated plasticity as a principle of systems memory consolidation
(
Spotlight
)
>
link
SlidesLive Video |
Jack Lindsey · Ashok Litwin-Kumar 🔗 |
Fri 8:30 a.m. - 8:35 a.m.
|
The Opportunistic PFC: Downstream Modulation of a Hippocampus-inspired Network is Optimal for Contextual Memory Recall
(
Spotlight
)
>
link
SlidesLive Video |
Hugo Chateau-Laurent · Frederic Alexandre 🔗 |
Fri 8:35 a.m. - 10:55 a.m.
|
Poster session + Lunch
(
In-person poster session
)
>
|
🔗 |
Fri 11:00 a.m. - 11:45 a.m.
|
Hava Siegelmann: "Lifelong learning and supporting memory"
(
Keynote
)
>
SlidesLive Video |
Hava Siegelmann 🔗 |
Fri 11:45 a.m. - 12:30 p.m.
|
Ida Mommenejad: "Neuro-inspired Memory in Reinforcement Learning: State of the art, Challenges, and Opportunities"
(
Keynote
)
>
SlidesLive Video |
Ida Momennejad 🔗 |
Fri 12:30 p.m. - 12:45 p.m.
|
Afternoon break
|
🔗 |
Fri 12:45 p.m. - 1:30 p.m.
|
Janice Chen: "Memory for narratives"
(
Keynote
)
>
SlidesLive Video |
Janice Chen 🔗 |
Fri 1:30 p.m. - 2:55 p.m.
|
Panel Discussion: Opportunities and Challenges
(
Discussion panel (in-person)
)
>
SlidesLive Video |
Kenneth Norman · Janice Chen · Samuel J Gershman · Albert Gu · Sepp Hochreiter · Ida Momennejad · Hava Siegelmann · Sainbayar Sukhbaatar 🔗 |
Fri 2:55 p.m. - 3:00 p.m.
|
Closing remarks
(
Talk
)
>
SlidesLive Video |
Mariya Toneva 🔗 |
-
|
Learning to Control Rapidly Changing Synaptic Connections: An Alternative Type of Memory in Sequence Processing Artificial Neural Networks ( Poster ) > link | Kazuki Irie · Jürgen Schmidhuber 🔗 |
-
|
Constructing compressed number lines of latent variables using a cognitive model of memory and deep neural networks ( Poster ) > link | Sahaj Singh Maini · James Mochizuki-Freeman · Chirag Shankar Indi · Brandon Jacques · Per B Sederberg · Marc Howard · Zoran Tiganj 🔗 |
-
|
Experiences from the MediaEval Predicting Media Memorability Task ( Poster ) > link | Alba Garcia Seco de Herrera · Mihai Gabriel Constantin · Claire-Helene Demarty · Camilo Fosco · Sebastian Halder · Graham Healy · Bogdan Ionescu · Ana Matran-Fernandez · Alan F Smeaton · Mushfika Sultana 🔗 |
-
|
Meta-Learning General-Purpose Learning Algorithms with Transformers ( Poster ) > link | Louis Kirsch · Luke Metz · James Harrison · Jascha Sohl-Dickstein 🔗 |
-
|
Evidence accumulation in deep RL agents powered by a cognitive model ( Poster ) > link | James Mochizuki-Freeman · Sahaj Singh Maini · Zoran Tiganj 🔗 |
-
|
Using Hippocampal Replay to Consolidate Experiences in Memory-Augmented Reinforcement Learning ( Poster ) > link | Chong Min John Tan · Mehul Motani 🔗 |
-
|
Training language models for deeper understanding improves brain alignment ( Poster ) > link | Khai Loong Aw · Mariya Toneva 🔗 |
-
|
Transformer needs NMDA receptor nonlinearity for long-term memory ( Poster ) > link | Dong-Kyum Kim · Jea Kwon · Meeyoung Cha · C. Lee 🔗 |
-
|
Neural networks learn an environment's geometry in latent space by performing predictive coding on visual scenes ( Poster ) > link | James Gornet · Matt Thomson 🔗 |
-
|
Low Resource Retrieval Augmented Adaptive Neural Machine Translation ( Poster ) > link | Vivek Harsha Lakkamaneni · Swair Shah · Anurag Beniwal · Narayanan Sadagopan 🔗 |
-
|
Exploring The Precision of Real Intelligence Systems at Synapse Resolution ( Poster ) > link | Mohammad Samavat · Tom Bartol · Kristen Harris · Terrence Sejnowski 🔗 |
-
|
Cache-memory gated graph neural networks ( Poster ) > link | Guixiang Ma · Vy Vo · Nesreen K. Ahmed · Theodore Willke 🔗 |
-
|
CL-LSG: Continual Learning via Learnable Sparse Growth ( Poster ) > link | Li Yang · Sen Lin · Junshan Zhang · Deliang Fan 🔗 |
-
|
Toward Semantic History Compression for Reinforcement Learning ( Poster ) > link | Fabian Paischer · Thomas Adler · Andreas Radler · Markus Hofmarcher · Sepp Hochreiter 🔗 |
-
|
Learning at Multiple Timescales ( Poster ) > link | Matt Jones 🔗 |
-
|
Neural Network Online Training with Sensitivity to Multiscale Temporal Structure ( Poster ) > link | Matt Jones · Tyler Scott · Gamaleldin Elsayed · Mengye Ren · Katherine Hermann · David Mayo · Michael Mozer 🔗 |
-
|
Mixed-Memory RNNs for Learning Long-term Dependencies in Irregularly-sampled Time Series ( Poster ) > link | Mathias Lechner · Ramin Hasani 🔗 |
-
|
Transformers generalize differently from information stored in context vs in weights ( Poster ) > link | Stephanie Chan · Ishita Dasgupta · Junkyung Kim · Dharshan Kumaran · Andrew Lampinen · Felix Hill 🔗 |
-
|
Memory in humans and deep language models: Linking hypotheses for model augmentation ( Poster ) > link | Omri Raccah · Phoebe Chen · Theodore Willke · David Poeppel · Vy Vo 🔗 |
-
|
Characterizing Verbatim Short-Term Memory in Neural Language Models ( Poster ) > link | Kristijan Armeni · Christopher J Honey · Tal Linzen 🔗 |
-
|
Informing generative replay for continual learning with long-term memory formation in the fruit fly ( Poster ) > link | Brian Robinson · Justin Joyce · Raphael Norman-Tenazas · Gautam Vallabha · Erik Johnson 🔗 |
-
|
Learning to Reason and Memorize with Self-Questioning ( Poster ) > link | Jack Lanchantin · Shubham Toshniwal · Jason E Weston · arthur szlam · Sainbayar Sukhbaatar 🔗 |
-
|
Recall-gated plasticity as a principle of systems memory consolidation ( Poster ) > link | Jack Lindsey · Ashok Litwin-Kumar 🔗 |
-
|
The Opportunistic PFC: Downstream Modulation of a Hippocampus-inspired Network is Optimal for Contextual Memory Recall ( Poster ) > link | Hugo Chateau-Laurent · Frederic Alexandre 🔗 |
-
|
Multiple Modes for Continual Learning ( Poster ) > link | Siddhartha Datta · Nigel Shadbolt 🔗 |
-
|
The Emergence of Abstract and Episodic Neurons in Episodic Meta-RL ( Poster ) > link | Badr AlKhamissi · Muhammad ElNokrashy · Michael Spranger 🔗 |
-
|
Self-recovery of memory via generative replay ( Poster ) > link | Zhenglong Zhou · Geshi Yeung · Anna Schapiro 🔗 |
-
|
Associative memory via covariance-learning predictive coding networks ( Poster ) > link | Mufeng Tang · Tommaso Salvatori · Yuhang Song · Beren Millidge · Thomas Lukasiewicz · Rafal Bogacz 🔗 |
-
|
Interpolating Compressed Parameter Subspaces ( Poster ) > link | Siddhartha Datta · Nigel Shadbolt 🔗 |
-
|
A Universal Abstraction for Hierarchical Hopfield Networks ( Poster ) > link | Benjamin Hoover · Duen Horng Chau · Hendrik Strobelt · Dmitry Krotov 🔗 |
-
|
Differentiable Neural Computers with Memory Demon ( Poster ) > link | Ari Azarafrooz 🔗 |
-
|
Leveraging Episodic Memory to Improve World Models for Reinforcement Learning ( Poster ) > link | Julian Coda-Forno · Changmin Yu · Qinghai Guo · Zafeirios Fountas · Neil Burgess 🔗 |
-
|
Biological Neurons vs Deep Reinforcement Learning: Sample efficiency in a simulated game-world ( Poster ) > link | Forough Habibollahi · Moein Khajehnejad · Amitesh Gaurav · Brett J. Kagan 🔗 |
-
|
Constructing Memory: Consolidation as Teacher-Student Training of a Generative Model ( Poster ) > link | Eleanor Spens · Neil Burgess 🔗 |