Workshop
Object Representations for Learning and Reasoning
William Agnew · Rim Assouel · Michael Chang · Antonia Creswell · Eliza Kosoy · Aravind Rajeswaran · Sjoerd van Steenkiste
Fri 11 Dec, 8 a.m. PST
Recent advances in deep reinforcement learning and robotics have enabled agents to achieve superhuman performance on a variety of challenging games and learn complex manipulation tasks. While these results are very promising, several open problems remain. In order to function in real-world environments, learned policies must be both robust to input perturbations and be able to rapidly generalize or adapt to novel situations. Moreover, to collaborate and live with humans in these environments, the goals and actions of embodied agents must be interpretable and compatible with human representations of knowledge. Hence, it is natural to consider how humans so successfully perceive, learn, and plan to build agents that are equally successful at solving real world tasks.
There is much evidence to suggest that objects are a core level of abstraction at which humans perceive and understand the world [8]. Objects have the potential to provide a compact, casual, robust, and generalizable representation of the world. Recently, there have been many advancements in scene representation, allowing scenes to be represented by their constituent objects, rather than at the level of pixels. While these works have shown promising results, there is still a lack of agreement on how to best represent objects, how to learn object representations, and how best to leverage them in agent training.
In this workshop we seek to build a consensus on what object representations should be by engaging with researchers from developmental psychology and by defining concrete tasks and capabilities that agents building on top of such abstract representations of the world should succeed at. We will discuss how object representations may be learned through invited presenters with expertise both in unsupervised and supervised object representation learning methods. Finally, we will host conversations and research on new frontiers in object learning.
Schedule
Fri 8:00 a.m. - 8:15 a.m.
|
Introduction
(
Introduction
)
>
|
William Agnew 🔗 |
Fri 8:15 a.m. - 9:00 a.m.
|
Keynote: Elizabeth Spelke
(
Talk
)
>
|
🔗 |
Fri 9:02 a.m. - 9:04 a.m.
|
Learning Object-Centric Video Models by Contrasting Sets
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:04 a.m. - 9:06 a.m.
|
Structure-Regularized Attention for Deformable Object Representation
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:06 a.m. - 9:08 a.m.
|
Learning Long-term Visual Dynamics with Region Proposal Interaction Networks
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:08 a.m. - 9:10 a.m.
|
Self-Supervised Attention-Aware Reinforcement Learning
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:10 a.m. - 9:12 a.m.
|
Emergence of compositional abstractions in human collaborative assembly
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:12 a.m. - 9:14 a.m.
|
Semantic State Representation for Reinforcement Learning
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:14 a.m. - 9:16 a.m.
|
Odd-One-Out Representation Learning
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:16 a.m. - 9:18 a.m.
|
Word(s) and Object(s): Grounded Language Learning In Information Retrieval
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:20 a.m. - 9:22 a.m.
|
Discrete Predictive Representation for Long-horizon Planning
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:22 a.m. - 9:24 a.m.
|
Dynamic Regions Graph Neural Networks for Spatio-Temporal Reasoning
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:26 a.m. - 9:28 a.m.
|
Dexterous Robotic Grasping with Object-Centric Visual Affordances
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:28 a.m. - 9:30 a.m.
|
Understanding designed objects by program synthesis
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:29 a.m. - 9:31 a.m.
|
Learning Embeddings that Capture Spatial Semantics for Indoor Navigation
(
Lightning
)
>
SlidesLive Video |
🔗 |
Fri 9:30 a.m. - 10:30 a.m.
|
Poster Session A in GatherTown ( Poster Session ) > link | 🔗 |
Fri 10:30 a.m. - 11:45 a.m.
|
Panel Discussion
(
Panel Discussion
)
>
|
Jessica Hamrick · Klaus Greff · Michelle A. Lee · Irina Higgins · Josh Tenenbaum 🔗 |
Fri 11:45 a.m. - 12:25 p.m.
|
Break in GatherTown link | 🔗 |
Fri 12:25 p.m. - 12:55 p.m.
|
Invited Talk: Jessica Hamrick
(
Talk
)
>
SlidesLive Video |
Jessica Hamrick 🔗 |
Fri 12:55 p.m. - 1:25 p.m.
|
Invited Talk: Irina Higgins
(
Talk
)
>
SlidesLive Video |
Irina Higgins 🔗 |
Fri 1:25 p.m. - 1:55 p.m.
|
Invited Talk: Sungjin Ahn
(
Talk
)
>
SlidesLive Video |
Sungjin Ahn 🔗 |
Fri 1:55 p.m. - 2:07 p.m.
|
Contributed Talk : A Symmetric and Object-Centric World Model for Stochastic Environments
(
Talk
)
>
SlidesLive Video |
🔗 |
Fri 2:07 p.m. - 2:19 p.m.
|
Contributed Talk : OGRE: An Object-based Generalization for Reasoning Environment
(
Talk
)
>
SlidesLive Video |
🔗 |
Fri 2:19 p.m. - 2:49 p.m.
|
Invited Talk: Wilka Carvalho
(
Talk
)
>
SlidesLive Video |
Wilka Carvalho 🔗 |
Fri 2:49 p.m. - 3:20 p.m.
|
Break in GatherTown link | 🔗 |
Fri 3:20 p.m. - 3:50 p.m.
|
Invited Talk: Renée Baillargeon
(
Talk
)
>
SlidesLive Video |
🔗 |
Fri 3:50 p.m. - 4:20 p.m.
|
Invited Talk: Dieter Fox
(
Talk
)
>
|
🔗 |
Fri 4:20 p.m. - 4:32 p.m.
|
Contributed Talk : Disentangling 3D Prototypical Networks for Few-Shot Concept Learning
(
Talk
)
>
SlidesLive Video |
🔗 |
Fri 4:32 p.m. - 4:44 p.m.
|
Contributed Talk : Deep Affordance Foresight: Planning for What Can Be Done Next
(
Talk
)
>
SlidesLive Video |
🔗 |
Fri 4:44 p.m. - 4:56 p.m.
|
Contributed talk : Estimating Mass Distribution of Articulated Objects using Non-prehensile Manipulation
(
Talk
)
>
SlidesLive Video |
🔗 |
Fri 4:56 p.m. - 6:10 p.m.
|
Panel
(
Panel
)
>
|
· Wilka Carvalho · Judith Fan · Tejas Kulkarni · Christopher Xie 🔗 |
Fri 6:10 p.m. - 6:15 p.m.
|
Concluding Remarks
(
Concluding Remarks
)
>
|
🔗 |
Fri 6:15 p.m. - 7:15 p.m.
|
Poster Session B in GatherTown ( Poster Session ) > link | 🔗 |