@SimonsInstituteTOC
  @SimonsInstituteTOC
Simons Institute | What can be passively learned about causality? @SimonsInstituteTOC | Uploaded 2 months ago | Updated 7 hours ago
Andrew Lampinen (Google DeepMind)
https://simons.berkeley.edu/talks/andrew-lampinen-google-deepmind-2024-06-25
Understanding Higher-Level Intelligence from AI, Psychology, and Neuroscience Perspectives

What could language models learn about causality and experimentation from their passive training? Observational learning is inherently limited. However, in this talk I will draw an important distinction between observational and passive learning, and argue that LMs learn passively, but from interventional data. I will then show empirically that agents trained via passive imitation on expert interventional data can learn generalizable causal strategies that they can apply at test time to discover causal structures never seen in training. This is possible even in a complex environment with high-dimensional observations, with the support of natural language explanations. Furthermore, explanations can even allow passive learners to generalize out-of-distribution from perfectly-confounded training data. Finally, I'll show that language models, trained only on passive next-word prediction, can generalize causal intervention strategies from a few-shot prompt containing examples of experimentation, together with explanations and reasoning. These results highlight the surprising power of passive learning of active causal strategies, and may help to understand the behaviors and capabilities of language models. I will close by reflecting on some of the open questions in how to enable AI to use explanations in a more human-like way.
What can be passively learned about causality?Computation with sequences of neuronal assembliesPanel: Reactive Synthesis - Open Questions and ChallengesDynamic Matching and (Ordered) Ruzsa-Szemerédi Graphs: Towards Constructive Matching SparsifiersTalk By Tal HermanWhistle variability and social acoustic interactions in bottlenose dolphinsThe Role of Prior Data in Rapid Learning of Motor SkillsData is as Data Does: The Influence of Computation on InferenceThe approximate structure of triangle-free graphsWhat neural machinery is needed for language acquisition (Virtual Talk)Faces, Flukes, Fins, and Flanks: How Multispecies Re-ID Models are Transforming Our Approach to AI..What Does Machine Learning Have to Offer Mathematics? | Theoretically Speaking

What can be passively learned about causality? @SimonsInstituteTOC

SHARE TO X SHARE TO REDDIT SHARE TO FACEBOOK WALLPAPER