Beyond the symbols vs signals debate

28 - 29 October 2024 09:00 - 17:00 The Royal Society Free Watch online
Book event
Background design of human profile and abstract elements on the subject of spirituality, science, creativity and the mind

Discussion meeting organised by Professor Marta Kwiatkowska FRS, Professor Peter Dayan FRS, Professor Tom Griffiths and Professor Doina Precup.

Building artificial intelligence systems that can emulate human intelligence will need to draw on the complementary strengths of machine learning and compositional reasoning, and take advantage of symbolic knowledge as well as raw signals. This meeting aims to advance our scientific understanding of how human learning, reasoning and cognition can be brought to bear on the engineering foundations of AI.

Poster session

There will be a poster session on Monday 28 October. If you would like to present a poster, please submit your proposed title, abstract (up to 200 words), author list, and the name of the proposed presenter and institution to the Scientific Programmes team no later than 17 September 2024. Please include the text 'Poster submission- Symbols vs signals' in the email subject line. 

Attending the meeting

This event is free to attend and intended for researchers in the field.

Both virtual and in-person attendance is available, but advance registration is essential.

Lunch is available on both days of the meeting and is optional. There are plenty of places to eat nearby if you would prefer to purchase food offsite. Participants are welcome to bring their own lunch to the meeting.

Enquiries: contact the Scientific Programmes team.

Organisers

  • Professor Marta Kwiatkowska FRS, University of Oxford, UK

    Professor Marta Kwiatkowska FRS, University of Oxford, UK

    Marta Kwiatkowska is Professor of Computing Systems and Fellow of Trinity College, University of Oxford. Her research is concerned with developing modelling and analysis methods for complex systems, such as those arising in computer networks, electronic devices and biological organisms. She is known for fundamental contributions to the theory and practice of model checking for probabilistic systems and is currently focusing on safety and robustness of automated decision-making in artificial intelligence. Kwiatkowska led the development of the PRISM model checker, which has been adopted in diverse fields, including security, robotics, healthcare and DNA computing. She has been the recipient of two ERC Advanced Grants, the Royal Society Milner Award, the BCS Lovelace Medal, and received an honorary doctorate from KTH Royal Institute of Technology. She holds the title of Professor awarded by the President of Poland and is a Fellow of the Royal Society, Fellow of ACM, Member of Academia Europea and American Academy of Arts and Sciences.

  • Professor Peter Dayan FRS, Max Planck Institute of Intelligent Systems, Germany

    Professor Peter Dayan FRS, Max Planck Institute for Biological Cybernetics, Germany

    Peter Dayan is a Director at the Max Planck Institute for Biological Cybernetics and a Professor at the University of Tübingen. His interests include affective decision making, neural reinforcement learning and computational psychiatry.

  • Professor Tom Griffiths, Princeton University, USA

    Professor Tom Griffiths, Princeton University, USA

    Tom Griffiths is the Henry R Luce Professor of Information Technology, Consciousness and Culture in the Departments of Psychology and Computer Science at Princeton University. His research explores connections between human and machine learning, using ideas from statistics and artificial intelligence to understand how people solve the challenging computational problems they encounter in everyday life. Tom completed his PhD in Psychology at Stanford University in 2005 and taught at Brown University and the University of California, Berkeley before moving to Princeton. He has received awards for his research from organisations ranging from the American Psychological Association to the National Academy of Sciences and is a co-author of the book Algorithms to Live By, introducing ideas from computer science and cognitive science to a general audience.

Schedule

Chair

Professor Marta Kwiatkowska FRS, University of Oxford, UK

Professor Marta Kwiatkowska FRS, University of Oxford, UK

09:00-09:10 Welcome by the Royal Society and lead organiser
09:10-09:15 Chair's introduction: Neurosymbolic
Professor Marta Kwiatkowska FRS, University of Oxford, UK

Professor Marta Kwiatkowska FRS, University of Oxford, UK

09:15-09:45 Educability

We seek to define the capability that has enabled humans to develop the civilisation we have, and that distinguishes us from other species. For this it is not enough to identify a distinguishing characteristic - we want a capability that is also explanatory of humanity's achievements. "Intelligence" does not work here because we have no agreed definition of what intelligence is or how an intelligent entity behaves. We need a concept that is behaviourally better defined. The definition will need to be computational in the sense that the expected outcomes of exercising the capability need to be both specifiable and computationally feasible. This formulation is related to the goals of AI research but is not synonymous with it, leaving out the many capabilities we share with other species.

We make a proposal for this essential human capability, which we call "educability." It synthesises abilities to learn from experience, to learn from others, and to chain together what we have learned in either mode and apply that to particular situations. It starts with the now standard notion of learning from examples, as captured by the Probably Approximately Correct model and used in machine learning. The ability of Large Language Models learning from examples to generate smoothly flowing prose lends encouragement to this approach. The basic question then is how to extend this approach to encompass broader human capabilities beyond learning from examples. This is what the educability notion aims to answer.

Professor Leslie Valiant FRS< Harvard University, USA

Professor Leslie Valiant FRS< Harvard University, USA

09:45-10:00 Discussion
10:00-10:30 How to make logics neurosymbolic

Neurosymbolic AI (NeSy) is regarded as the third wave in AI. It aims at combining knowledge representation and reasoning with neural networks. Numerous approaches to NeSy are being developed and there exists an `alphabet-soup' of different systems, whose relationships are often unclear. I will discuss the state-of-the art in NeSy and argue that there are many similarities with statistical relational AI (StarAI).

Taking inspiring from StarAI, and exploiting these similarities, I will argue that Neurosymbolic AI = Logic + Probability + Neural Networks.  I will also provide a recipe for developing NeSy approaches: start from a logic, add a probabilistic interpretation, and then turn neural networks into `neural predicates'.

Probability is interpreted broadly here, and is necessary to provide a quantitative and differentiable component to the logic. At the semantic and the computation level, one can then combine logical circuits (aka proof structures) labelled with probability, and neural networks in computation graphs. 

I will illustrate the recipe with NeSy systems such as DeepProbLog, a deep probabilistic extension of Prolog, and DeepStochLog, a neural network extension of stochastic definite clause grammars (or stochastic logic programs).

Professor Luc De Raedt, KU Leuven and Örebro University, Belgium

Professor Luc De Raedt, KU Leuven and Örebro University, Belgium

10:30-10:45 Discussion
10:45-11:15 Break
11:15-11:45 Planning, reasoning, and generalisation in deep learning

What do we need to build artificial agents which can reason effectively and generalise to new situations? An oft-cited claim, both in cognitive science and in machine learning, is that a key ingredient for reasoning and generalisation is planning with a model of the world. In this talk, Dr Hamrick will evaluate this claim in the context of model-based reinforcement learning, presenting evidence that demonstrates the utility of planning for certain classes of problems (e.g. in-distribution learning and procedural generalisation in reinforcement learning), as well as evidence that planning is not a silver bullet for out-of-distribution generalisation. In particular, generalisation performance is limited by the generalisation abilities of the individual components required for planning (e.g., the policy, reward model, and world model), which in turn are dependent on the diversity of data those components are trained on. Moreover, generalisation is strongly dependent on choosing the appropriate level of abstraction. These concerns may be partially addressed by leveraging new state-of-the-art foundation models, which are trained on both an unprecedented breadth of data and at a higher level of abstraction than before.

Dr Jessica Hamrick, Google DeepMind, UK

Dr Jessica Hamrick, Google DeepMind, UK

11:45-12:00 Discussion
12:00-12:30 The role of rationality in modern AI

The classical approach to AI was to design systems that were rational at run-time: they had explicit representations of beliefs, goals, and plans and ran inference algorithms, online, to select actions. The rational approach was criticised (by the behaviourists) and modified (by the probabilists) but persisted in some form. Now the overwhelming success of the connectionist approach in so many areas presents evidence that the rational view may no longer have a role to play in AI. This talk examines this question from several perspectives, including whether the rationality is present at design-time and/or at run-time, and whether systems with run-time rationality might be useful from the perspectives of computational efficiency, cognitive modelling and safety. It will present some current research focused on understanding the roles of learning in runtime-rational systems with the ultimate aim of constructing general-purpose human-level intelligent robots.

Professor Leslie Pack Kaelbling, Massachusetts Institute of Technology, USA

Professor Leslie Pack Kaelbling, Massachusetts Institute of Technology, USA

12:30-12:45 Discussion

Chair

Professor Peter Dayan FRS, Max Planck Institute of Intelligent Systems, Germany

Professor Peter Dayan FRS, Max Planck Institute for Biological Cybernetics, Germany

13:30-13:35 Chair's introduction: Neuroscience
Professor Peter Dayan FRS, Max Planck Institute for Biological Cybernetics, Germany

Professor Peter Dayan FRS, Max Planck Institute for Biological Cybernetics, Germany

13:35-14:05 Representing the future

To flexibly adapt to new situations, our brains must understand the regularities in the world, but also in our own patterns of behaviour. A wealth of findings is beginning to reveal the algorithms we use to map the outside world. In contrast, the biological algorithms that map the complex structured behaviours we compose to reach our goals remain enigmatic. Here we reveal a neuronal implementation of an algorithm for mapping abstract behavioural structure and transferring it to new scenarios. We trained mice on many tasks which shared a common structure organising a sequence of goals but differed in the specific goal locations. Animals discovered the underlying task structure, enabling zero-shot inferences on the first trial of new tasks. The activity of most neurons in the medial Frontal cortex tiled progress-to-goal, akin to how place cells map physical space. These “goal-progress cells” generalised, stretching and compressing their tiling to accommodate different goal distances. In contrast, progress along the overall sequence of goals was not encoded explicitly. Instead, a subset of goal-progress cells was further tuned such that individual neurons fired with a fixed task-lag from a particular behavioural step. Together these cells implemented an algorithm that instantaneously encoded the entire sequence of future behavioural steps, and whose dynamics automatically retrieved the appropriate action at each step. These dynamics mirrored the abstract task structure both on-task and during offline sleep. Our findings suggest that goal-progress cells in the medial frontal cortex may be elemental building blocks of schemata that can be sculpted to represent complex behavioural structures.

Professor Timothy Behrens FRS, University of Oxford and University College London, UK

Professor Timothy Behrens FRS, University of Oxford and University College London, UK

14:05-14:20 Discussion
14:20-14:50 Language is distinct from thought in the human brain

Dr Fedorenko seeks to understand how humans understand and produce language, and how language relates to, and works together with, the rest of human cognition. She will discuss the ‘core’ language network, which includes left-hemisphere frontal and temporal areas, and show that this network is ubiquitously engaged during language processing across input and output modalities, strongly interconnected, and causally important for language. This language network plausibly stores language knowledge and supports linguistic computations related to accessing words and constructions from memory and combining them to interpret (decode) or generate (encode) linguistic messages. Importantly, the language network is sharply distinct from higher-level systems of knowledge and reasoning. First, the language areas show little neural activity when individuals solve math problems, infer patterns from data, or reason about others’ minds. And second, some individuals with severe aphasia lose the ability to understand and produce language but can still do math, play chess, and reason about the world. Thus, language does not appear to be necessary for thinking and reasoning. Human thinking instead relies on several brain systems, including the network that supports social reasoning and the network that supports abstract formal reasoning. These systems are sometimes engaged during language use—and thus have to work with the language system—but are not language-selective. Many exciting questions remain about the representations and computations in the systems of thought and about how the language system interacts with these higher-level systems. Furthermore, the sharp separation between language and thought in the human brain has implications for how we think about this relationship in the context of AI models, and for what we can expect from neural network models trained solely on linguistic input with the next-word prediction objective.

Professor Evelina Fedorenko, Massachusetts Institute of Technology, USA

Professor Evelina Fedorenko, Massachusetts Institute of Technology, USA

14:50-15:05 Discussion
15:05-15:30 Break
15:30-16:00 Neither nature nor nurture: the semiotic infrastructure of symbolic reference

Use of the symbol concept suffers from a conflation of two interpretations: a) a conventional sign vehicle (an alphanumeric character), and b) a conventional reference relationship (word meaning). Both are often mischaracterised in terms of "arbitrarity," a negative attribute. When your computer begins randomly displaying characters (a) on your screen, they are generally interpreted as indications of malfunction (or the operation of a "viral" algorithm). And yet when LLMs print out strings of characters that are iconic of interpretable sentences, we assume (b) that they are more than mere icons and indices of an algorithmic (aka mechanistic) process. This begs the question of what distinguishes symbolic interpretation from iconic and indexical interpretation and how they are related. Conventional relations are not just "given," however, they must be acquired. As a result, they are dependent on prior non-conventional referential relations (i.e. iconic and indexical interpretive processes) to extrinsically "ground" the reference of these intrinsically "ungrounded" sign vehicles. This semiotic infrastructure exemplifies the hierarchic complexity of symbolic reference, why it is cognitively difficult for non-humans, and hints at the special neurological architecture that aids human symbolic cognition. It also is relevant for understanding the difference between how humans and generative AI systems produce and process the tokens used as sign vehicles. So far, LLMs and their generative cousins are structured by token-token iconic and indexical relations only (though of extremely high dimensionality), not externally grounded by iconic and indexical pragmatic relations, even though the token-token relations of the training data have been.

Professor Terrence W Deacon, University of California, USA

Professor Terrence W Deacon, University of California, USA

16:15-16:45 Neural models of compositional learning
Professor Alexandre Pouget, University of Geneva, Switzerland

Professor Alexandre Pouget, University of Geneva, Switzerland

16:45-17:00 Discussion
09:00-09:05 Chair's introduction: Reinforcement learning
09:05-09:35 Meta-learning as bridging the neuro-symbolic gap in AI

Meta-learning, the ability to acquire and utilise prior knowledge to facilitate new learning, is a hallmark of human and animal cognition. This capability is also evident in deep reinforcement learning agents and large language models (LLMs). While recurrent neural networks have offered insights into the neural underpinnings of learning, understanding how LLMs, trained on vast amounts of human-generated text and code, achieve rapid in-context learning remains a challenge. The inherent structure present in these training data sources—reflecting symbolic knowledge embedded in language and cultural artifacts—potentially plays a crucial role in enabling LLMs to generalise effectively. Therefore, examining the role of structured data in large-scale model training through a cognitive science lens offers crucial insights into how these models acquire and generalise knowledge, mirroring aspects of human learning. 

This talk will discuss how these findings not only deepen our understanding of deep learning models but also offer potential avenues for integrating machine learning and symbolic reasoning, through the lens of meta-learning and cognitive science. Insights from meta-learning research can inform the development of embodied AI agents, such as those in the Scalable, Instructable, Multiworld Agent (SIMA) project, by incorporating structured knowledge representations and meta-learning capabilities to potentially enhance their ability to follow instructions, generalise to novel tasks, and interact more effectively within many complex 3D environments.

Dr Jane X Wang, Google DeepMind, UK

Dr Jane X Wang, Google DeepMind, UK

09:35-09:50 Discussion
09:50-10:20 Dynamic deep learning

Deep learning and large language models have dramatically shifted the conversation about Signals vs Symbols in favour of numerical methods. Nevertheless, current deep learning methods are limited; they have great difficulty learning during their normal operation. In this talk, Sutton argues that this is not an inherent limitation of neural networks, but just of the algorithms currently used, and he proposes new neural-network algorithms specifically designed for continual learning.

Professor Richard S Sutton FRS, University of Alberta, Canada

Professor Richard S Sutton FRS, University of Alberta, Canada

10:20-10:35 Discussion
10:35-11:05 Break
11:35-11:50 Discussion
11:50-12:20 Associate Professor Emma Brunskill, Stanford University, USA
Professor Emma Brunskill, Stanford University, USA

Professor Emma Brunskill, Stanford University, USA

12:20-12:35 Discussion

Chair

Professor Tom Griffiths, Princeton University, USA

Professor Tom Griffiths, Princeton University, USA

13:30-13:35 Chair's introduction: Alignment
13:35-14:05 Professor Sheila McIlraith, University of Toronto and Vector Institute, Canada
Professor Sheila McIlraith, University of Toronto and Vector Institute, Canada

Professor Sheila McIlraith, University of Toronto and Vector Institute, Canada

14:05-14:20 Discussion
14:50-15:05 Discussion
15:05-15:35 Break
15:35-16:05 The emerging science of benchmarks

Benchmarks have played a central role in the progress of machine learning research since the 1980s. Although there's much researchers have done with them, we still know little about how and why benchmarks work. In this talk, I will trace the rudiments of an emerging science of benchmarks through selected empirical and theoretical observations. Looking back at the ImageNet era, I'll discuss what we learned about the validity of model rankings and the role of label errors. Looking ahead, I'll talk about new challenges to benchmarking and evaluation in the era of large language models. The results we'll encounter challenge conventional wisdom and underscore the benefits of developing a science of benchmarks.

Professor Moritz Hardt, Max Planck Institute of Intelligent Systems, Germany

Professor Moritz Hardt, Max Planck Institute of Intelligent Systems, Germany

16:05-16:20 Discussion
16:20-17:00 Panel discussion