User profiles for Adam Santoro
Adam SantoroGoogle DeepMind Verified email at google.com Cited by 16199 |
A simple neural network module for relational reasoning
Relational reasoning is a central component of generally intelligent behavior, but has
proven difficult for neural networks to learn. In this paper we describe how to use Relation …
proven difficult for neural networks to learn. In this paper we describe how to use Relation …
Meta-learning with memory-augmented neural networks
A Santoro, S Bartunov, M Botvinick… - International …, 2016 - proceedings.mlr.press
Despite recent breakthroughs in the applications of deep neural networks, one setting that
presents a persistent challenge is that of" one-shot learning." Traditional gradient-based …
presents a persistent challenge is that of" one-shot learning." Traditional gradient-based …
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models
Language models demonstrate both quantitative improvement and new qualitative
capabilities with increasing scale. Despite their potentially transformative impact, these new …
capabilities with increasing scale. Despite their potentially transformative impact, these new …
Relational inductive biases, deep learning, and graph networks
Artificial intelligence (AI) has undergone a renaissance recently, making major progress in
key domains such as vision, language, control, and decision-making. This has been due, in …
key domains such as vision, language, control, and decision-making. This has been due, in …
Backpropagation and the brain
During learning, the brain modifies synapses to improve behaviour. In the cortex, synapses
are embedded within multilayered networks, making it difficult to determine the effect of an …
are embedded within multilayered networks, making it difficult to determine the effect of an …
Data distributional properties drive emergent in-context learning in transformers
Large transformer-based models are able to perform in-context few-shot learning, without
being explicitly trained for it. This observation raises the question: what aspects of the training …
being explicitly trained for it. This observation raises the question: what aspects of the training …
Relational recurrent neural networks
Memory-based neural networks model temporal data by leveraging an ability to remember
information for long periods. It is unclear, however, whether they also have an ability to …
information for long periods. It is unclear, however, whether they also have an ability to …
Assessing the scalability of biologically-motivated deep learning algorithms and architectures
S Bartunov, A Santoro, B Richards… - Advances in neural …, 2018 - proceedings.neurips.cc
The backpropagation of error algorithm (BP) is impossible to implement in a real brain. The
recent success of deep networks in machine learning and AI, however, has inspired …
recent success of deep networks in machine learning and AI, however, has inspired …
Relational deep reinforcement learning
We introduce an approach for deep reinforcement learning (RL) that improves upon the
efficiency, generalization capacity, and interpretability of conventional approaches through …
efficiency, generalization capacity, and interpretability of conventional approaches through …
Hyperbolic attention networks
We introduce hyperbolic attention networks to endow neural networks with enough capacity
to match the complexity of data with hierarchical and power-law structure. A few recent …
to match the complexity of data with hierarchical and power-law structure. A few recent …