Geoffrey Hinton (Google Scientific Advisor) – Does the brain do backpropogation? (May 2019)


Chapters

00:00:00 Neuroscience and AI: The Impact of Geoff Hinton
00:05:13 Understanding Deep Learning: A Paradigm Shift in Artificial Intelligence
00:10:07 Deep Learning: From Theory to Applications
00:15:43 Neural Nets: Imaginative Story Generation and Contextual Understanding
00:20:17 Fast Weights and Temporary Context
00:22:50 Neural Networks and Brain Function
00:31:01 Neural Networks: Noise and Generalization
00:34:05 Neural Firing Rates and Error Derivatives in Neural Networks
00:37:21 Autoencoders and Spiking Neurons Learning Rule
00:41:35 Neurons Represent Error Derivatives
00:44:11 Neuroscience and Innate Knowledge: A Critique
00:52:06 Evolutionary Pressure in AI: From Learning to Hardwiring
00:59:49 Bridging Brains and Computers: An Exploration of Computational Neuroscience
01:02:58 Exploring Neural Networks, Hierarchy, and Statistical Intuition in AI
01:16:27 Deep Learning Challenges and Opportunities
01:18:52 Deep Learning Dropout Techniques

Abstract

Bridging Neuroscience and AI: Insights from Geoffrey Hinton and the Evolution of Deep Learning

The intersection of neuroscience and artificial intelligence (AI) is an increasingly fertile ground for scientific discovery, as highlighted by the recent public lecture by Paul Franklin, a neuroscientist at SickKids and Program Chair for the Canadian Neuroscience Meeting. Central to this discourse is the work of Geoffrey Hinton, a pivotal figure in the field of deep learning. Hinton’s contributions, including the development of neural networks, the backpropagation algorithm, and innovations like dropout regularization, have not only advanced AI but also offered insights into understanding the human brain. This article delves into the key concepts and debates shaping this intersection, exploring the parallels between brain functions and computational models, the evolution of deep learning, and Hinton’s vision for the future convergence of neuroscience and AI.

Introduction to Neuroscience and AI Interface

The traditional curtain-raiser for the Canadian Neuroscience Meeting, this year’s public lecture, spotlights the vibrant interface between neuroscience and AI. Toronto, a hub for AI research, has been significantly influenced by Geoffrey Hinton’s work in neural networks and machine learning.

Geoff Hinton: A Trailblazer in the Convergence of Neuroscience and AI

Geoff Hinton is a pioneering figure in the field of deep learning, often referred to as the godfather of the field. With a BA in experimental psychology from Cambridge and a PhD in artificial intelligence from Edinburgh, Hinton’s career has been dedicated to neural networks as models of the mind and as tools for artificial intelligence. His tenure-track position at Carnegie Mellon in the 1980s was followed by a move to the University of Toronto due to ethical objections to DARPA funding. Throughout the 1990s, Hinton continued to champion neural networks and machine learning, establishing the Gatsby Computational Neuroscience Unit at University College London in 1998 before returning to Toronto in 2001. Hinton’s contributions have earned him widespread recognition, including the prestigious Turing Award, the Order of Canada, and a distinguished fellowship from the Canadian Institute for Advanced Research. His work has garnered over 270,000 citations, surpassing those of Einstein, Ramon y Cajal, and Alan Turing combined. Hinton’s influence on other scientists is remarkable, with many of his graduate students and postdocs becoming prominent figures in artificial intelligence. His insights have reshaped the way computer scientists and neuroscientists approach their respective fields. Hinton’s groundbreaking work in deep learning has revolutionized artificial intelligence and has the potential to greatly impact neuroscience in the coming years.

Geoffrey Hinton recently presented a neural net-generated story that he claims showcases its impressive text generation abilities. The story revolves around unicorns in Argentina speaking English and involves fictional details like “Dr. Jorge Perez” and “the University of La Paz.” Despite these inaccuracies, the story demonstrates the neural net’s ability to remember and incorporate initial context, unlike recurrent neural nets. This phenomenon raises questions about the net’s understanding of magic realism, as seen in its inclusion of unicorns and the reference to a “lost race” in South America. The story also highlights the net’s strengths and limitations, evident in its mix-up of “dialect” and “dialectic” and the erroneous attribution of unicorns to Argentina instead of Bolivia. Nevertheless, Hinton predicts that a larger neural net with 50 billion connections, trained on Google’s latest cloud hardware, will produce even more impressive and accurate results.

Deep Learning: A Paradigm Shift

Deep learning represents a paradigm shift from traditional reasoning and logic-based AI to models that emulate brain connections. This shift marks a move from programmed instructions to learning from examples, with neural networks at the forefront. Key components of these networks include simplified neuron models, weight training through backpropagation, and the application of neural nets in fields like speech and image recognition, and machine translation.

History of Deep Learning and AI

Deep learning emerged as one of two AI paradigms, the other being reasoning and logic. The approach of adapting connections between artificial neurons eventually yielded better results than symbolic AI. This led to deep learning’s emphasis on example-based programming, allowing computers to learn without explicit instructions. Neural networks have demonstrated capabilities in tasks like image captioning, which posed challenges for conventional AI.

Simplified Model of a Neuron

At the core of neural networks is a simplified model of a neuron. These artificial neurons receive input, multiply it by weights, and sum it up (depolarization), resulting in an output proportional to the input. Networks are created by connecting neurons in layers.

Training Neural Networks with Backpropagation

The training of neural networks involves an iterative process called backpropagation. It begins with random weights, and inputs are presented with desired outputs. One weight is slightly changed, and inputs are shown again to check for performance improvement. Weight changes that improve performance are retained. Backpropagation’s advantages lie in its speed, predictability of weight change effects, and efficient calculation of the cost function gradient with respect to the weights.

Advantages of Neural Networks

Neural networks offer several advantages over traditional AI approaches. Their ability to adapt and learn from examples makes them suitable for complex tasks like image recognition and language translation. Neural networks’ generalization capabilities allow them to perform well on tasks not explicitly encountered during training. Additionally, neural networks can be trained incrementally, making them suitable for online learning.

Innovations and Challenges in Neural Network Models

Hinton’s work introduced several computational innovations, such as transformers and techniques to prevent overfitting, significantly enhancing neural network performance. Moreover, large-scale neural networks with billions of connections demonstrate impressive text generation capabilities, advancing the field’s understanding of machine learning. However, the concepts of Hebbian learning, supervision signals, real-valued activities, and dropout regularization present both opportunities and challenges in aligning AI models more closely with brain functions.

New Algorithm for Weight Adjustment

Hinton’s work on neural network models includes a new algorithm for weight adjustment. This algorithm allows for simultaneous weight adjustment based on output improvement, with incremental updates to optimize output quality.

Online Algorithm for Efficient Learning

An online algorithm developed by Hinton’s team updates weights after each case or small batch instead of waiting to process all examples. This approach has proven effective, even for large datasets.

Neural Nets for Complex Tasks

Contrary to initial predictions, neural nets with random weights can learn complex tasks such as language translation and image recognition. In 2009, researchers demonstrated improved speech recognition using neural nets. Neural nets also outperformed traditional computer vision techniques in image recognition tasks in 2012, becoming the standard for object recognition. In 2014, a neural net was developed for translating fragments of words between languages, surpassing symbolic AI in translation quality and leading to its adoption by Google for machine translation.

Factors Contributing to Neural Net Advancements

The advancements in neural nets can be attributed to increased computational speed, larger datasets, clever tricks and techniques, and the availability of powerful graphics processing units (GPUs).

Examples of Neural Net Capabilities

OpenAI trained a neural net with 1.5 billion learnable connections on billions of words of English text, enabling it to predict the next word in a sequence based on probabilities. By feeding predictions back into the net and prompting it to generate more text, researchers can uncover the net’s underlying knowledge and beliefs.

Further Innovations

Transformers and improved overfitting prevention techniques are among the latest advancements in neural networks.

Hebbian Learning and Fast Weights

Hinton proposes the concept of fast weights as a mechanism for storing long-term temporal context in neural networks. This approach, which involves temporary weight changes decaying over time, offers a potential explanation for short-term memory storage in the brain. It aligns with the idea that backpropagation could be a plausible learning algorithm for the brain, addressing challenges related to supervision signals, real-valued activities, reciprocal connections, and the number of parameters.

Storing Long-term Memory

Hinton challenges the traditional view of storing long-term memory as copies of neural activity patterns. Instead, he proposes that long-term memory is stored in temporary changes in synaptic weights, or “fast weights.” These fast weights are constantly modified through Hebbian learning, allowing the brain to store and access information over long periods of time.

Location of Short-term Memory

Hinton suggests that short-term memory is stored in the temporary changes to synaptic weights, rather than in the activities of neurons themselves. This allows for a much larger capacity for memory storage compared to using individual neurons. The temporary changes in synaptic weights decay over time, providing a natural mechanism for forgetting.

Bridging the Gap Between Brains and Computers

Hinton’s goal is to bridge the gap between brain computation and computer simulation. This involves making artificial neural nets more biological, contrasting with conventional AI’s approach. The discussion includes the utility of Hebbian synapses in AI, reinforcement learning with dopamine signals, and the importance of hierarchies in processing complex information.

Evolutionary Discovery Through Learning and the Baldwin Effect

A key aspect of Hinton’s theory is the Baldwin effect, where learning leads to genetic pressure to hardwire certain solutions. This perspective challenges the traditional view of innate knowledge being solely a product of evolution, suggesting that learning plays a crucial role in shaping the brain’s innate knowledge.

The Future of Neuroscience and AI

Looking forward, Hinton envisions a new cognitive science focusing on general intelligence through brain-centric models, predicting a convergence between neuroscience and AI. This vision includes the importance of powerful computing capabilities, alongside clever algorithms and network architectures, reflecting the potential of the Baldwin effect in AI.

Further Supplemental Information Updates:

Noise and Generalization

* Adding noise to a large model, like the brain, improves generalization compared to a small model with no noise.

* Poisson neurons with a firing rate and added noise lead to better generalization.

Dropout and Backpropagation

* Dropout models, where neurons are randomly dropped out during training, are a form of noise addition that improves generalization.

* These models are trained using backpropagation, making them compatible with the backpropagation algorithm.

Neurons and Feature Representation

* Neurons in a neural network represent the presence of features in the current input.

* The same neuron cannot represent both the feature value and the error derivative simultaneously.

Backpropagation Using Firing Rates

* Neurons communicate through firing rates, which are stochastically communicated by spikes.

* The rate of change of the firing rate represents the error derivative.

* This allows for positive and negative derivatives without changing the signs of synapses.

Learning with Autoencoders

* Each layer of the autoencoder learns to reconstruct the activity in the layer below.

* A stack of autoencoders is built, allowing each layer to reconstruct the activity in the layer below.

Top-Down Passes

* Two top-down passes are performed.

* The first pass reconstructs activities everywhere based on the predicted output.

* The second pass changes the output to be more like the desired output.

Backpropagation

* The difference between the two reconstructed activities is the signal needed for backpropagation.

* The learning rule involves changing the synapse by the presynaptic activity and the rate of change of the postsynaptic activity.

Spiking Neurons

* Spiking neurons represent underlying firing rates that change.

* The learning rule for spiking neurons involves measuring the rate of change of the postsynaptic firing rate.

Derivative Filter

* The learning rule for spiking neurons is a derivative filter that measures the rate of change of the postsynaptic firing rate.

Temporal Derivatives

* Using the rate of change of a neuron to represent an error derivative prevents the use of temporal derivatives to communicate the temporal derivatives of what the neuron represents.

Conclusion

Geoffrey Hinton’s contributions and perspectives offer a unique lens through which to view the confluence of neuroscience and AI. His work not only propels the development of deep learning but also provides critical insights into the brain’s functioning. As we stand at the crossroads of these two fields, Hinton’s vision of a future where neuroscience and AI converge promises to unlock further mysteries of the human brain and the potential of artificial intelligence.


Notes by: oganesson