Jeff Dean (Google Senior Fellow) – Deep Learning to Solve Challending Problems | ETH Zurich (Sep 2018)
Chapters
00:00:15 Deep Learning: Solving Challenging Problems and Grand Engineering Challenges
Introduction: Deep learning, a modern rebranding of artificial neural networks, is experiencing a surge in interest and growth. Google’s research approach includes basic research, infrastructure building, product collaboration, and emerging area exploration.
Machine Learning Field Explosion: Machine learning archive papers are growing faster than Moore’s law, indicating the rapid advancement of ideas. Deep learning, with its simple trainable units, has gained significant attention.
Early Excitement and Limitations of Neural Networks: Neural nets faced challenges in the past due to limited computational power. The current availability of massive compute has enabled impressive results with artificial neural networks.
Capabilities of Neural Networks: Computer vision: Neural nets can categorize images, generate captions, and transcribe audio. Language translation: Neural nets can translate text from one language to another. Progress in Computer Vision: ImageNet contest showcases remarkable improvement in object recognition accuracy.
U.S. National Academy of Engineering Grand Challenges: The talk focuses on applying deep learning to address the U.S. National Academy of Engineering’s grand challenges for the 21st century.
00:07:25 Machine Learning's Role in Solving Global Issues
Machine Learning’s Role in Addressing Global Challenges: Machine learning can significantly contribute to solving global challenges, such as improving urban infrastructure, promoting healthier living, and mitigating climate change.
Self-Driving Cars: A Revolution in Transportation: Self-driving cars are being actively tested and developed, with Waymo leading the way in real-world trials. These cars use sensor data to understand their surroundings, predict object movements, and make safe driving decisions.
Vision and Robotics: Vision is crucial for robots to navigate and interact with their environment. Researchers are developing hardware labs of robotic arms that can experiment and pool their collective experience to improve their ability to grasp objects and perform various tasks.
Grasping as a Fundamental Task: Grasping is a fundamental skill for robots, as it enables them to manipulate objects and perform complex actions. Researchers are focusing on improving robots’ ability to grasp objects effectively.
00:10:09 Machine Learning in Healthcare and Robotics
Robotics: Researchers explored supervised learning with parallel robotic arms, enabling self-supervised grasping tasks. Grasp success rates improved from mid-60s to 96% for unseen objects through algorithm enhancements. Learning approaches are being explored to give robots new skills by observing videos, demonstrating promising results in emulating human actions.
Healthcare: Machine learning has tremendous potential in healthcare for improved decision-making. In medical imaging, computer vision techniques aid in diagnostic tasks like diabetic retinopathy screening. An ML model achieved par or better performance compared to US board-certified ophthalmologists in grading retinal images. Advanced models can identify patterns in retinal images that ophthalmologists may not be aware of, providing new insights into patient health. Machine learning models can predict future outcomes for patients based on electronic medical record data, aiding in decision-making and treatment planning. Applications of these ML models include Smart Reply in Gmail, machine translation, and predicting various aspects of patient care using de-identified medical records.
00:20:40 Scientific Discovery Acceleration Through Machine Learning
Predicting Patient Mortality: Traditional methods use 20 hand-designed features to predict mortality risk. Using all data in the medical record can improve prediction accuracy. Earlier assessment of patients’ medical risk is possible with this approach.
Accelerating Quantum Chemistry Simulations: Quantum chemists use computationally intensive simulations to predict molecule properties. Training a neural network with simulation data can achieve indistinguishable accuracy. The neural network is 300,000 times faster than the original simulator.
Exploiting Simulators for Scientific Discovery: Replacing simulation components with neural networks can lead to significant speedups. Screening large datasets becomes feasible with faster simulations.
Engineering Tools for Scientific Discovery: The grand challenge aims to develop tools that facilitate scientific discovery. The speaker suggests exploring the use of neural networks to accelerate simulations.
00:24:10 Automating Machine Learning with TensorFlow
TensorFlow: TensorFlow is an open-source machine learning framework developed by Google for expressing machine learning research ideas and deploying them in real-world environments. It has gained significant interest and adoption, with over 16 million downloads and a vibrant community contributing to its development. TensorFlow is flexible and runs on various platforms, including mobile phones, desktops, and data centers.
Democratizing Machine Learning: There is a shortage of machine learning experts compared to the vast number of organizations that could benefit from machine learning applications. Traditional machine learning problem-solving involves human experts making numerous decisions about modeling, data preparation, and experimentation. To make machine learning more accessible, researchers are exploring ways to automate some of these tasks and provide data-plus-computation solutions.
Neural Architecture Search: Neural architecture search involves using a model-generating model to automatically design neural network architectures. The model-generating model creates descriptions of network structures, which are then trained on a specific problem. The accuracy of these trained models is used as feedback to guide the model-generating model towards better architectures. This iterative process leads to the development of increasingly accurate models over time.
Results and Insights: Neural architecture search can produce models with unique connectivity structures that are not typically designed by human experts. These “organic style skip connections” allow information to flow more directly from the input to the output, improving model performance. Experiments on ImageNet, a popular image classification dataset, demonstrate the effectiveness of neural architecture search in achieving state-of-the-art results.
00:30:56 Advances in AutoML and Neural Architecture Search
AutoML’s Achievements: AutoML has revolutionized machine learning and computer vision, delivering state-of-the-art results through collaborative efforts of top researchers. Its impact is evident across various applications, from high-accuracy models to cost-efficient models suitable for resource-constrained environments.
AutoML in Google’s Cloud Products: Google has integrated AutoML into its cloud products, enabling industries without machine learning expertise to utilize it. Users can label image datasets, upload them, and obtain highly accurate trained models for various tasks, such as part recognition in factories.
Exploring Alternatives to Reinforcement Learning: Researchers are investigating evolutionary algorithms as an alternative to reinforcement learning in AutoML. This approach aims to determine the effectiveness of evolutionary algorithms in comparison to reinforcement learning.
Learning Symbolic Optimization Rules: AutoML can learn and combine different optimization update rules symbolically, resulting in new and improved optimization functions. It can also discover non-linearity functions beyond those traditionally explored in neural network optimization.
Optimizing Models for Latency and Accuracy: AutoML can be adapted to consider both latency and accuracy when searching for optimal neural architectures. This enables the creation of models that meet specific latency requirements while maintaining high accuracy.
Data Augmentation Policies: AutoML utilizes data augmentation techniques to expand the training dataset and improve model performance. It applies transformations such as resizing, flipping, and cropping images while preserving their labels.
Exploring Multiple Architectures Simultaneously: To reduce computational costs, AutoML can explore multiple neural architectures simultaneously, sharing parameters among them. This approach allows for efficient exploration of a broader range of architectures.
TPU Background: Deep learning algorithms are tolerant of reduced precision, requiring only one decimal digit of precision, unlike traditional HPC codes. Neural algorithms utilize a handful of specific kinds of matrix operations and linear algebra operations. Low-precision linear algebra operations can be used to express various machine learning algorithms.
TPU Motivation: Increased speech recognition usage could potentially double the number of computers needed for that feature alone. A hardware accelerator was required for low-latency inference needs.
TPUv1: First TPU was for inference only, not training. Designed to fit a trained model in memory and perform low-precision number multiplication. Used in production for four years, including search queries, translations, speech, and image recognition. Published in ISCA 2017.
TPUv2: Designed for both training and inference, targeting large problems. Features four chips with about 180 teraflops of compute, connected together. Matrix multiply unit provides 65,000 scalar results in 128×128 matrix per cycle. Designed to be connected into pods, such as 64 chips arranged in a 16×16 mesh.
TPUv3: Announced in May, with improvements to the core chip and water cooling for higher power. Can be connected into a larger pod configuration, about four times bigger than TPUv2. Delivers eight times the performance of the earlier pod, achieving 100 petaflops of compute. Enables large-scale machine learning problems and quick experimentation results.
Programming and Accessibility: Programmed via TensorFlow, allowing the same program to run with minor modifications on different platforms. Scales via synchronous data parallelism, making it easy for machine learning researchers to use pods without modification. DAWNBench benchmark shows that TPUs hold the top three spots for training time and training costs. 1,000 devices are made available for free to top researchers committed to open machine learning research.
00:41:14 Machine Learning Techniques for System Optimization
Introduction to Learned Index Data Structures: Traditional index data structures, such as B-trees, are used in databases and key-value stores to efficiently find data based on a key. Machine learning offers an opportunity to create learned index data structures that can replace traditional ones and provide improved performance and space efficiency.
How Learned Index Data Structures Work: Learned index data structures use a small neural network to make predictions about the location of a key in a sorted list of keys. The neural network takes in a key as input and outputs a prediction of where that key is likely to fall in the sorted list. This prediction is used to perform the same function as a B-tree, which is to identify the disk page where the key must exist if it exists at all.
Benefits of Learned Index Data Structures: Learned index data structures can be significantly faster than traditional B-trees, with one example being 60% faster. They can also be much smaller in size, with the same example being 1/20th the space of a B-tree. The accuracy and size of the learned index can be traded off to achieve the desired performance and space requirements.
Potential Applications of Learned Index Data Structures: Learned index data structures have the potential to revolutionize how data is indexed and accessed in computer systems. They could lead to smaller and faster databases, key-value stores, and other data storage systems. The ability to adapt to the actual usage patterns of the system could further improve performance and efficiency.
00:44:05 Machine Learning Fundamentals and Societal Applications
Opportunities for Learning-Based Systems: The speaker emphasizes the potential of using learning instead of heuristics in various domains, including instruction scheduling, register allocation, loop nest parallelization, networking, operating systems, ASIC design, and more. These systems could be self-adaptive and high-performance, eliminating the need for handwritten heuristics.
Challenges of AI in Society: Machine learning systems are becoming more prevalent and influential in society, raising concerns about safety, bias, interpretability, and privacy. Medical diagnostic systems require interpretable predictions for clinicians to trust and understand the reasoning behind decisions. Machine learning systems can be vulnerable to privacy issues, especially with sensitive data.
Google’s Principles for AI: Google has established principles to guide the responsible development and use of AI, addressing fairness, safety, interpretability, and privacy. These principles are not always straightforward to apply, and ongoing research is needed to develop methodologies for ensuring the safety of learned systems.
Fairness in Machine Learning: Machine learning systems trained on biased data can perpetuate and amplify existing biases. Research is being conducted to explore how to mitigate bias and promote fairness in machine learning predictions.
Conclusion: Machine learning hardware is a rapidly growing field with significant potential for practical applications. Machine learning is revolutionizing various disciplines by providing new approaches to solving complex problems. The speaker invites questions from the audience.
Abstract
Exploring the Frontiers of Machine Learning: From Deep Learning to Real-World Applications
Introduction: Deep Learning’s Ascendance in Modern Technology
Deep learning, a modern rebranding of artificial neural networks, is experiencing a surge in interest and growth. Google’s research approach includes basic research, infrastructure building, product collaboration, and emerging area exploration. This synergy between foundational research and real-world application is a recurring theme in the evolution of machine learning.
The Exponential Growth of Machine Learning
Machine learning, particularly its subset deep learning, is undergoing an explosive growth, outpacing even Moore’s Law. This is evident in the surging number of research papers and advancements. Deep learning thrives on its ability to solve complex tasks through a collection of trainable mathematical units, showcasing versatility and adaptability.
Revolutionizing Fields with Neural Networks
Neural networks have made remarkable strides in various domains such as computer vision, audio transcription, and language translation. A notable achievement is the drastic reduction of error rates in computer vision, evidenced by the ImageNet contest results, plummeting from 26% in 2011 to just 3% in 2016.
Addressing Grand Challenges with Machine Learning
Deep learning’s potential to solve the U.S. National Academy of Engineering’s grand challenges is significant. Machine learning can significantly contribute to solving global challenges, such as improving urban infrastructure, promoting healthier living, and mitigating climate change. From improving urban infrastructure to revolutionizing self-driving cars, machine learning is at the forefront. Robotics, benefitting from advancements in vision technology, is moving towards more autonomous and adaptable systems, with robotic grasping systems now achieving up to 96% success rates. Self-driving cars are being actively tested and developed, with Waymo leading the way in real-world trials. These cars use sensor data to understand their surroundings, predict object movements, and make safe driving decisions.
Predicting Patient Mortality: Deep Learning in Healthcare
Traditional methods use a limited set of hand-designed features to predict mortality risk, while deep learning approaches utilize all data in the medical record, improving prediction accuracy. This enables earlier assessment of patients’ medical risk. Deep learning can predict patient mortality more accurately using all data in the medical record, allowing for earlier assessment of patients’ medical risk.
Accelerating Quantum Chemistry Simulations
Quantum chemists use computationally intensive simulations to predict molecule properties. Deep learning offers a faster alternative. Training a neural network with simulation data can achieve indistinguishable accuracy, while being 300,000 times faster than the original simulator.
Exploiting Simulators for Scientific Discovery
Replacing simulation components with neural networks can lead to significant speedups, making it feasible to screen large datasets and accelerating scientific discovery.
Healthcare: A Prime Beneficiary of Machine Learning
In healthcare, machine learning, especially in medical imaging diagnosis, stands on the brink of setting new care standards. Models trained on specialist-labeled images are matching or even surpassing human experts, offering insights into patterns previously unnoticed by ophthalmologists. Moreover, the application extends to electronic medical records, enhancing patient outcome predictions and treatment recommendations. Machine learning has tremendous potential in healthcare for improved decision-making. In medical imaging, computer vision techniques aid in diagnostic tasks like diabetic retinopathy screening. An ML model achieved par or better performance compared to US board-certified ophthalmologists in grading retinal images. Advanced models can identify patterns in retinal images that ophthalmologists may not be aware of, providing new insights into patient health. Machine learning models can predict future outcomes for patients based on electronic medical record data, aiding in decision-making and treatment planning. Applications of these ML models include Smart Reply in Gmail, machine translation, and predicting various aspects of patient care using de-identified medical records.
Expanding Machine Learning’s Reach with TensorFlow and AutoML
TensorFlow, Google’s open-source machine learning platform, has catalyzed the democratization of machine learning research and application. Its flexibility has led to diverse applications, from agricultural innovations to healthcare diagnostics. TensorFlow is an open-source machine learning framework that enables users to express machine learning research ideas and deploy them in real-world environments. With over 16 million downloads and a vibrant community, it runs on various platforms, including mobile phones, desktops, and data centers. Complementing this is the advent of AutoML, aiming to make machine learning accessible to non-experts through automated model design and neural architecture search, generating models with unique structures and efficient computational use.
Enhancing Computational Efficiency with TPUs
Google’s development of Tensor Processing Units (TPUs) marks a significant advancement in machine learning hardware. From the inference-focused TPUv1 to the more comprehensive TPUv2 and the enhanced TPUv3 with water cooling for increased performance, these specialized accelerators have revolutionized machine learning research by enabling faster and more efficient training and experimentation.
Machine Learning for System Optimization
Machine learning is not just transforming applications but also systems themselves. Learned indices, utilizing neural networks for key position prediction, are proving more efficient than traditional B-trees. This concept extends to instruction scheduling, register allocation, and various system optimization domains, where learning-based approaches can significantly outperform traditional heuristics.
The Ethical Dimension of AI in Society
As AI permeates various societal aspects, ensuring its interpretability, safety, and fairness becomes paramount. AI systems, especially in critical domains like medical diagnostics, must be transparent and explainable. Moreover, addressing data bias to prevent unfairness in machine learning predictions is an ongoing research focus.
The Future of Machine Learning
In conclusion, machine learning, particularly through deep learning, is making significant strides across various disciplines. The development of specialized hardware like TPUs is enhancing computational efficiency, while applications in fields from healthcare to system optimization showcase the technology’s breadth. However, the responsible and ethical use of AI is crucial to address challenges around interpretability, safety, bias, and privacy, ensuring AI’s beneficial integration into society.
Supplemental Updates from Recent Research and Developments:
TPU Development Journey:
Google’s TPU development has been instrumental in advancing machine learning hardware. TPUv1, designed for inference, enabled low-latency needs. TPUv2, capable of both training and inference, offered increased compute power and scalability. The latest TPUv3, with water cooling and enhanced performance, enables large-scale machine learning problems and quick experimentation results.
Machine Learning for Systems: Learned Index Data Structures:
Learned index data structures, utilizing neural networks for key position prediction, offer significant improvements over traditional B-trees. They can be faster, smaller, and adapt to the system’s usage patterns, leading to enhanced performance and efficiency.
Reflections on the Wide-Ranging Applications and Societal Implications of Machine Learning:
Machine learning’s potential for learning-based systems in various domains, from instruction scheduling to ASIC design, holds promise for high-performance and self-adaptive systems. However, concerns about AI’s safety, bias, interpretability, and privacy require ongoing research and the development of methodologies to ensure the responsible use of learned systems.
Google's search systems and infrastructure have evolved significantly, driven by hardware improvements, distributed architectures, and innovative techniques like MapReduce and Spanner. The focus on scalability, availability, and performance optimization has set benchmarks in web search and data processing, inspiring future innovations in large-scale data handling....
Deep learning has evolved from theoretical insights to practical applications, and its future holds promise for further breakthroughs with increased compute power and large-scale efforts. The intersection of image and language understanding suggests a potential convergence towards a unified architectural approach in the future....
Machine learning and artificial intelligence advancements are revolutionizing various fields, including healthcare, automotive, and scientific research. Google's TensorFlow, computational techniques, and neural architecture search are key players in this transformation....
Geoffrey Hinton's groundbreaking work in neural networks revolutionized AI by mimicking the brain's learning process and achieving state-of-the-art results in tasks like speech recognition and image processing. His approach, inspired by the brain, laid the foundation for modern AI and raised questions about the potential and limitations of neural networks....
Neural networks have revolutionized various fields, from language translation and speech recognition to healthcare and finance, by outperforming logic-based AI systems in learning and adapting from vast data sets. They face challenges such as adversarial attacks, explainability, and regulatory compliance, but hold great promise for the future, including self-driving vehicles,...
Geoffrey Hinton, a pioneer in deep learning, has made significant contributions to AI and neuroscience, leading to a convergence between the two fields. His work on neural networks, backpropagation, and dropout regularization has not only enhanced AI but also provided insights into understanding the human brain....
Neural networks, empowered by backpropagation, have revolutionized computing, enabling machines to learn from data and adapt to various applications, influencing fields like image recognition, natural language processing, and healthcare. These networks excel in tasks that involve complex data patterns and have exceeded human performance in certain domains....