Artificial Intelligence Engines
A tutorial introduction to the mathematics of deep learning
About this book
The brain has always had a fundamental advantage over conventional computers: it can learn. However, a new generation of artificial intelligence algorithms, in the form of deep neural networks, is rapidly eliminating that advantage.
Deep neural networks rely on adaptive algorithms to master a wide variety of tasks, including cancer diagnosis, object recognition, speech recognition, robotic control, chess, poker, backgammon and Go, at super-human levels of performance.
In this richly illustrated book, key neural network learning algorithms are explained informally first, followed by detailed mathematical analyses. Topics include both historically important neural networks (perceptrons, Hopfield nets, Boltzmann machines and backpropagation networks), and modern deep neural networks (variational autoencoders, convolutional networks, generative adversarial networks, and reinforcement learning using SARSA and Q-learning).
Online computer programs, collated from open source repositories, give hands-on experience of neural networks, and PowerPoint slides provide support for teaching. Written in an informal style, with a comprehensive glossary, tutorial appendices (eg Bayes' theorem, maximum likelihood estimation), and a list of further readings, this is an ideal introduction to the algorithmic engines of modern artificial intelligence.
Published 1 April 2019
Paperback ISBN: 9780956372819
Hardback ISBN: 9780956372826
Download Chapter 1 (PDF, 9.7MB)
List of pseudocode examples
Online code examples
1. Artificial neural networks
1.2 What is an artificial neural network?
1.3 The origins of neural networks
1.4 From backprop to deep learning
1.5 An overview of Chapters
2. Linear associative networks
2.2 Setting one connection weight
2.3 Learning one association
2.4 Gradient descent
2.5 Learning two associations
2.6 Learning many associations
2.7 Learning photographs
3.2 The perceptron learning algorithm
3.3 The exclusive OR problem
3.4 Why exclusive OR matters
4. The backpropagation algorithm
4.2 The backpropagation algorithm
4.3 Why use sigmoidal hidden units?
4.4 Generalisation and overfitting
4.5 Vanishing gradients
4.6 Speeding up backprop
4.7 Local and global mimima
4.8 Temporal backprop
4.9 Early backprop achievements
5. Hopfield nets
5.2 The Hopfield net
5.3 Learning one network state
5.4 Content addressable memory
5.5 Tolerance to damage
5.6 The energy function
6. Boltzmann machines
6.2 Learning in generative models
6.3 The Boltzmann machine energy function
6.4 Simulated annealing
6.5 Learning by sculpting distributions
6.6 Learning in Boltzmann machines
6.7 Learning by maximising likelihood
6.8 Autoencoder networks
7. Deep RBMs
7.2 Restricted Boltzmann machines
7.3 Training restricted Boltzmann machines
7.4 Deep autoencoder networks
8. Variational autoencoders
8.2 Why favour independent features?
8.3 Overview of variational autoencoders
8.4 Latent variables and manifolds
8.5 Key quantities
8.6 How variational autoencoders work
8.7 The evidence lower bound
8.8 An alternative derivation
8.9 Maximising the lower bound
8.10 Conditional variational autoencoders
9. Deep backprop networks
9.2 Convolutional neural networks
9.8 Ladder autoencoder networks
9.9 Denoising autoencoder networks
9.10 Fooling neural networks
9.11 Generative adversarial networks
9.12 Temporal deep neural networks
9.13 Capsule neural networks
10. Reinforcement learning
10.2 What's the problem?
10.3 Key quantities
10.4 Markov decision processes
10.5 Formalising the problem
10.6 The Bellman equation
10.7 Learning state-value functions
10.8 Eligibility traces
10.9 Learning action-value functions
10.10 Balancing a pole
11. The emperor’s new AI?
11.1 Artificial intelligence
11.2 Yet another revolution?
B. Mathematical symbols
C. A vector matrix tutorial
D. Maximum likelihood estimation
E. Bayes' theorem
"Authoritative, funny, and concise."
Steven Strogatz, Professor of Applied Mathematics, Cornell University
"Artificial Intelligence Engines will introduce you to the rapidly growing field of deep learning networks: how to build them, how to use them and how to think about them. James Stone will guide you from the basics to the outer reaches of a technology that is changing the world."
Professor Terrence Sejnowski, Director of the Computational Neurobiology Laboratory, Salk Institute, USA and author of The Deep Learning Revolution, MIT Press, 2018
"This book manages the impossible: it is a fun read, intuitive and engaging, light-hearted and delightful, and cuts right through the hype and turgid terminology. Unlike many texts, this is not a shallow cookbook for some particular deep learning program-du-jure. Instead, it crisply and painlessly imparts the principles, intuitions and background needed to understand existing machine-learning systems, learn new tools, and invent novel architectures, with ease."
Professor Barak Pearlmutter, Brain and Computation Laboratory, National University of Ireland Maynooth, Ireland
"This text provides an engaging introduction to the mathematics underlying neural networks. It is meant to be read from start to finish, as it carefully builds up, chapter by chapter, the essentials of neural network theory. After first describing classic linear networks and nonlinear multilayer perceptrons, Stone gradually introduces a comprehensive range of cutting edge technologies in use today. Written in an accessible and insightful manner, this book is a pleasure to read, and I will certainly be recommending it to my students."
Dr Stephen Eglen, Department of Applied Mathematics and Theoretical Physics (DAMTP), Cambridge Computational Biology Institute (CCBI), Cambridge University, UK