Book description
With the reinvigoration of neural networks in the 2000s, deep learning has become an extremely active area of research, one that’s paving the way for modern machine learning. In this practical book, author Nikhil Buduma provides examples and clear explanations to guide you through major concepts of this complicated field.
Companies such as Google, Microsoft, and Facebook are actively growing in-house deep-learning teams. For the rest of us, however, deep learning is still a pretty complex and difficult subject to grasp. If you’re familiar with Python, and have a background in calculus, along with a basic understanding of machine learning, this book will get you started.
- Examine the foundations of machine learning and neural networks
- Learn how to train feed-forward neural networks
- Use TensorFlow to implement your first neural network
- Manage problems that arise as you begin to make networks deeper
- Build neural networks that analyze complex images
- Perform effective dimensionality reduction using autoencoders
- Dive deep into sequence analysis to examine language
- Understand the fundamentals of reinforcement learning
Publisher resources
Table of contents
- Preface
- 1. The Neural Network
- 2. Training Feed-Forward Neural Networks
-
3. Implementing Neural Networks in TensorFlow
- What Is TensorFlow?
- How Does TensorFlow Compare to Alternatives?
- Installing TensorFlow
- Creating and Manipulating TensorFlow Variables
- TensorFlow Operations
- Placeholder Tensors
- Sessions in TensorFlow
- Navigating Variable Scopes and Sharing Variables
- Managing Models over the CPU and GPU
- Specifying the Logistic Regression Model in TensorFlow
- Logging and Training the Logistic Regression Model
- Leveraging TensorBoard to Visualize Computation Graphs and Learning
- Building a Multilayer Model for MNIST in TensorFlow
- Summary
-
4. Beyond Gradient Descent
- The Challenges with Gradient Descent
- Local Minima in the Error Surfaces of Deep Networks
- Model Identifiability
- How Pesky Are Spurious Local Minima in Deep Networks?
- Flat Regions in the Error Surface
- When the Gradient Points in the Wrong Direction
- Momentum-Based Optimization
- A Brief View of Second-Order Methods
- Learning Rate Adaptation
- The Philosophy Behind Optimizer Selection
- Summary
-
5. Convolutional Neural Networks
- Neurons in Human Vision
- The Shortcomings of Feature Selection
- Vanilla Deep Neural Networks Don’t Scale
- Filters and Feature Maps
- Full Description of the Convolutional Layer
- Max Pooling
- Full Architectural Description of Convolution Networks
- Closing the Loop on MNIST with Convolutional Networks
- Image Preprocessing Pipelines Enable More Robust Models
- Accelerating Training with Batch Normalization
- Building a Convolutional Network for CIFAR-10
- Visualizing Learning in Convolutional Networks
- Leveraging Convolutional Filters to Replicate Artistic Styles
- Learning Convolutional Filters for Other Problem Domains
- Summary
-
6. Embedding and Representation Learning
- Learning Lower-Dimensional Representations
- Principal Component Analysis
- Motivating the Autoencoder Architecture
- Implementing an Autoencoder in TensorFlow
- Denoising to Force Robust Representations
- Sparsity in Autoencoders
- When Context Is More Informative than the Input Vector
- The Word2Vec Framework
- Implementing the Skip-Gram Architecture
- Summary
-
7. Models for Sequence Analysis
- Analyzing Variable-Length Inputs
- Tackling seq2seq with Neural N-Grams
- Implementing a Part-of-Speech Tagger
- Dependency Parsing and SyntaxNet
- Beam Search and Global Normalization
- A Case for Stateful Deep Learning Models
- Recurrent Neural Networks
- The Challenges with Vanishing Gradients
- Long Short-Term Memory (LSTM) Units
- TensorFlow Primitives for RNN Models
- Implementing a Sentiment Analysis Model
- Solving seq2seq Tasks with Recurrent Neural Networks
- Augmenting Recurrent Networks with Attention
- Dissecting a Neural Translation Network
- Summary
-
8. Memory Augmented Neural Networks
- Neural Turing Machines
- Attention-Based Memory Access
- NTM Memory Addressing Mechanisms
- Differentiable Neural Computers
- Interference-Free Writing in DNCs
- DNC Memory Reuse
- Temporal Linking of DNC Writes
- Understanding the DNC Read Head
- The DNC Controller Network
- Visualizing the DNC in Action
- Implementing the DNC in TensorFlow
- Teaching a DNC to Read and Comprehend
- Summary
-
9. Deep Reinforcement Learning
- Deep Reinforcement Learning Masters Atari Games
- What Is Reinforcement Learning?
- Markov Decision Processes (MDP)
- Explore Versus Exploit
- Policy Versus Value Learning
- Pole-Cart with Policy Gradients
-
Q-Learning and Deep Q-Networks
- The Bellman Equation
- Issues with Value Iteration
- Approximating the Q-Function
- Deep Q-Network (DQN)
- Training DQN
- Learning Stability
- Target Q-Network
- Experience Replay
- From Q-Function to Policy
- DQN and the Markov Assumption
- DQN’s Solution to the Markov Assumption
- Playing Breakout wth DQN
- Building Our Architecture
- Stacking Frames
- Setting Up Training Operations
- Updating Our Target Q-Network
- Implementing Experience Replay
- DQN Main Loop
- DQNAgent Results on Breakout
- Improving and Moving Beyond DQN
- Summary
- Index
Product information
- Title: Fundamentals of Deep Learning
- Author(s):
- Release date: June 2017
- Publisher(s): O'Reilly Media, Inc.
- ISBN: 9781491925614
You might also like
book
Deep Learning Essentials
Get to grips with the essentials of deep learning by leveraging the power of Python About …
book
Advanced Deep Learning with Python
Gain expertise in advanced deep learning domains such as neural networks, meta-learning, graph neural networks, and …
book
Hands-On Deep Learning Algorithms with Python
Understand basic-to-advanced deep learning algorithms, the mathematical principles behind them, and their practical applications Key Features …
book
Interpretable Machine Learning with Python
A deep and detailed dive into the key aspects and challenges of machine learning interpretability, complete …