Category Archives: Deep Learning

LM101-085: Ch7: How to Guarantee your Batch Learning Algorithm Converges

This 85th episode of Learning Machines 101 discusses formal convergence guarantees for a broad class of machine learning algorithms designed to minimize smooth non-convex objective functions using batch learning methods. In particular, a broad class of unsupervised, supervised, and reinforcement machine learning algorithms which iteratively update their parameter vector by adding a perturbation based upon all of the training data. This process is repeated, making a perturbation of the parameter vector based upon all of the training data until a parameter vector is generated which exhibits improved predictive performance. The magnitude of the perturbation at each learning iteration is called the “stepsize” or “learning rate” and the identity of the perturbation vector is called the “search direction”. Simple mathematical formulas are presented based upon research from the late 1960s by Philip Wolfe and G. Zoutendijk that ensure convergence of the generated sequence of parameter vectors. These formulas may be used as the basis for the design of artificially intelligent smart automatic learning rate selection algorithms. The material in this podcast is designed to provide an overview of Chapter 7 of my new book “Statistical Machine Learning” and is based upon material originally presented in Episode 68 of Learning Machines 101!

LM101-083: Ch5: How to Use Calculus to Design Learning Machines

Episode Summary: This particular podcast covers the material from Chapter 5 of my new book “Statistical Machine Learning: A unified framework” which is now available! The book chapter shows how matrix calculus is very useful for the analysis and design of both linear and nonlinear learning machines with lots of examples. Show Notes: Hello everyone! Welcome to the… Read More »

LM101-069: What Happened at the 2017 Neural Information Processing Systems Conference?

LM101-069: What Happened at the 2017 Neural Information Processing Systems Conference?   Episode Summary: This 69th episode of Learning Machines 101 provides a short overview of the 2017 Neural Information Processing Systems conference with a focus on the development of methods for teaching learning machines rather than simply training them on examples. In addition, a book review of… Read More »

LM101-068: How to Design Automatic Learning Rate Selection for Gradient Descent Type Machine Learning Algorithms

LM101-068: How to Design Automatic Learning Rate Selection for Gradient Descent Type Machine Learning Algorithms Episode Summary: This 68th episode of Learning Machines 101 discusses a broad class of unsupervised, supervised, and reinforcement machine learning algorithms which iteratively update their parameter vector by adding a perturbation based upon all of the training data. This process is repeated, making… Read More »

LM101-065: How to Design Gradient Descent Learning Machines (Rerun)

LM101-065: How to Design Gradient Descent Learning Machines (Rerun) Episode Summary: In this episode we introduce the concept of gradient descent which is the fundamental principle underlying learning in the majority of deep learning and neural network learning algorithms. Show Notes: Hello everyone! Welcome to the sixteenth podcast in the podcast series Learning Machines 101. In this series… Read More »

LM101-063: How to Transform a Supervised Learning Machine into a Policy Gradient Reinforcement Learning Machine

LM101-063: How to Transform a Supervised Learning Machine into a Policy Gradient Reinforcement Learning Machine Episode Summary: This 63rd episode of Learning Machines 101 discusses how to build reinforcement learning machines which become smarter with experience but do not use this acquired knowledge to modify their actions and behaviors. This episode explains how to build reinforcement learning machines… Read More »

LM101-062: How to Transform a Supervised Learning Machine into a Value Function Reinforcement Learning Machine

LM101-062: How to Transform a Supervised Learning Machine into a Value Function Reinforcement Learning Machine Episode Summary: This 62nd episode of Learning Machines 101 discusses how to design reinforcement learning machines using your knowledge of how to build supervised learning machines! Specifically, we focus on Value Function Reinforcement Learning Machines which estimate the unobservable total penalty associated with… Read More »

LM101-061: What happened at the Reinforcement Learning Tutorial? (RERUN)

LM101-061: What happened at the Reinforcement Learning Tutorial? (RERUN) Episode Summary: This is the third of a short subsequence of podcasts providing a summary of events associated with Dr. Golden’s recent visit to the 2015 Neural Information Processing Systems Conference. This is one of the top conferences in the field of Machine Learning. This episode reviews and discusses… Read More »

LM101-059: How to Properly Introduce a Neural Network

LM101-059: How to Properly Introduce a Neural Network Episode Summary: I discuss the concept of a “neural network” by providing some examples of recent successes in neural network machine learning algorithms and providing a historical perspective on the evolution of the neural network concept from its biological origins. Show Notes: Hello everyone! Welcome to the fifty-ninth podcast in… Read More »

LM101-052: How to Use the Kernel Trick to Make Hidden Units Disappear

LM101-052: How to Use the Kernel Trick to Make Hidden Units Disappear Episode Summary: Today, we discuss a simple yet powerful idea which began popular in the machine learning literature in the 1990s which is called “The Kernel Trick”. The basic idea behind “The Kernel Trick” is that an impossible machine learning problem can be transformed into an… Read More »

LM101-051: How to Use Radial Basis Function Perceptron Software for Supervised Learning [Rerun]

LM101-051: How to Use Radial Basis Function Perceptron Software for Supervised Learning [Rerun] Episode Summary: In this episode we describe how to download and use free nonlinear machine learning software for implementing a Perceptron learning machine with a single layer of Radial Basis Function hidden units for the purposes of supervised learning. Show Notes: Welcome to the 51st podcast… Read More »

LM101-046: How to Optimize Student Learning using Recurrent Neural Networks (Educational Technology)

LM101-046: How to Optimize Student Learning using Recurrent Neural Networks (Educational Technology) Episode Summary: In this episode, we briefly review Item Response Theory and Bayesian Network Theory methods for the assessment and optimization of student learning and then describe a poster presented on the first day of the Neural Information Processing Systems conference in December 2015 in Montreal… Read More »

LM101-045: How to Build a Deep Learning Machine for Answering Questions about Images

LM101-045: How to Build a Deep Learning Machine for Answering Questions about Images Episode Summary: This is the fourth of a short subsequence of podcasts which provides a summary of events associated with Dr. Golden’s recent visit to the 2015 Neural Information Processing Systems Conference. This is one of the top conferences in the field of Machine Learning. This… Read More »

LM101-044: What happened at the Deep Reinforcement Learning Tutorial at the 2015 Neural Information Processing Systems Conference?

LM101-044: What happened at the Deep Reinforcement Learning Tutorial at the 2015 Neural Information Processing Systems Conference? Episode Summary: This is the third of a short subsequence of podcasts providing a summary of events associated with Dr. Golden’s recent visit to the 2015 Neural Information Processing Systems Conference. This is one of the top conferences in the field… Read More »

LM101-041: What happened at the 2015 Neural Information Processing Systems Deep Learning Tutorial?

LM101-041: What happened at the 2015 Neural Information Processing Systems Deep Learning Tutorial? Episode Summary: This is the first of a short subsequence of podcasts which provides a summary of events at the recent 2015 Neural Information Processing Systems Conference. This is one of the top conferences in the field of Machine Learning. This episode introduces the Neural… Read More »

LM101-036: How to Predict the Future from the Distant Past using Recurrent Neural Networks

LM101-036: How to Predict the Future from the Distant Past using Recurrent Neural Networks Episode Summary: In this episode, we discuss the problem of predicting the future from not only recent events but also from the distant past using Recurrent Neural Networks (RNNs). A example RNN is described which learns to label images with simple sentences. A learning… Read More »

LM101-034: How to Use Nonlinear Machine Learning Software to Make Predictions (Feedforward Perceptrons with Radial Basis Functions)[Rerun]

LM101-034: How to Use Nonlinear Machine Learning Software to Make Predictions (Feedforward Perceptrons with Radial Basis Functions)[Rerun] Episode Summary: In this episode we describe how to download and use free nonlinear machine learning software which is more advanced than the linear machine software introduced in Episode 13. Show Notes: Welcome to the 34th podcast in the podcast series… Read More »

LM101-030: How to Improve Deep Learning Performance with Artificial Brain Damage (Dropout and Model Averaging)

LM101-030: How to Improve Deep Learning Performance with Artificial Brain Damage (Dropout and Model Averaging) Episode Summary: Deep learning machine technology has rapidly developed over the past five years due in part to a variety of factors such as: better technology, convolutional net algorithms, rectified linear units, and a relatively new learning strategy called “dropout” in which hidden… Read More »

LM101-029: How to Modernize Deep Learning with Rectilinear units, Convolutional Nets, and Max-Pooling

LM101-029: How to Modernize Deep Learning  with Rectilinear units,  Convolutional Nets, and Max-Pooling Episode Summary This podcast discusses the topics of rectilinear units, convolutional nets, and max-pooling relevant to deep learning which were inspired by my recent visit to the 3rd International Conference on Learning Representations (May 7-9, 2015) in San Diego. Specifically, commonly used techniques shared by… Read More »

LM101-023: How to Build a Deep Learning Machine (Function Approximation)

Episode Summary: In this episode we discuss how to design and build “Deep Learning Machines” which can autonomously discover useful ways to represent knowledge of the world. Show Notes: Hello everyone! Welcome to the twenty-third podcast in the podcast series Learning Machines 101. In this series of podcasts my goal is to discuss important concepts of artificial intelligence… Read More »