Category Archives: Gradient Descent Learning

LM101-085: Ch7: How to Guarantee your Batch Learning Algorithm Converges

This 85th episode of Learning Machines 101 discusses formal convergence guarantees for a broad class of machine learning algorithms designed to minimize smooth non-convex objective functions using batch learning methods. In particular, a broad class of unsupervised, supervised, and reinforcement machine learning algorithms which iteratively update their parameter vector by adding a perturbation based upon all of the training data. This process is repeated, making a perturbation of the parameter vector based upon all of the training data until a parameter vector is generated which exhibits improved predictive performance. The magnitude of the perturbation at each learning iteration is called the “stepsize” or “learning rate” and the identity of the perturbation vector is called the “search direction”. Simple mathematical formulas are presented based upon research from the late 1960s by Philip Wolfe and G. Zoutendijk that ensure convergence of the generated sequence of parameter vectors. These formulas may be used as the basis for the design of artificially intelligent smart automatic learning rate selection algorithms. The material in this podcast is designed to provide an overview of Chapter 7 of my new book “Statistical Machine Learning” and is based upon material originally presented in Episode 68 of Learning Machines 101!

LM101-083: Ch5: How to Use Calculus to Design Learning Machines

Episode Summary: This particular podcast covers the material from Chapter 5 of my new book “Statistical Machine Learning: A unified framework” which is now available! The book chapter shows how matrix calculus is very useful for the analysis and design of both linear and nonlinear learning machines with lots of examples. Show Notes: Hello everyone! Welcome to the… Read More »

LM101-069: What Happened at the 2017 Neural Information Processing Systems Conference?

LM101-069: What Happened at the 2017 Neural Information Processing Systems Conference?   Episode Summary: This 69th episode of Learning Machines 101 provides a short overview of the 2017 Neural Information Processing Systems conference with a focus on the development of methods for teaching learning machines rather than simply training them on examples. In addition, a book review of… Read More »

LM101-068: How to Design Automatic Learning Rate Selection for Gradient Descent Type Machine Learning Algorithms

LM101-068: How to Design Automatic Learning Rate Selection for Gradient Descent Type Machine Learning Algorithms Episode Summary: This 68th episode of Learning Machines 101 discusses a broad class of unsupervised, supervised, and reinforcement machine learning algorithms which iteratively update their parameter vector by adding a perturbation based upon all of the training data. This process is repeated, making… Read More »

LM101-065: How to Design Gradient Descent Learning Machines (Rerun)

LM101-065: How to Design Gradient Descent Learning Machines (Rerun) Episode Summary: In this episode we introduce the concept of gradient descent which is the fundamental principle underlying learning in the majority of deep learning and neural network learning algorithms. Show Notes: Hello everyone! Welcome to the sixteenth podcast in the podcast series Learning Machines 101. In this series… Read More »

LM101-049: How to Experiment with Lunar Lander Software

LM101-049: How to Experiment with Lunar Lander Software Episode Summary: In this episode we continue the discussion of learning when the actions of the learning machine can alter the characteristics of the learning machine’s statistical environment. We describe how to download free lunar lander software so you can experiment with an autopilot for a lunar lander module that… Read More »

LM101-041: What happened at the 2015 Neural Information Processing Systems Deep Learning Tutorial?

LM101-041: What happened at the 2015 Neural Information Processing Systems Deep Learning Tutorial? Episode Summary: This is the first of a short subsequence of podcasts which provides a summary of events at the recent 2015 Neural Information Processing Systems Conference. This is one of the top conferences in the field of Machine Learning. This episode introduces the Neural… Read More »

LM101-036: How to Predict the Future from the Distant Past using Recurrent Neural Networks

LM101-036: How to Predict the Future from the Distant Past using Recurrent Neural Networks Episode Summary: In this episode, we discuss the problem of predicting the future from not only recent events but also from the distant past using Recurrent Neural Networks (RNNs). A example RNN is described which learns to label images with simple sentences. A learning… Read More »

LM101-034: How to Use Nonlinear Machine Learning Software to Make Predictions (Feedforward Perceptrons with Radial Basis Functions)[Rerun]

LM101-034: How to Use Nonlinear Machine Learning Software to Make Predictions (Feedforward Perceptrons with Radial Basis Functions)[Rerun] Episode Summary: In this episode we describe how to download and use free nonlinear machine learning software which is more advanced than the linear machine software introduced in Episode 13. Show Notes: Welcome to the 34th podcast in the podcast series… Read More »

LM101-031: How to Analyze and Design Learning Rules using Gradient Descent Methods (RERUN)

LM101-031: How to Analyze and Design Learning Rules using Gradient Descent Methods (RERUN) Episode Summary: In this episode we introduce the concept of gradient descent which is the fundamental principle underlying learning in the majority of machine learning algorithms. Show Notes: Hello everyone! Welcome to the sixteenth podcast in the podcast series Learning Machines 101. In this series… Read More »

LM101-025: How to Build a Lunar Lander Autopilot Learning Machine (adaptive control)

LM101-025: How to Build a Lunar Lander Autopilot Learning Machine (adaptive control) Episode Summary: In this episode we consider the problem of learning when the actions of the learning machine can alter the characteristics of the learning machine’s statistical environment. We illustrate the solution to this problem by designing an autopilot for a lunar lander module that learns… Read More »

LM101-020: How to Use Nonlinear Machine Learning Software to Make Predictions

Episode Summary: In this episode we describe how to download and use free nonlinear machine learning software which is more advanced than the linear machine software introduced in Episode 13. Show Notes: Hello everyone! Welcome to the twentieth podcast in the podcast series Learning Machines 101. In this series of podcasts my goal is to discuss important concepts… Read More »

LM101-016: How to Analyze and Design Learning Rules using Gradient Descent Methods

Episode Summary: In this episode we introduce the concept of gradient descent which is the fundamental principle underlying learning in the majority of machine learning algorithms. Show Notes: Hello everyone! Welcome to the sixteenth podcast in the podcast series Learning Machines 101. In this series of podcasts my goal is to discuss important concepts of artificial intelligence and… Read More »