This article aims to contextualize machine learning (ML) for hardware and embedded engineers, what it is, how it works, why it matters, and how TinyML fits in.
June 05, 2022 by Brandon Satrom
This article gives you essential information on a mathematical technique that plays an absolutely fundamental role in system design and signal processing.
July 30, 2020 by Robert Keim
Learn the key parts of an autoencoder, how a variational autoencoder improves on it, and how to build and train a variational autoencoder using TensorFlow.
April 06, 2020 by Henry Ansah Fordjour
This article discusses a complication that can prevent your Perceptron from achieving adequate classification accuracy.
February 06, 2020 by Robert Keim
This article shows you how to add bias values to a multilayer Perceptron implemented in a high-level programming language such as Python.
February 05, 2020 by Robert Keim
In this article, we’ll perform some classification experiments and gather data on the relationship between hidden-layer dimensionality and network performance.
February 04, 2020 by Robert Keim
This article provides guidelines for configuring the hidden portion of a multilayer Perceptron.
January 31, 2020 by Robert Keim
In this article, we’ll use Excel-generated samples to train a multilayer Perceptron, and then we’ll see how the network performs with validation samples.
January 30, 2020 by Robert Keim
This article explains why validation is particularly important when we’re processing data using a neural network.
January 28, 2020 by Robert Keim
This article takes you step by step through a Python program that will allow us to train a neural network and perform advanced classification.
January 19, 2020 by Robert Keim
This article discusses the Perceptron configuration that we will use for our experiments with neural-network training and classification, and we’ll also look at the related topic of bias nodes.
January 09, 2020 by Robert Keim
This article presents the equations that we use when performing weight-update computations, and we’ll also discuss the concept of backpropagation.
December 27, 2019 by Robert Keim
We can greatly enhance the performance of a Perceptron by adding a layer of hidden nodes, but those hidden nodes also make training a bit more complicated.
December 26, 2019 by Robert Keim
In this article, we’ll see why we need a new activation function for a neural network that is trained via gradient descent.
December 25, 2019 by Robert Keim
This article explains why high-performance neural networks need an extra “hidden” layer of computational nodes.
December 24, 2019 by Robert Keim
This article discusses learning rate, which plays an important role in neural-network training.
December 19, 2019 by Robert Keim
This article presents an algorithm intended for humans to calibrate the internal oscillator of an MCU, with the help of an oscilloscope and a spreadsheet. An example experiment with numbers is also shown.
December 13, 2019 by Eduardo Corpeño
In this article, we’ll explore Perceptron training from a more theoretical perspective, focusing on the “error bowl.”
December 05, 2019 by Robert Keim
In this article, we’ll review some important aspects of neural-net training, and then we’ll discuss the concept of overtraining.
November 26, 2019 by Robert Keim
This article presents Python code that allows you to automatically generate weights for a simple neural network.
November 24, 2019 by Robert Keim
Don't have an AAC account? Create one now.
Forgot your password? Click here.