Home > On-Demand Archives > Talks >

Introduction to Machine Learning and Deep Learning

Peter McLaughlin - Watch Now - DSP Online Conference 2021 - Duration: 36:33

In 2016 the Google supercomputer AlphaGo beat the world champion of the board game Go, a highly complex mathematical game. This milestone demonstrated the possibilities of Artificial Intelligence and set the scene for new technologies which are now transforming our lives, from the way we drive to the way we buy clothes. Thanks to recent advances in graphics acceleration hardware and neural network development tools, the benefits of Artificial Intelligence are within reach for any business. This presentation introduces the underlying theory of Machine Learning and Deep Learning and explains how to practically apply it. Topics covered include the training process, model types, development tools, common pitfalls and real-life examples. Attendees will walk away with a kick start to help them apply Machine Learning and Deep Learning in their projects.

This guide was created with the help of AI, based on the presentation's transcript. Its goal is to give you useful context and background so you can get the most out of the session.

What this presentation is about and why it matters

Peter McLaughlin’s talk, "Introduction to Machine Learning and Deep Learning," explains the practical ideas behind teaching computers to learn instead of hand-coding every rule. He contrasts classic rules-based systems (Deep Blue) with modern data-driven approaches (AlphaGo) to show why, for many signal-processing problems, building a model from data is more scalable and flexible than writing rules. For engineers working with signals (images, audio, sensor streams), this matters because the industry is shifting from hand-crafted feature pipelines and filters toward learned representations that can absorb far more raw information and adapt to changing conditions. The talk covers the training loop, model types (supervised vs unsupervised; classification vs regression), deep learning’s ability to work from raw pixels/samples, and common deployment pitfalls such as overfitting, bias, and hidden business costs.

Who will benefit the most from this presentation

  • Signal-processing engineers who want to apply ML/DL to image, audio, or sensor tasks (e.g., detection, classification, regression).
  • Early-career engineers or students with basic DSP/math background who need a practical introduction to model workflows and trade-offs.
  • Project leads evaluating whether to use rules-based algorithms or machine learning for industrial automation.
  • Engineers transitioning from feature-based pipelines (hand-crafted descriptors, filters) to neural-network-based solutions.

What you need to know

The talk is introductory but technical. To get the most from it, you should be comfortable with:

  • Basic linear algebra and vectors (signals represented as vectors or matrices).
  • Elementary probability and statistics (distributions, mean, variance, conditional probability).
  • Core DSP concepts: sampling, filtering, feature extraction (edges, texture descriptors, spectral features).
  • Basic programming experience (Python or similar) and familiarity with the idea of a dataset (training/validation/test splits).

Two short equations used in the talk are helpful to recall:

  • Linear model (simple regression / perceptron view): $y = \mathbf{w}^\top \mathbf{x} + b$, where $\mathbf{x}$ is the input feature vector and $\mathbf{w}$ and $b$ are learned parameters.
  • Mean squared error (a common loss): $L = \frac{1}{N}\sum_{i=1}^N (y_i - \hat{y}_i)^2$, which the optimizer minimizes by updating weights.

Also be aware of these workflow ideas: extract or ingest features (machine learning often uses hand-crafted descriptors; deep learning often ingests raw signals), choose model class, train using a loss and an optimizer, validate on held-out data, and evaluate real-world cost of errors before deployment.

Glossary

  • Model — a mathematical simplification of data that maps inputs (signals/features) to outputs (labels or values).
  • Feature — a number or set of numbers extracted from a signal (e.g., spectral energy, texture descriptor, pixel values) used as model input.
  • Supervised learning — training where each input has a labeled target (e.g., defective/not defective); common for detection and regression tasks.
  • Unsupervised learning — discovering structure (clusters, latent factors) in unlabeled signals; useful for anomaly detection and dimensionality reduction.
  • Classification — predicting discrete classes from signal inputs (e.g., material A vs B).
  • Regression — predicting continuous values from signals (e.g., liquid level, temperature).
  • Loss function — a scalar that measures prediction error during training (e.g., MSE, cross-entropy) and guides optimization.
  • Overfitting — when a model learns training-specific noise and fails to generalize to new signals; a central deployment risk.
  • Optimizer — algorithm (SGD, Adam, etc.) that updates model parameters to reduce loss during training.
  • Inference — running a trained model on new signals to produce predictions; practical constraints include latency and compute cost.

Final thoughts

Peter’s talk strikes a useful balance between conceptual framing and practical advice. He uses clear real-world examples (chess vs Go, industrial vision tasks) to explain why data-driven models can outperform hand-coded rules for complex or changing signal problems, and he emphasizes the engineering realities: data quality, compute trade-offs, validation practices, and business costs. If you work on signal-processing systems and are wondering whether to hand-tune filters and descriptors or start building learned pipelines, this presentation is a concise, pragmatic primer that will help you decide where to invest effort and how to avoid common traps. It’s especially helpful for engineers who want to move quickly from idea to a monitored prototype: narrow the problem, do a quick sniff test, and iterate.

M↓ MARKDOWN HELP
italicssurround text with
*asterisks*
boldsurround text with
**two asterisks**
hyperlink
[hyperlink](https://example.com)
or just a bare URL
code
surround text with
`backticks`
strikethroughsurround text with
~~two tilde characters~~
quote
prefix with
>

SlightlyChaotic
Score: 0 | 4 years ago | 1 reply

Thanks, excellent talk, and very well suited to my introductory level. First question:
Since we have much more computing power and familiarity with machine learning today, if we had the same situation as in 1996, where people were trying to write software to beat Garry Kasparov, would they decide that rules-based programming is too much work, and opt for machine learning instead?

Peter_McLaughlinSpeaker
Score: 0 | 4 years ago | no reply

Hello, Chess being relatively simple compared to Go, it can be solved with a rules-based approach. A Deep Learning based system would most likely out perform a rules-based system but it's much more work to build the dataset. In our computer vision work, we generally stick to rules based algorithms if we can for simplicity.

SlightlyChaotic
Score: 0 | 4 years ago | 1 reply

Second /comment/question: before machine learning became popular, I remember the term "fuzzy logic" was popular. Are you familiar with that, and can you tell me what that was about? I remember thinking, at the end of the day, it was just a more complicated decision tree in your program than simply "is it equal to 1"? At that point, it did not seem like anything noteworthy to me. Was I mistaken?

Peter_McLaughlinSpeaker
Score: 0 | 4 years ago | no reply

Hello Fuzzy Logic still has to be explicitly defined rather than being trained from a data set. It is therefore easier to modify than Machine Learning / Deep Learning but it can't tackle as complex problems. It is still heavily used in control system decision making.

mpuchlik
Score: 0 | 4 years ago | 1 reply

Hi, great talk! What kind of software or libraries are you using while dealing with machine learning especially deep learning? Cheers!

Peter_McLaughlinSpeaker
Score: 0 | 4 years ago | no reply

Hello, the main libraries are TensorFlow (and Keras which wraps TensorFlow), PyTorch and Caffe. Python is typically used for training and C++ for inference.

SlightlyChaotic
Score: 0 | 4 years ago | 1 reply

And my third (and final?) question: In the "Machine Learning vs. Deep Learning" slide at 24:53, you have the neural network type diagram under the deep learning. To give me something concrete, could you give some simple actual examples of what might be in a few of those decision bubbles?
Thanks again, it was the best introductory exposure to machine learning vs deep learning I have had so far.

Peter_McLaughlinSpeaker
Score: 0 | 4 years ago | no reply

Hello, the nodes of a neural network contain weights for each of the inputs to that node, followed by an "activation function" which determines the node's output value. Here are some examples of activation functions: https://en.wikipedia.org/wiki/Activation_function.

Peter_McLaughlinSpeaker
Score: 0 | 4 years ago | no reply

Thank you for watching. Apologies if the audio quality was bad at times.

marek_klemes
Score: 0 | 4 years ago | no reply

Will the presentation charts become available soon?

dcomer
Score: 0 | 4 years ago | 1 reply

Had issues hearing the speaker. Someone was speaking in the background.

Stephane.Boucher
Score: 0 | 4 years ago | 1 reply

Do you mean in the video itself or in your own physical environment?

dcomer
Score: 0 | 4 years ago | 1 reply

In the video itself. To verify, I paused and unpaused several times and the background voice tracked. It didn't seem like part of the presentation as it was interrupting. I could be wrong. I was trying to listen in the background. Maybe the speaker was using audio for his input?

Stephane.Boucher
Score: 0 | 4 years ago | 1 reply

Let me know if you ever locate the timestamps for these, I'm curious. Thanks.

dcomer
Score: 0 | 4 years ago | 1 reply

I will re-listen later this afternoon and let you know. Did you have a chance to read my email about access to the 2020 conference?

dcomer
Score: 0 | 4 years ago | no reply

I relisten to the section I heard the background talking. It was about 32min into the video. However, relistening to the video at that point the background talking was not present.. It should similar to what you would experience if someone had not muted their mic. in a Zoom presentation. However, since this was not Zoom, and pre-recorded, I have to assume 1) an abnormality, or 2) I heard a voice in my head? Hard to believe, but pressing the pause/play and having the voice track, maybe some metaphysical connection? :~V

SlightlyChaotic
Score: 0 | 4 years ago | no reply

And can you provide the slides for download?

Stephane.Boucher
Score: 0 | 4 years ago | no reply

The Q&A for this session had to be canceled. Please feel free to ask questions here.