Kolmogorov-Arnold Networks: MLP vs KAN, Math, B-Splines, Universal Approximation Theorem

Umar Jamil · Beginner ·📐 ML Fundamentals ·1y ago
In this video, I will be explaining Kolmogorov-Arnold Networks, a new type of network that was presented in the paper "KAN: Kolmogorov-Arnold Networks" by Liu et al. I will start the video by reviewing Multilayer Perceptrons, to show how the typical Linear layer works in a neural network. I will then introduce the concept of data fitting, which is necessary to understand Bézier Curves and then B-Splines. Before introducing Kolmogorov-Arnold Networks, I will also explain what is the Universal Approximation Theorem for Neural Networks and its equivalent for Kolmogorov-Arnold Networks called Kolm…
Watch on YouTube ↗ (saves to browser)

Chapters (14)

Introduction
1:10 Multilayer Perceptron
11:08 Introduction to data fitting
15:36 Bézier Curves
28:12 B-Splines
40:42 Universal Approximation Theorem
45:10 Kolmogorov-Arnold Representation Theorem
46:17 Kolmogorov-Arnold Networks
51:55 MLP vs KAN
55:20 Learnable functions
58:06 Parameters count
1:00:44 Grid extension
1:03:37 Interpretability
1:10:42 Continual learning
Calculating Pi with Skittles (and census data!?) - Numberphile
Next Up
Calculating Pi with Skittles (and census data!?) - Numberphile
Numberphile