Kolmogorov-Arnold Networks: MLP vs KAN, Math, B-Splines, Universal Approximation Theorem
In this video, I will be explaining Kolmogorov-Arnold Networks, a new type of network that was presented in the paper "KAN: Kolmogorov-Arnold Networks" by Liu et al.
I will start the video by reviewing Multilayer Perceptrons, to show how the typical Linear layer works in a neural network. I will then introduce the concept of data fitting, which is necessary to understand Bézier Curves and then B-Splines.
Before introducing Kolmogorov-Arnold Networks, I will also explain what is the Universal Approximation Theorem for Neural Networks and its equivalent for Kolmogorov-Arnold Networks called Kolm…
Watch on YouTube ↗
(saves to browser)
Chapters (14)
Introduction
1:10
Multilayer Perceptron
11:08
Introduction to data fitting
15:36
Bézier Curves
28:12
B-Splines
40:42
Universal Approximation Theorem
45:10
Kolmogorov-Arnold Representation Theorem
46:17
Kolmogorov-Arnold Networks
51:55
MLP vs KAN
55:20
Learnable functions
58:06
Parameters count
1:00:44
Grid extension
1:03:37
Interpretability
1:10:42
Continual learning
DeepCamp AI