Function Approximation: Neural Network vs Kernel Regression

True function
Neural Network
Kernel Regression (NTK)
Training Points

Kernel Matrix (NTK)

Training Convergence

Key Insight: The Neural Tangent Kernel (NTK) theory shows that infinitely wide neural networks trained with gradient descent are equivalent to kernel regression with a deterministic kernel. As network width increases, the NTK stays nearly constant during training, and the network's predictions converge to those of kernel regression. This demo visualizes this correspondence.

Network Architecture

Training

NTK Kernel

Data

NN Loss: -
Kernel Loss: -
Prediction MSE: -
Epoch: 0