SineKAN: A Flexible Machine Learning Model for Modelling $L^2$ Functions and Beyond
ORAL
Abstract
The multi-layer perceptron (MLP) is near-ubiquitous in modern neural network architectures. Recently, an alternative to the MLP, Kolmogorov-Arnold Networks (KANs), was proposed, motivated by the Kolmogorov-Arnold representation theorem. The representation theorem states that over a finite domain it is possible to model any continuous multi-variate function as a superposition of many continuous functions of a single variable. However, the theorem doesn't predict what the ideal functional form should be for such a representation leaving much room for investigation. Here, we present the SineKAN model which uses a functional form of sinusoidal functions with learnable amplitudes, frequencies, and biases. We find that these models can successfully model $L^2$ functions (which includes normalized wave functions), and functions bounded over a finite domain. We further find that these functions can achieve competitive results on several tasks when compared with MLP and basis-spline KAN models at competitive sizes and speeds.
–
Publication: @misc{reinhardt2024sinekankolmogorovarnoldnetworksusing,
title={SineKAN: Kolmogorov-Arnold Networks Using Sinusoidal Activation Functions},
author={Eric A. F. Reinhardt and P. R. Dinesh and Sergei Gleyzer},
year={2024},
eprint={2407.04149},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2407.04149},
}
Presenters
-
Eric Allen Friss Reinhardt
University of Alabama
Authors
-
Eric Allen Friss Reinhardt
University of Alabama
-
Sergei V Gleyzer
University of Alabama
-
Dinesh P. R.
University of Alabama
-
Hanh V Nguyen
University of Alabama