199070-Thumbnail Image.png
Description

The curse of dimensionality poses a significant challenge to modern multilayer perceptron-based architectures, often causing performance stagnation and scalability issues. Addressing this limitation typically requires vast amounts of data. In contrast, Kolmogorov-Arnold Networks have gained attention in the machine learning

The curse of dimensionality poses a significant challenge to modern multilayer perceptron-based architectures, often causing performance stagnation and scalability issues. Addressing this limitation typically requires vast amounts of data. In contrast, Kolmogorov-Arnold Networks have gained attention in the machine learning community for their bold claim of being unaffected by the curse of dimensionality. This paper explores the Kolmogorov-Arnold representation theorem and the mathematical principles underlying Kolmogorov-Arnold Networks, which enable their scalability and high performance in high-dimensional spaces. We begin with an introduction to foundational concepts necessary to understand Kolmogorov-Arnold Networks, including interpolation methods and Basis-splines, which form their mathematical backbone. This is followed by an overview of perceptron architectures and the Universal approximation theorem, a key principle guiding modern machine learning. This is followed by an overview of the Kolmogorov-Arnold representation theorem, including its mathematical formulation and implications for overcoming dimensionality challenges. Next, we review the architecture and error-scaling properties of Kolmogorov-Arnold Networks, demonstrating how these networks achieve true freedom from the curse of dimensionality. Finally, we discuss the practical viability of Kolmogorov-Arnold Networks, highlighting scenarios where their unique capabilities position them to excel in real-world applications. This review aims to offer insights into Kolmogorov-Arnold Networks' potential to redefine scalability and performance in high-dimensional learning tasks.

Reuse Permissions
  • 2.33 MB application/pdf

    Download restricted. Please sign in.

    Details

    Title
    • KAT to KANs: a review of Kolmogorov-Arnold Networks and the neural leap forward
    Contributors
    Date Created
    2024-11-15
    Keywords
    • Kolmogorov-Arnold Networks
    • Kolmogorov-Arnold Representation Theorem
    • Universal Approximation Theorem
    • Multi-layer Perceptrons
    Resource Type
  • Text
  • Collections this item is in
    Identifier

    Citation and reuse

    Cite this item

    This is a suggested citation. Consult the appropriate style guide for specific citation guidelines.

    Basina, D., Vishal, J. R., Choudhary, A., & Chakravarthi, B. (2024). KAT to KANs: A Review of Kolmogorov-Arnold Networks and the Neural Leap Forward. https://hdl.handle.net/2286/R.2.N.199070 [Preprint]

    Also available in arXiv as:

    Basina, D., Vishal, J. R., Choudhary, A., & Chakravarthi, B. (2024). KAT to KANs: A Review of Kolmogorov-Arnold Networks and the Neural Leap Forward (arXiv:2411.10622; Version 1). arXiv. https://doi.org/10.48550/arXiv.2411.10622

    Statement of Responsibility

    Divesh Basina, Joseph Raj Vishal, Aarya Choudhary, Bharatesh Chakravarthi
    Arizona State University

    Machine-readable links