all AI news
[R] KAN: Kolmogorov-Arnold Networks
May 1, 2024, 5:03 p.m. | /u/SeawaterFlows
Machine Learning www.reddit.com
**Code**: [https://github.com/KindXiaoming/pykan](https://github.com/KindXiaoming/pykan)
**Quick intro**: [https://kindxiaoming.github.io/pykan/intro.html](https://kindxiaoming.github.io/pykan/intro.html)
**Documentation**: [https://kindxiaoming.github.io/pykan/](https://kindxiaoming.github.io/pykan/)
**Abstract**:
>Inspired by the Kolmogorov-Arnold representation theorem, we propose **Kolmogorov-Arnold Networks** (**KANs**) as promising alternatives to Multi-Layer Perceptrons (MLPs). While MLPs have *fixed* activation functions on *nodes* ("neurons"), KANs have *learnable* activation functions on *edges* ("weights"). KANs have no linear weights at all -- every weight parameter is replaced by a univariate function parametrized as a spline. We show that this seemingly simple change makes KANs outperform MLPs in terms of …
abstract every function functions layer linear machinelearning networks neurons nodes representation show spline theorem while
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York