Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
129
result(s) for
"Kolmogorov–Arnold networks"
Sort by:
KAN–EEG: towards replacing backbone–MLP for an effective seizure detection system
by
Huang, Zhaojing
,
Kavehei, Omid
,
Nikpour, Armin
in
Adaptability
,
Artificial intelligence
,
Datasets
2025
The landscape of artificial intelligence (AI) research is witnessing a transformative shift with the emergence of the Kolmogorov–Arnold network (KAN), presenting a novel architectural paradigm aimed to redefine the structural foundations of AI models, which are based on multilayer perceptron (MLP). Through rigorous experimentation and evaluation, we introduce the KAN–electroencephalogram (EEG) model, a tailored design for efficient seizure detection. Our proposed network is tested and successfully generalized on three different datasets, one from the USA, one from Europe, and one from Oceania, recorded with different front-end hardware. All datasets are scalp EEG in adults and are from patients living with epilepsy. Our empirical findings reveal that while both architectures demonstrate commendable performance in seizure detection, the KAN model exhibits high-level out-of-sample generalization across datasets from diverse geographical regions, underscoring its inherent efficacy and adaptability at the backbone level. Furthermore, we demonstrate the resilience of the KAN architecture to model size reduction and shallow network configurations, highlighting its versatility and efficiency by preventing over-fitting in-sample datasets. This study advances our understanding of innovative neural network architectures and underscores the pioneering potential of KANs in critical domains such as medical diagnostics.
Journal Article
Multifidelity Kolmogorov–Arnold networks
by
Stinis, Panos
,
Howard, Amanda A
,
Jacob, Bruno
in
Accuracy
,
Kolmogorov–Arnold networks
,
multifidelity
2025
We develop a method for multifidelity Kolmogorov–Arnold networks (KANs), which use a low-fidelity model along with a small amount of high-fidelity data to train a model for the high-fidelity data accurately. Multifidelity KANs (MFKANs) reduce the amount of expensive high-fidelity data needed to accurately train a KAN by exploiting the correlations between the low- and high-fidelity data to give accurate and robust predictions in the absence of a large high-fidelity dataset. In addition, we show that MFKANs can be used to increase the accuracy of physics-informed KANs, without the use of training data.
Journal Article
SPIKANs: separable physics-informed Kolmogorov–Arnold networks
by
Stinis, Panos
,
Howard, Amanda A
,
Jacob, Bruno
in
Complexity
,
Computation
,
Computer architecture
2025
Physics-Informed Neural Networks (PINNs) have emerged as a promising method for solving partial differential equations (PDEs) in scientific computing. While PINNs typically use multilayer perceptrons (MLPs) as their underlying architecture, recent advancements have explored alternative neural network structures. One such innovation is the Kolmogorov–Arnold Network (KAN), which has demonstrated benefits over traditional MLPs, including faster neural scaling and better interpretability. The application of KANs to physics-informed learning has led to the development of Physics-Informed KANs (PIKANs), enabling the use of KANs to solve PDEs. However, despite their advantages, KANs often suffer from slower training speeds, particularly in higher-dimensional problems where the number of collocation points grows exponentially with the dimensionality of the system. To address this challenge, we introduce Separable Physics-Informed Kolmogorov–Arnold Networks (SPIKANs). This novel architecture applies the principle of separation of variables to PIKANs, decomposing the problem such that each dimension is handled by an individual KAN. This approach drastically reduces the computational complexity of training without sacrificing accuracy, facilitating their application to higher-dimensional PDEs. Through a series of benchmark problems, we demonstrate the effectiveness of SPIKANs, showcasing their superior scalability and performance compared to PIKANs and highlighting their potential for solving complex, high-dimensional PDEs in scientific computing.
Journal Article
Uncertainty quantification with Bayesian higher order ReLU-KANs
2025
We introduce the first method of uncertainty quantification in the domain of Kolmogorov–Arnold Networks, specifically focusing on (Higher Order) ReLU-KANs to enhance computational efficiency given the computational demands of Bayesian methods. The method we propose is general in nature, providing access to both epistemic and aleatoric uncertainties. It is also capable of generalization to other various basis functions. We validate our method through a series of closure tests, commonly found in the KAN literature, including simple one-dimensional functions and application to the domain of (stochastic) partial differential equations. Referring to the latter, we demonstrate the method’s ability to correctly identify functional dependencies introduced through the inclusion of a stochastic term.
Journal Article
Accurately Models the Relationship Between Physical Response and Structure Using Kolmogorov–Arnold Network
2025
Artificial intelligence (AI) in science is a key area of modern research. However, many current machine learning methods lack interpretability, making it difficult to grasp the physical mechanisms behind various phenomena, which hampers progress in related fields. This study focuses on the Poisson's ratio of a hexagonal lattice elastic network as it varies with structural deformation. By employing the Kolmogorov–Arnold Network (KAN), the transition of the network's Poisson's ratio from positive to negative as the hexagonal structural element shifts from a convex polygon to a concave polygon was accurately predicted. The KAN provides a clear mathematical framework that describes this transition, revealing the connection between the Poisson's ratio and the geometric properties of the hexagonal element, and accurately identifying the geometric parameters at which the Poisson's ratio equals zero. This work demonstrates the significant potential of the KAN network to clarify the mathematical relationships that underpin physical responses and structural behaviors.
The present study underscores the remarkable capabilities of the Kolmogorov‐Arnold Network (KAN) in formula fitting for physical systems. KAN predicts the relationship between the Poisson's ratio of a metamaterial structure and its geometric parameters, demonstrating the transition of Poisson's ratio from positive to negative as the hexagonal component changes shape, illustrating its adaptability in analyzing complex systems.
Journal Article
A Local–Global Graph KAN for Multi‐Class Prediction of PPI
by
Liu, Minghui
,
Qu, Ying
in
asymmetric loss function
,
global graphs and local subgraphs
,
Graph Kolmogorov‐Arnold Network
2025
ABSTRACT
Traditional experimental methods for identifying protein–protein interactions (PPI) are expensive and time‐consuming. Therefore, using machine learning to treat multiple PPI predictions as binary classifications has become an alternative, but there is a problem of data imbalance. The proposed GLGKAN‐PPI method integrates features from both global graphs and local subgraphs to capture the complex structural information of PPI networks comprehensively. Specifically, the method utilizes the pre‐trained model MASSA to extract multimodal features of proteins. The global graph features are extracted using the GKAN (Graph Kolmogorov‐Arnold Network) algorithm. Meanwhile, the local subgraph features are extracted using the MOE‐GKAN (Mixture of Experts‐Graph Kolmogorov‐Arnold Network) algorithm. To mitigate data imbalance, an asymmetric loss function is utilized to better handle minority classes and improve overall prediction accuracy. Experimental results demonstrate that GLGKAN‐PPI outperforms a range of existing intelligent approaches across multiple datasets and partitioning strategies.
The GLGKAN‐PPI method integrates global and local graph features with multimodal protein attributes to enhance multi‐class protein–protein interaction prediction. Using an asymmetric loss function to address class imbalance, it achieves superior accuracy across datasets, highlighting the importance of combined structural and sequence information for robust PPI prediction.
Journal Article
HyperKAN: Kolmogorov–Arnold Networks Make Hyperspectral Image Classifiers Smarter
by
Butt, Muhammad A.
,
Khonina, Svetlana
,
Khabibullin, Roman
in
Accuracy
,
Classification
,
Comparative analysis
2024
In traditional neural network designs, a multilayer perceptron (MLP) is typically employed as a classification block following the feature extraction stage. However, the Kolmogorov–Arnold Network (KAN) presents a promising alternative to MLP, offering the potential to enhance prediction accuracy. In this paper, we studied KAN-based networks for pixel-wise classification of hyperspectral images. Initially, we compared baseline MLP and KAN networks with varying numbers of neurons in their hidden layers. Subsequently, we replaced the linear, convolutional, and attention layers of traditional neural networks with their KAN-based counterparts. Specifically, six cutting-edge neural networks were modified, including 1D (1DCNN), 2D (2DCNN), and 3D convolutional networks (two different 3DCNNs, NM3DCNN), as well as transformer (SSFTT). Experiments conducted using seven publicly available hyperspectral datasets demonstrated a substantial improvement in classification accuracy across all the networks. The best classification quality was achieved using a KAN-based transformer architecture.
Journal Article
How to Learn More? Exploring Kolmogorov–Arnold Networks for Hyperspectral Image Classification
by
Ghamisi, Pedram
,
Hong, Danfeng
,
Jamali, Ali
in
Accuracy
,
Algorithms
,
Artificial neural networks
2024
Convolutional neural networks (CNNs) and vision transformers (ViTs) have shown excellent capability in complex hyperspectral image (HSI) classification. However, these models require a significant number of training data and are computational resources. On the other hand, modern Multi-Layer Perceptrons (MLPs) have demonstrated a great classification capability. These modern MLP-based models require significantly less training data compared with CNNs and ViTs, achieving state-of-the-art classification accuracy. Recently, Kolmogorov–Arnold networks (KANs) were proposed as viable alternatives for MLPs. Because of their internal similarity to splines and their external similarity to MLPs, KANs are able to optimize learned features with remarkable accuracy, in addition to being able to learn new features. Thus, in this study, we assessed the effectiveness of KANs for complex HSI data classification. Moreover, to enhance the HSI classification accuracy obtained by the KANs, we developed and proposed a hybrid architecture utilizing 1D, 2D, and 3D KANs. To demonstrate the effectiveness of the proposed KAN architecture, we conducted extensive experiments on three newly created HSI benchmark datasets: QUH-Pingan, QUH-Tangdaowan, and QUH-Qingyun. The results underscored the competitive or better capability of the developed hybrid KAN-based model across these benchmark datasets over several other CNN- and ViT-based algorithms, including 1D-CNN, 2DCNN, 3D CNN, VGG-16, ResNet-50, EfficientNet, RNN, and ViT.
Journal Article
Exploring Kolmogorov–Arnold Networks for Interpretable Time Series Classification
2025
Time‐series classification is a relevant step supporting decision‐making processes in various domains, and deep neural models have shown promising performance in this respect. Despite significant advancements in deep learning, the theoretical understanding of how and why complex architectures function remains limited, prompting the need for more interpretable models. Recently, the Kolmogorov–Arnold Networks (KANs) have been proposed as a more interpretable alternative to deep learning. While KAN‐related research is significantly rising, to date, the study of KAN architectures for time‐series classification has been limited. In this paper, we aim to conduct a comprehensive and robust exploration of the KAN architecture for time‐series classification utilizing 117 datasets from UCR benchmark archive, from multiple different domains. More specifically, we investigate (a) the transferability of reference architectures designed for regression to classification tasks, (b) the hyperparameter and implementation configurations for an architecture that best generalizes across 117 datasets, (c) the associated complexity trade‐offs, and (d) KANs interpretability. Our results demonstrate that (1) the Efficient KAN outperforms MLPs in both performance and training times, showcasing its suitability for classification tasks. (2) Efficient KAN exhibits greater stability than the original KAN across grid sizes, depths, and layer configurations, especially when lower learning rates are employed. (3) KAN achieves competitive accuracy compared to state‐of‐the‐art models such as HIVE‐COTE2 and InceptionTime, while maintaining smaller architectures and faster training times, highlighting its favorable balance of performance and transparency. (4) The interpretability of the KAN model, as confirmed by SHAP analysis, reinforces its capacity for transparent decision‐making.
Journal Article
PFSKANs: A Novel Pixel-Level Feature Selection Model Based on Kolmogorov–Arnold Networks
2025
Inspired by the interpretability of Kolmogorov–Arnold Networks (KANs), a novel Pixel-level Feature Selection (PFS) model based on KANs (PFSKANs) is proposed as a fundamentally distinct alternative from trainable Convolutional Neural Networks (CNNs) and transformers in the computer vision tasks. We modify the simplification techniques of KANs to detect key pixels with high contribution scores directly at the input image. Specifically, a trainable selection procedure is intuitively visualized and performed only once, since the obtained interpretable pixels can subsequently be identified and dimensionally standardized using the proposed mathematical approach. Experiments on the image classification tasks using the MNIST, Fashion-MNIST, CIFAR-10, and CIFAR-100 datasets demonstrate that PFSKANs achieve comparable performance to CNNs in terms of accuracy, parameter efficiency, and training time.
Journal Article