ViKANformer replaces standard MLP blocks in Vision Transformers with Kolmogorov-Arnold expansions (SineKAN, Fast-KAN, etc.), leveraging the Kolmogorov-Arnold theorem for enhanced representational power. For full details on the methodology and experiments, please refer to the arXiv paper.
For conda:
$ conda create -n <env_name> For virtualenv:
$ python -m venv <env_name>where env_name will be the name of the directory that you will use as the virtual environment.
For conda:
$ conda install --file requirements.txtFor virualenv:
$ pip install -r requirements.txt$ python train.pyTo follow these steps correctly, make sure you are in the root directory of the repository.
Control variables:
- Dataset used: MNIST
- Transformations: None
- GPU Used: Tesla P100
If you find ViKANformer helpful in your research, please cite our paper:
@misc{s2025vikanformerembeddingkolmogorovarnold,
title={ViKANformer: Embedding Kolmogorov Arnold Networks in Vision Transformers for Pattern-Based Learning},
author={Shreyas S and Akshath M},
year={2025},
eprint={2503.01124},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2503.01124},
}We give credit to all the other papers and projects that we have referenced in order to write the paper. This paper is covered by the MIT License, granting full access to using this repository for any use whatsoever.