Proceedings of ESCAPE 35ISSN: 2818-4734
Volume: 4 (2025)
Table of Contents
LAPSE:2025.0372
Published Article
LAPSE:2025.0372
Kolmogorov Arnold Networks (KANs) as surrogate models for global process optimization
Tanuj Karia, Giacomo Lastrucci, Artur M. Schweidtmann
June 27, 2025
Abstract
Surrogate models are widely used to improve the tractability of process optimization. Some commonly used surrogate models are obtained via machine learning such as multi-layer perceptrons (MLPs), Gaussian processes, and decision trees. Recently, a new class of machine learning models named Kolmogorov Arnold Networks (KANs) have been proposed. Broadly, KANs are similar to MLPs, yet they are based on the Kolmogorov representation theorem instead of the universal approximation theorem for the MLPs. Compared to MLPs, it was reported that KANs require significantly fewer parameters to approximate a given input/output relationship. One of the bottlenecks preventing the embedding of MLPs into optimization formulations is that MLPs with a high number of parameters (larger width or depth) are more challenging to globally optimize. We investigate whether the parameter efficiency of KANs relative to MLPs can be translated to computational benefits when embedding them into optimization problems and solving them to global optimality. We propose a Mixed-Integer Nonlinear Programming (MINLP) formulation of a KAN and test its effectiveness using a case study on the optimization of an auto-thermal reforming process. We observe that KANs are a promising alternative as surrogate models particularly for cases with low number of inputs or outputs. For surrogate models with higher inputs or outputs, stronger formulations must be developed to improve global optimization of models with KANs embedded.
Keywords
Deterministic Global Optimization, Kolmogorov Arnold Networks, Mixed-Integer Nonlinear Programming, Surrogate modeling
Suggested Citation
Karia T, Lastrucci G, Schweidtmann AM. Kolmogorov Arnold Networks (KANs) as surrogate models for global process optimization. Systems and Control Transactions 4:1371-1376 (2025) https://doi.org/10.69997/sct.195815
Author Affiliations
Karia T: Process Intelligence Research Group, Department of Chemical Engineering, Delft University of Technology, Van der Maasweg 9, 2629 HZ, Delft, The Netherlands
Lastrucci G: Process Intelligence Research Group, Department of Chemical Engineering, Delft University of Technology, Van der Maasweg 9, 2629 HZ, Delft, The Netherlands
Schweidtmann AM: Process Intelligence Research Group, Department of Chemical Engineering, Delft University of Technology, Van der Maasweg 9, 2629 HZ, Delft, The Netherlands
Journal Name
Systems and Control Transactions
Volume
4
First Page
1371
Last Page
1376
Year
2025
Publication Date
2025-07-01
Version Comments
Original Submission
Other Meta
PII: 1371-1376-1255-SCT-4-2025, Publication Type: Journal Article
Record Map
Published Article

LAPSE:2025.0372
This Record
External Link

https://doi.org/10.69997/sct.195815
Article DOI
Download
Files
Jun 27, 2025
Main Article
License
CC BY-SA 4.0
Meta
Record Statistics
Record Views
906
Version History
[v1] (Original Submission)
Jun 27, 2025
 
Verified by curator on
Jun 27, 2025
This Version Number
v1
Citations
Most Recent
This Version
URL Here
http://psecommunity.org/LAPSE:2025.0372
 
Record Owner
PSE Press
Links to Related Works
Directly Related to This Work
Article DOI
References Cited
  1. Misener R, Biegler L. Formulating data-driven surrogate models for process optimization. Comput Chem Eng. 2023 Nov 1;179:108411 https://doi.org/10.1016/j.compchemeng.2023.108411
  2. Schweidtmann AM, Bongartz D, Grothe D, Kerkenhoff T, Lin X, Najman J, et al. Deterministic global optimization with Gaussian processes embedded. Math Program Comput [Internet]. 2021 Sep 1 [cited 2025 Jan 29];13(3):553-81. Available from: https://link.springer.com/article/10.1007/s12532-021-00204-y https://doi.org/10.1007/s12532-021-00204-y
  3. Schweidtmann AM, Weber JM, Wende C, Netze L, Mitsos A. Obey validity limits of data-driven models through topological data analysis and one-class classification. Optimization and Engineering [Internet]. 2022 Jun 1 [cited 2025 Jan 29];23(2):855-76. Available from: https://link.springer.com/article/10.1007/s11081-021-09608-0 https://doi.org/10.1007/s11081-021-09608-0
  4. Thebelt A, Kronqvist J, Mistry M, Lee RM, Sudermann-Merx N, Misener R. ENTMOOT: A framework for optimization over ensemble tree models. Comput Chem Eng. 2021 Aug 1;151:107343 https://doi.org/10.1016/j.compchemeng.2021.107343
  5. Schweidtmann AM, Bongartz D, Mitsos A. Optimization with Trained Machine Learning Models Embedded. Encyclopedia of Optimization [Internet]. 2023 [cited 2025 Jan 24];1-8. Available from: https://link.springer.com/referenceworkentry/10.1007/978-3-030-54621-2_735-1 https://doi.org/10.1007/978-3-030-54621-2_735-1
  6. Schweidtmann AM, Esche E, Fischer A, Kloft M, Repke JU, Sager S, et al. Machine Learning in Chemical Engineering: A Perspective. Chemie Ingenieur Technik [Internet]. 2021 Dec 1 [cited 2025 Jan 29];93(12):2029-39. Available from: https://onlinelibrary.wiley.com/doi/full/10.1002/cite.202100083 https://doi.org/10.1002/cite.202100083
  7. Hornik K, Stinchcombe M, White H. Multilayer feedforward networks are universal approximators. Neural Networks. 1989 Jan 1;2(5):359-66 https://doi.org/10.1016/0893-6080(89)90020-8
  8. Schweidtmann AM, Mitsos A. Deterministic Global Optimization with Artificial Neural Networks Embedded. J Optim Theory Appl [Internet]. 2019 Mar 15 [cited 2025 Jan 24];180(3):925-48. Available from: https://link.springer.com/article/10.1007/s10957-018-1396-0 https://doi.org/10.1007/s10957-018-1396-0
  9. Webinar: Using Trained Machine Learning Predictors in Gurobi - YouTube [Internet]. [cited 2025 Jan 24]. Available from: https://www.youtube.com/watch?v=jaux5Oo4qHU
  10. Liu Z, Wang Y, Vaidya S, Ruehle F, Halverson J, Soljaci'c MS, et al. KAN: Kolmogorov-Arnold Networks. 2024 Apr 30 [cited 2025 Jan 24]; Available from: https://arxiv.org/abs/2404.19756v4
  11. Arnold VI. On the representation of functions of several variables as a superposition of functions of a smaller number of variables. Collected Works [Internet]. 2009 [cited 2025 Jan 24];25-46. Available from: https://link.springer.com/chapter/10.1007/978-3-642-01742-1_5 https://doi.org/10.1007/978-3-642-01742-1_5
  12. López-Flores FJ, Ramírez-Márquez C, Ponce-Ortega JM. Process Systems Engineering Tools for Optimization of Trained Machine Learning Models: Comparative and Perspective. Ind Eng Chem Res [Internet]. 2024 Aug 14 [cited 2025 Jan 24];63(32):13966-79. Available from: https://pubs.acs.org/doi/full/10.1021/acs.iecr.4c00632 https://doi.org/10.1021/acs.iecr.4c00632
  13. Bolusani S, Besançon M, Bestuzheva K, Chmiela A, Dionísio J, Donkiewicz T, et al. The SCIP Optimization Suite 9.0. 2024 Feb 27 [cited 2025 Jan 24]; Available from: https://arxiv.org/abs/2402.17702v2
  14. Grimstad B. A MIQCP formulation for B-spline constraints. Optim Lett [Internet]. 2018;12(4):713-25. Available from: https://doi.org/10.1007/s11590-017-1190-1
  15. Ceccon F, Jalving J, Haddad J, Thebelt A, Tsay C, Laird CD, et al. OMLT: Optimization & Machine Learning Toolkit. Journal of Machine Learning Research [Internet]. 2022 [cited 2025 Jan 27];23(349):1-8. Available from: http://jmlr.org/papers/v23/22-0277.html
  16. Bynum ML, Hackebeil GA, Hart WE, Laird CD, Nicholson BL, Siirola JD, et al. Pyomo - Optimization Modeling in Python. 2021 [cited 2025 Jan 27];67. Available from: http://link.springer.com/10.1007/978-3-030-68928-5 https://doi.org/10.1007/978-3-030-68928-5_5