Development of a reinforcement learning-based adaptive scheduling algorithm for commercial smart kitchens
Article Sidebar
Open full text
Issue Vol. 15 No. 3 (2025)
-
Objects features extraction by singular projections of data tensor to matrices
Yuriy Bunyak, Roman Kvуetnyy, Olga Sofina, Volodymyr Kotsiubynskyi5-9
-
A hybrid approach combining generalized normal distribution optimization algorithm and fuzzy C-means with Calinski-Harabasz index for clustering optimization
Moatasem Mahmood Ibrahim, Omar Saber Qasim, Talal Fadhil Hussein10-14
-
Method of hybrid logical classification trees based on group selection of discrete features
Igor Povkhan, Andrii Leheza, Oksana Mulesa, Olena Melnyk, Aliya Kintonova15-21
-
Optimal control of electric energy quality, based on lexicographic approach
Yurii Voitiuk, Anatolii Volotskyi, Iuliia Shullie, Maiia Kovalchuk, Laura Duissembayeva22-28
-
Implementation of energy-saving modes for electro-radiation drying of oil-containing material using automation tools
Borys Kotov, Roman Kalinichenko, Serhii Stepanenko, Vasyl Lukach, Volodymyr Hryshchenko, Alvian Kuzmych, Yurii Pantsyr, Ihor Garasymchuk, Volodymyr Vasylyuk29-32
-
Methods and means of laser layer Jones matrix mapping of polycrystalline films of biological fluids
Olexander Ushenko, Iryna Soltys, Olexander Dubolazov, Sergii Pavlov, Vasyl Garasym, Alona Kolomiiets, Bakhyt Yeraliyeva33-37
-
Alzheimer’s disease classification from MRI using vision transformer
Mohith Reddy Kandi, Sree Vijaya Lakshmi Kothapalli, Sivamsh Pavan Rajanala, Suvarna Vani Koneru, Vishnu Pramukh Vattikunta38-44
-
Enhancing early Parkinson’s disease diagnosis through handwriting analysis
Asma Ouabd, Abdelilah Jilbab, Achraf Benba, Ahmed Hammouch45-49
-
Biomechanical foundations and benefits of active orthoses in the treatment of idiopathic scoliosis
Patrycja Tymińska-Wójcik50-54
-
Application of facial recognition technologies for enhancing control in information security systems
Nurzhigit Smailov, Rashida Kadyrova, Kamila Abdulina, Fatima Uralova, Nurgul Kubanova, Akezhan Sabibolda55-58
-
IoT system with frequency converters of physical quantities on FPGA
Oleksandr V. Osadchuk, Iaroslav O. Osadchuk, Valentyn K. Skoshchuk59-66
-
Research on the possibility of reducing the error in measuring the phase shift of radio signals
Sergey Matvienko, Grygoriy Tymchyk, Konstantin Vonsevych, Nataliia Stelmakh67-72
-
Implementation of fiber-optic sensing systems in structural health monitoring of concrete
Nurzhigit Smailov, Akmaral Tolemanova, Amir Aziskhan, Beibarys Sekenov, Akezhan Sabibolda73-76
-
Modelling the working cycle of a heat pump scroll compressor
Bohdan Sydorchuk, Oleksandr Naumchuk77-80
-
Modeling of interception parking lots
Larysa Gumeniuk, Volodymyr Lotysh, Pavlo Humeniuk, Oleksandr Reshetylo, Yuriy Syrota81-86
-
Development and research of W-parameters of potentially unstable four-poles based on the mathematical model of W-parameters of field-effect transistors in the high-frequency range
Oleksandr Voznyak, Kateryna Kovalova, Yurii Polievoda, Liudmyla Kolianovska, Svitlana Ovsienko, Alla Solomon87-90
-
Detection confidential information by large language models
Oleh Deineka, Oleh Harasymchuk, Andrii Partyka, Yurii Dreis, Yuliia Khokhlachova, Yuriy Pepa91-99
-
Ethical simulation of a phishing attack
Justyna Kęczkowska, Karol Wykrota, Mirosław Płaza100-104
-
The effectiveness of machine learning in detecting phishing websites
Jacek Łukasz Wilk-Jakubowski, Aleksandra Sikora, Dawid Maciejski105-109
-
Contemporary approaches to integrating AI agents into library information processes
Mariia Sokil, Andriy Andrukhiv110-116
-
Development of a reinforcement learning-based adaptive scheduling algorithm for commercial smart kitchens
Konrad Kabala, Piotr Dziurzanski, Agnieszka Konrad117-122
-
Optimizing deep learning techniques with stacking BiLSTM and BiGRU models for gold price prediction
Iqbal Kharisudin, Nike Yustina Oktaviani123-130
-
Websites with virtual church tours in Poland – usability and accessibility analysis
Michał Mitura, Mariusz Dzieńkowski131-137
-
Study of feed granulation process based on system analysis – justification of optimization criteria
Mahil Isa Mammadov138-142
Archives
-
Vol. 15 No. 3
2025-09-30 24
-
Vol. 15 No. 2
2025-06-27 24
-
Vol. 15 No. 1
2025-03-31 26
-
Vol. 14 No. 4
2024-12-21 25
-
Vol. 14 No. 3
2024-09-30 24
-
Vol. 14 No. 2
2024-06-30 24
-
Vol. 14 No. 1
2024-03-31 23
-
Vol. 13 No. 4
2023-12-20 24
-
Vol. 13 No. 3
2023-09-30 25
-
Vol. 13 No. 2
2023-06-30 14
-
Vol. 13 No. 1
2023-03-31 12
-
Vol. 12 No. 4
2022-12-30 16
-
Vol. 12 No. 3
2022-09-30 15
-
Vol. 12 No. 2
2022-06-30 16
-
Vol. 12 No. 1
2022-03-31 9
-
Vol. 11 No. 4
2021-12-20 15
-
Vol. 11 No. 3
2021-09-30 10
-
Vol. 11 No. 2
2021-06-30 11
-
Vol. 11 No. 1
2021-03-31 14
Main Article Content
DOI
Authors
Abstract
Reinforcement learning (RL) is a machine learning method in which a model optimizes its decision-making strategy based on rewards or penalties received for the actions it takes in an environment, often simulated. An example of an optimized process could be work scheduling in a restaurant, with the cost function being the absolute error of the difference between the scheduled and actual delivery times of an order. In task planning, RL stands out for its ability to handle problems requiring a complex sequence of actions, where traditional planning algorithms may struggle. RL models can effectively explore the solution space, adjusting their decisions to changing conditions, which enables dynamic and adaptive task execution management. RL is a broad class encompassing various approaches to achieving a goal, and in this research, we focus on selected ones. Three popular RL methods named DQN, SARSA and TD-AC have been implemented and evaluated. The study was conducted in a simulated environment designed to replicate a "delivery-based" restaurant business model. The kitchen simulation model has been developed based on 65,845 recorded food preparation processes performed in 30 restaurants located throughout Poland. A rule-based, queue-driven model (FIFO) served as the baseline for absolute quality comparison of the generated schedules. The results show that, for the defined problem, the quality of the scheduling outcomes varies significantly depending on the choice of learning algorithm. Notably, the hybrid approach performed best under simulation conditions, considerably reducing the total completion time in a scenario reflecting the operations of a small, typical restaurant.
Keywords:
References
[1] Deguchi A. et al.: What is society 5.0. Society 5.0, 2020, 1–24. DOI: https://doi.org/10.1007/978-981-15-2989-4_1
[2] Dziurzanski P., Zhao S., Indrusiak L. S.: Integrated Process Planning and Scheduling in Commercial Smart Kitchens. arXiv preprint arXiv:1910.03322, 2019.
[3] Dziurzanski P., Kabala K., Konrad A.: A stochastic interval algebra for smart factory process. Informatyka, Automatyka, Pomiary w Gospodarce i Ochronie Środowiska – IAPGOS 1, 2025, 33–38. DOI: https://doi.org/10.35784/iapgos.6555
[4] Hamada R., et al.: Cooking navi: assistant for daily cooking in kitchen. 13th annual ACM international conference on Multimedia. 2005. DOI: https://doi.org/10.1145/1101149.1101228
[5] Kiesel J.: The internet of things in restaurants. 2018 [https://www.manufacturingtomorrow.com/article/2017/03/the-internet-of-things-in-restaurants/9261].
[6] Lei J., Ren X., Fox D.: Fine-grained kitchen activity recognition using rgb-d. ACM Conference on Ubiquitous Computing. 2012. DOI: https://doi.org/10.1145/2370216.2370248
[7] Mizrahi M., et al.: Digital gastronomy: Methods & recipes for hybrid cooking. 29th Annual Symposium on User Interface Software and Technology. 2016. DOI: https://doi.org/10.1145/2984511.2984528
[8] Mohan A., Zhang A., Lindauer M.: Structure in Deep Reinforcement Learning: A Survey and Open Problems. Journal of Artificial Intelligence Research 79, 2024, 1167–1236. DOI: https://doi.org/10.1613/jair.1.15703
[9] Moritz P., et al.: Ray: A distributed framework for emerging {AI} applications. 13th USENIX symposium on operating systems design and implementation –OSDI 18. 2018.
[10] Padakandla S.: A survey of reinforcement learning algorithms for dynamically varying environments. ACM Computing Surveys – CSUR 54(6), 2021, 1–25. DOI: https://doi.org/10.1145/3459991
[11] Sutton R. S., Barto A. G.: Reinforcement learning: An introduction. MIT press, 2018.
[12] Ustundag A., Cevikcan E.: Industry 4.0: managing the digital transformation. Springer, 2017. DOI: https://doi.org/10.1007/978-3-319-57870-5
[13] Van Hasselt H., Guez A., Silver D.: Deep reinforcement learning with double q-learning. AAAI conference on artificial intelligence 30(1), 2016. DOI: https://doi.org/10.1609/aaai.v30i1.10295
[14] Iterative.ai. Data Version Control (DVC). Iterative.ai, 2025 [http://dvc.org] (accessed: 21.08.2025).
[15] MLflow Project (part of LF Projects, LLC). MLflow. MLflow Project, 2025 [http://mlflow.org] (accessed: 21.08.2025).
[16] Project Jupyter. Project Jupyter. Project Jupyter, 2025 [http://jupyter.org] (accessed: 21.08.2025).
[17] PyTorch Foundation. PyTorch. PyTorch Foundation, 2025 [http://pytorch.org] (accessed: 21.08.2025).
[18] Rynek wewnętrzny w 2022 r., Analizy statystyczne. Główny Urząd Statystyczny. 2023.
Article Details
Abstract views: 124

