Manufacturing & Industrial Engineering Faculty Publications
Document Type
Article
Publication Date
3-9-2026
Abstract
Ensuring reliable anomaly detection in industrial robots is critical for safe and autonomous manufacturing operations. However, it remains challenging due to temporal dependencies and class imbalance in sensor data. This study presents a reinforcement learning approach using Deep Q-Network (DQN) enhanced with Long Short-Term Memory (LSTM) and Gradient Boosting Machine (GBM) for robust anomaly detection in robotic systems. The proposed framework integrates an LSTM into the DQN policy to capture temporal patterns. It also introduces a novel GBM-based reward mechanism that mitigates class imbalance by applying SMOTE (Synthetic Minority Over-sampling Technique) after removing temporal dependencies. Experimental results demonstrate that this hybrid DQN-GBM framework achieves superior performance in precision, recall, and F1-score compared to standalone DQN and DQN-LSTM variants. Beyond technical improvements, this approach enables truly autonomous manufacturing environments by providing adaptive, real-time anomaly detection that reduces human intervention and prevents costly production failures, ultimately contributing to more resilient and self-optimizing industrial systems.
Recommended Citation
Messaoudi, Salma, Ahmed Bendaouia, El Hassan Abdelwahed, Mohammed Ameksa, Hajar Mousannif, and Jianzhi Li. "Reinforcement Learning for Imbalanced Data in Robotic Anomaly Detection within Autonomous Manufacturing." SN Computer Science 7, no. 3 (2026): 266. https://doi.org/10.1007/s42979-026-04858-0
Creative Commons License

This work is licensed under a Creative Commons Attribution 4.0 International License.
Publication Title
SN Computer Science
DOI
10.1007/s42979-026-04858-0

Comments
This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made.