|
JAIST Repository >
b. 情報科学研究科・情報科学系 >
b10. 学術雑誌論文等 >
b10-1. 雑誌掲載論文 >
このアイテムの引用には次の識別子を使用してください:
https://hdl.handle.net/10119/20347
|
| タイトル: | Explainable and transferable deep reinforcement learning for adaptive patrol of rail-guided robot system |
| 著者: | Lee, Hosun Kwon, Jaesung Chong, Nak Young Yang, Woosung |
| キーワード: | Rail-guided patrol robot Adaptive speed control Deep reinforcement learning (DRL) DDPG-based robot control Simulation-to-real transfer CycleGAN-based domain adaptation Explainable artificial intelligence (XAI) Grad-CAM visualization Facility monitoring automation |
| 発行日: | 2026-04-14 |
| 出版者: | PeerJ |
| 誌名: | PeerJ Computer Science |
| 巻: | 12 |
| 開始ページ: | e3722 |
| DOI: | 10.7717/peerj-cs.3722 |
| 抄録: | Intelligent facility management systems can reduce the workload of human operators by enabling autonomous operation. However, the lack of transparency in existing machine learning-based systems often hinders user trust, especially in safety-critical environments such as industrial and public facilities. To ensure reliability and accountability, autonomous systems must not only perform effectively but also provide human-understandable explanations for their actions. This article presents an explainable deep reinforcement learning framework for a rail-guided patrol robot that adaptively controls its speed based on the visual complexity of its surroundings. The proposed system employs the Deep Deterministic Policy Gradient (DDPG) algorithm to learn a continuous speed-control policy directly from image-based observations. To enhance transparency, Gradient-weighted Class Activation Mapping (Grad-CAM) is integrated into the actor network to visualize which spatial regions of the input most strongly influence speed decisions, providing post hoc explanations of the model’s decisions. To support real-world deployment, we incorporate a Cycle-Consistent Generative Adversarial Network (CycleGAN)-based domain adaptation module that transforms real camera images into a simulation-compatible visual style, enabling the trained policy to operate without additional retraining. Grad-CAM is also used to assess the semantic consistency of translated images and verify that domain adaptation preserves task-relevant visual cues. Because the proposed framework is designed around lightweight visual inputs and compact neural networks, its computational demand remains modest and suitable for embedded execution. Grad-CAM analysis is used for explainability rather than for action generation, and its computation does not affect the timing of the control loop. The framework is evaluated through extensive experiments in both simulation and a physical testbed environment. Results demonstrate that the robot successfully adjusts its patrol speed in response to scene complexity and that the learned policy provides coherent and meaningful visual explanations. These findings highlight the potential of combining deep reinforcement learning, visual domain adaptation, and explainable AI to realize trustworthy and adaptable autonomous patrol systems. |
| Rights: | Copyright (c) 2026 Authors. Hosun Lee, Jaesung Kwon, Nak Young Chong and Woosung Yang. PeerJ Computer Science 12:e3722. This is an Open Access article distributed under the terms of Creative Commons Licence CC-BY [https://creativecommons.org/licenses/by/4.0/]. Original publication is available on PeerJ via https://doi.org/10.7717/peerj-cs.3722. |
| URI: | https://hdl.handle.net/10119/20347 |
| 資料タイプ: | publisher |
| 出現コレクション: | b10-1. 雑誌掲載論文 (Journal Articles)
|
このアイテムのファイル:
| ファイル |
記述 |
サイズ | 形式 |
| I-CHONG-N-0414.pdf | | 6761Kb | Adobe PDF | 見る/開く |
|
当システムに保管されているアイテムはすべて著作権により保護されています。
|