Loading [a11y]/accessibility-menu.js
Multimodal Task Attention Residual Reinforcement Learning: Advancing Robotic Assembly in Unstructured Environment | IEEE Journals & Magazine | IEEE Xplore

Multimodal Task Attention Residual Reinforcement Learning: Advancing Robotic Assembly in Unstructured Environment


Abstract:

Robotic assembly in dynamic and unstructured environments poses challenges for recent methods, due to background noise and wide-ranging errors. Directly learning from env...Show More

Abstract:

Robotic assembly in dynamic and unstructured environments poses challenges for recent methods, due to background noise and wide-ranging errors. Directly learning from environments relies on complex models and extensive training iterations to adapt. Representation selection approaches, which depend on expert knowledge, can reduce training costs but suffer from poor robustness and high manual costs, limiting scalability. In response, this letter proposes a system that integrates task attention into residual reinforcement learning to address these challenges. By effectively segmenting task-relevant information from the background to leverage task attention, our approach mitigates the impact of environmental variability. Additionally, compared with existing baselines, our task attention mechanism based on instance segmentation and prompt-guided selection does not require additional offline training or local fine-tuning. Experimental evaluations conducted in both simulated and real environments demonstrate the superiority of our method over various baselines. Specifically, our system achieves high efficiency and effectiveness in learning and executing assembly tasks in dynamic and unstructured environments.
Published in: IEEE Robotics and Automation Letters ( Volume: 10, Issue: 4, April 2025)
Page(s): 3900 - 3907
Date of Publication: 03 March 2025

ISSN Information:

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.