TUHH Open Research
Help
  • Log In
    New user? Click here to register.Have you forgotten your password?
  • English
  • Deutsch
  • Communities & Collections
  • Publications
  • Research Data
  • People
  • Institutions
  • Projects
  • Statistics
  1. Home
  2. TUHH
  3. Publication References
  4. Residual reinforcement learning for robot control
 
Options

Residual reinforcement learning for robot control

Publikationstyp
Conference Paper
Date Issued
2019-05
Sprache
English
Author(s)
Johannink, Tobias  
Bahl, Shikhar  
Nair, Ashvin  
Luo, Jianlan  
Kumar, Avinash  
Loskyll, Matthias  
Ojea, Juan Aparicio  
Solowjow, Eugen  
Levine, Sergey  
Institut
Mechanik und Meerestechnik M-13  
TORE-URI
http://hdl.handle.net/11420/10865
Volume
2019
Start Page
6023
End Page
6029
Article Number
8794127
Citation
IEEE International Conference on Robotics and Automation (ICRA 2019): 8794127, 60023-6029 (2019)
Contribution to Conference
IEEE International Conference on Robotics and Automation, ICRA 2019  
Publisher DOI
10.1109/ICRA.2019.8794127
Scopus ID
2-s2.0-85071448649
Publisher
IEEE
Conventional feedback control methods can solve various types of robot control problems very efficiently by capturing the structure with explicit models, such as rigid body equations of motion. However, many control problems in modern manufacturing deal with contacts and friction, which are difficult to capture with first-order physical modeling. Hence, applying control design methodologies to these kinds of problems often results in brittle and inaccurate controllers, which have to be manually tuned for deployment. Reinforcement learning (RL) methods have been demonstrated to be capable of learning continuous robot controllers from interactions with the environment, even for problems that include friction and contacts. In this paper, we study how we can solve difficult control problems in the real world by decomposing them into a part that is solved efficiently by conventional feedback control methods, and the residual which is solved with RL. The final control policy is a superposition of both control signals. We demonstrate our approach by training an agent to successfully perform a real-world block assembly task involving contacts and unstable objects.
TUHH
Weiterführende Links
  • Contact
  • Send Feedback
  • Cookie settings
  • Privacy policy
  • Impress
DSpace Software

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science
Design by effective webwork GmbH

  • Deutsche NationalbibliothekDeutsche Nationalbibliothek
  • ORCiD Member OrganizationORCiD Member Organization
  • DataCiteDataCite
  • Re3DataRe3Data
  • OpenDOAROpenDOAR
  • OpenAireOpenAire
  • BASE Bielefeld Academic Search EngineBASE Bielefeld Academic Search Engine
Feedback