Reinforcement learning for angle-only intercept guidance of maneuvering targets
Affiliation
Univ Arizona, Dept Syst & Ind EngnIssue Date
2020-04Keywords
Reinforcement learningReinforcement meta-learning
Exo-atmospheric Intercept
Missile terminal guidance
Passive seeker
Metadata
Show full item recordCitation
Gaudet, B., Furfaro, R., & Linares, R. (2020). Reinforcement learning for angle-only intercept guidance of maneuvering targets. Aerospace Science and Technology, 99, 105746.Journal
AEROSPACE SCIENCE AND TECHNOLOGYRights
© 2020 Elsevier Masson SAS. All rights reserved.Collection Information
This item from the UA Faculty Publications collection is made available by the University of Arizona with support from the University of Arizona Libraries. If you have questions, please contact us at repository@u.library.arizona.edu.Abstract
We present a novel guidance law that uses observations consisting solely of seeker line-of-sight angle measurements and their rate of change. The policy is optimized using reinforcement meta-learning and demonstrated in a simulated terminal phase of a mid-course exo-atmospheric interception. Importantly, the guidance law does not require range estimation, making it particularly suitable for passive seekers. The optimized policy maps stabilized seeker line-of-sight angles and their rate of change directly to commanded thrust for the missile's divert thrusters. Optimization with reinforcement meta-learning allows the optimized policy to adapt to target acceleration, and we demonstrate that the policy performs better than augmented zero-effort miss guidance with perfect target acceleration knowledge. The optimized policy is computationally efficient and requires minimal memory, and should be compatible with today's flight processors. (C) 2020 Elsevier Masson SAS. All rights reserved.Note
24 month embargo; published online: 30 January 2020ISSN
1270-9638Version
Final accepted manuscriptae974a485f413a2113503eed53cd6c53
10.1016/j.ast.2020.105746
