Reinforcement Learning Agents with Generalizing Behavior
##plugins.pubIds.doi.readerDisplayName##:
https://doi.org/10.32473/flairs.37.1.135591摘要
We explore the generality of Reinforcement Learning (RL) agents on unseen environment configurations by analyzing
the behavior of an agent tasked with traversing a graph based environment from a starting position to a goal position. We find that training on a single task is likely to result in inflexible policies that do not respond well to change. Instead, training on a wide variety of scenarios offers the best chance of developing a flexible policy, at the expense of increased training difficulty.
##submission.downloads##
已出版
##submission.howToCite##
##submission.license##
##submission.copyrightStatement##
##submission.license.cc.by-nc4.footer##