TY - GEN
T1 - Generalizing to New Tasks via One-Shot Compositional Subgoals
AU - Xihan, Bian
AU - Mendez, Oscar
AU - Lianpin, Zhang
AU - Hadfield, Simon
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/2
Y1 - 2024/2
N2 - Generalizing to new tasks with little supervision is a challenge in machine learning and a requirement for future 'General AI' agents. Reinforcement and imitation learning is used to adapt to new tasks, but this is difficult for complex tasks that require long-term planning. However, this can be challenging for complex tasks often requiring many timesteps or large numbers of subtasks. This leads to long episodes with long-horizon tasks which are difficult to learn. In this work, we attempt to address these issues by training an Imitation Learning agent using in-episode 'near future' subgoals. These sub goals are re-calculated at each step using compositional arithmetic in a learned latent representation space. In addition to improving learning efficiency for standard long-term tasks, this approach also makes it possible to perform one-shot generalization to previously unseen tasks, given only a single reference trajectory for the task in a different environment. Our experiments show that the proposed approach consistently outperforms the previous state-of-the-art compositional Imitation Learning approach by 30%. While capable of learning from long episodes where the SOTA fails.
AB - Generalizing to new tasks with little supervision is a challenge in machine learning and a requirement for future 'General AI' agents. Reinforcement and imitation learning is used to adapt to new tasks, but this is difficult for complex tasks that require long-term planning. However, this can be challenging for complex tasks often requiring many timesteps or large numbers of subtasks. This leads to long episodes with long-horizon tasks which are difficult to learn. In this work, we attempt to address these issues by training an Imitation Learning agent using in-episode 'near future' subgoals. These sub goals are re-calculated at each step using compositional arithmetic in a learned latent representation space. In addition to improving learning efficiency for standard long-term tasks, this approach also makes it possible to perform one-shot generalization to previously unseen tasks, given only a single reference trajectory for the task in a different environment. Our experiments show that the proposed approach consistently outperforms the previous state-of-the-art compositional Imitation Learning approach by 30%. While capable of learning from long episodes where the SOTA fails.
KW - Compositional Model
KW - Imitation Learning
KW - Planning
KW - Reinforcement Learning
UR - http://www.scopus.com/inward/record.url?scp=85197349120&partnerID=8YFLogxK
UR - https://ieeexplore.ieee.org/abstract/document/10552980
U2 - 10.1109/ICARA60736.2024.10552980
DO - 10.1109/ICARA60736.2024.10552980
M3 - Conference Proceeding
AN - SCOPUS:85197349120
T3 - 2024 10th International Conference on Automation, Robotics, and Applications, ICARA 2024
SP - 491
EP - 495
BT - 2024 10th International Conference on Automation, Robotics, and Applications, ICARA 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 10th International Conference on Automation, Robotics, and Applications, ICARA 2024
Y2 - 22 February 2024 through 24 February 2024
ER -