Abstract

In this paper, we combine context-based Meta-Reinforcement Learning with task-aware representation to efficiently overcome data-inefficiency and limited generalization in the hyperparameter optimization problem. First, we propose a new context-based meta-RL model that disentangles task inference and control, which improves the meta-training efficiency and accelerates the learning process for unseen tasks. Second, the task properties are inferred on-line, which includes not only the dataset representation but also the task-solving experience, thus encouraging the agent to explore in a much smarter fashion. Third, we employ amortized meta-learning to meta-train the agent, which is simple and runs faster than the gradient-based meta-training method. Experimental results suggest that our method can search for the optimal hyperparameter configuration with limited computational cost in a reasonable time.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.