[ICML '17] Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks

minha·2022년 11월 17일
0

RL

목록 보기
3/6
post-thumbnail
  • 풀고자 하는 문제
    Trained 모델이 new task에 대해 fine-tuning 되는 과정을 가속화시켜줌. 즉, new task에 대해 fine-tuning 될 때 새로운 데이터가 많이 필요 없음. 다시 말해 few-shot learning의 범주에 든다고 생각하면 됨 (Few-shot & Rapid adaptation)
    논문에 쓰인 구절을 차용하자면, our artificial agents should be able to learn and adapt quickly from only a few examples, and continue to adapt as more data becomes available!
  • Main contributions
    • Train the model's initial parameters such that the model has maximal performance on a new task after the parameters have been updated through one or more gradient steps computed with a small amount of data from the new task
    • Maximize the sensitivity of the loss functions of new tasks with respect to the parameters
    • Model-agnostic & Task-agnostic
      • Supervised regression
      • Supervised classification
      • Reinforcement learning e.g., locomotion
        • Accelerate fine-tuning for policy gradient reinforcement learning with NN policies
        • A new task might involve achieving a new goal or succeeding on a previously trained goal in a new environment
  • MAML 알고리즘
  • 코드: 오피셜 있음, Tensorflow
    github.com/cbfinn/maml
    github.com/cbfinn/maml_rl

0개의 댓글