Model-based meta-learning in neural networks

Aluno: Willian Wang

Orientadora: Nina S. T. Hirata

Meta-learning, often referred to as “learning to learn,” aims to improve the performance and efficiency of learning algorithms by enabling them to learn more effectively new tasks. This approach is particularly relevant in the context of deep learning, where classic challenges such as data and computational bottlenecks often limit model performance and training speed.[1]

Unlike optimization-based methods, which focus on improving gradient backpropagation, and metric-based methods, which seek to find an optimal metric function for a specific problem, model-based (or “black box”) meta-learning relies primarily on its architecture. This approach leverages the design of the network to update its weights more efficiently.

The goal of this study is to focus on model-based methods of meta-learning within neural networks. It will examine a range of historically significant architectures in deep learning to establish a foundation, explore recent advancements in meta-learning using these techniques, and implement a few pertinent algorithms to understand them better. The plan is:

  1. Study of Memory-Augmented Neural Networks (MANNs): These techniques are particularly important to this work because there are multiple attempts of using memory storage enhanced networks to improve the learning process.

  2. Analysis of Meta-Learning Methods Relative to Conventional Techniques: This involves a detailed examination of meta-learning models influenced by established frameworks, including MANNs and other models. [2][3]

  3. Investigation of Non-Traditional Meta-Learning Techniques: This includes the study of strategies that diverge from mainstream algorithms, such as hypernetworks and self-referential networks [4][5]

  4. Selective Implementation of Algorithms: Given the impracticality of implementing every algorithm, a few selected algorithms will be implemented to understand this field better.

Activity April May June July August September October November December
1 X X X
2 X X X X
3 X X X X
4 X X X X
report writing X X X

  1. T. Hospedales, A. Antoniou, P. Micaelli and A. Storkey, “Meta-Learning in Neural Networks: A Survey,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 9, pp. 5149-5169, 1 Sept. 2022, doi: 10.1109/TPAMI.2021.3079209.

  2. Adam Santoro, Sergey Bartunov, Matthew Botvinick, Daan Wierstra, and Timothy Lillicrap. 2016. Meta-learning with memory-augmented neural networks. In Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48 (ICML’16). JMLR.org, 1842–1850.

  3. Sergey Bartunov, Jack Rae, Simon Osindero, and Timothy Lillicrap. 2020. Meta-Learning Deep Energy-Based Memory Models. In International Conference on Learning Representations.

  4. Louis Kirsch and Jürgen Schmidhuber. 2021. Meta Learning Backpropagation And Improving It. In Advances in Neural Information Processing Systems, Volume 34, pages 14122–14134. Curran Associates, Inc.

  5. Mark Sandler, Max Vladymyrov, Andrey Zhmoginov, Nolan Andrew Miller, Andrew Jackson, Tom Madams, and Blaise Hilary Aguera-Arcas. 2021. Meta-Learning Bidirectional Update Rules. In ICML 2021.