Maml and anil provably learn representations

L Collins, A Mokhtari, S Oh… - … on Machine Learning, 2022 - proceedings.mlr.press
ANIL requires only constant error. We also show that ANIL learns the ground-truth representation
with … Now we present new intuition for MAML and ANIL’s representation learning ability …

First-order ANIL provably learns representations despite overparametrization

OK Yüksel, E Boursier, N Flammarion - arXiv preprint arXiv:2303.01335, 2023 - arxiv.org
… We study FO-ANIL in a linear shared representation model introduced in Sec… MAML,
remains open for future work. Lastly, our work presents a provable shared representation learning

Provable generalization of overparameterized meta-learning trained with sgd

Y Huang, Y Liang, L Huang - Advances in Neural …, 2022 - proceedings.neurips.cc
learning is still limited. This paper studies the generalization of a widely used meta-learning
approach, Model-Agnostic MetaLearning (MAML… ] studied MAML from a representation point …

LaANIL: ANIL with Look-Ahead Meta-Optimization and Data Parallelism

V Tammisetti, K Bierzynski, G Stettinger… - Electronics, 2024 - mdpi.com
… not considerably modify the learned representations of a fully … ANIL offers numerous
advantages compared to MAML but faces challenges such as accuracy variance, slow initial learning

Few-shot Multi-Task Learning of Linear Invariant Features with Meta Subspace Pursuit

C Zhang, L Liu, X Zhang - arXiv preprint arXiv:2409.02708, 2024 - arxiv.org
… -SP), that provably learns this invariant subspace shared by … , are still capable of learning
the shared representation, or the … As shown in our numerical experiments, ANIL requires more …

Adaptation: Blessing or Curse for Higher Way Meta-Learning

A Aimen, S Sidheekh, B Ladrecha… - IEEE Transactions …, 2023 - ieeexplore.ieee.org
… We first assess the quality of the feature extractors being learned by MAML, ANIL, ProtoNet,
Maml and anil provably learn representations. In International Conference on Machine …

Global convergence of maml and theory-inspired neural architecture search for few-shot learning

H Wang, Y Wang, R Sun, B Li - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
… ) under metalearning, and RN stands for the representation power … We use MAML-kernel
and ANIL-kernel for the 5-cells and 8-… Provable guarantees for gradient-based meta-learning. In …

Sharp-maml: Sharpness-aware model-agnostic meta learning

M Abbas, Q Xiao, L Chen, PY Chen… - … on machine learning, 2022 - proceedings.mlr.press
… head with a frozen representation network from the meta-… 2020), we ask if incorporating
Sharp-MAMLlow in ANIL can … MAML-ANIL is comparable in performance to SharpMAML

The power of adaptivity in multi-task representation learning

LG Collins - 2024 - repositories.lib.utexas.edu
… for representation learning. Surprisingly, our analysis reveals that MAML and FedAvg are
indeed capable of learning … multi-task representation learning provably occurs for nonlinear …

Consistent meta-regularization for better meta-knowledge in few-shot learning

P Tian, W Li, Y Gao - … on Neural Networks and Learning …, 2021 - ieeexplore.ieee.org
… each few-shot learning task can be considered as a biased representation of the whole data.
… If we consider that the base learner is a neural network, eg, MAML or ANIL, which contains …