by , , ,
Meta-Learning aims to accelerate the learning on new tasks by acquiring useful inductive biases from related data sources. In practice, the number of tasks available for meta-learning is often small. Yet, most of the existing approaches rely on an abundance of meta-training tasks, making them prone to overfitting. How to regularize the meta-learner to ensure generalization to unseen tasks, is a central question in the literature. We provide a theoretical analysis using the PAC-Bayesian framework and derive the first bound for meta-learners with unbounded loss functions. Crucially, our bounds allow us to derive the PAC-optimal hyper-posterior (PACOH) - the closed-form-solution of the PAC-Bayesian meta-learning problem, thereby avoiding the reliance on nested optimization, giving rise to an optimization problem amenable to standard variational methods that scale well. Our experiments show that, when instantiating the PACOH with Gaussian processes and Bayesian Neural Networks as base learners, the resulting methods are more scalable, and yield state-of-the-art performance, both in terms of predictive accuracy and the quality of uncertainty estimates. Finally, thanks to the principled treatment of uncertainty, our meta-learners can also be successfully employed for sequential decision problems.
PAC-Bayesian Meta-Learning: From Theory to Practice J. Rothfuss, M. Josifoski, V. Fortuin, A. KrauseIn arXiv preprint arXiv:2211.07206, 2022
Bibtex Entry:
	author = {Rothfuss, Jonas and Josifoski, Martin and Fortuin, Vincent and Krause, Andreas},
	journal = {arXiv preprint arXiv:2211.07206},
	month = {December},
	title = {PAC-Bayesian Meta-Learning: From Theory to Practice},
	year = {2022}}