@inproceedings{8e304706a8b340f9b503cf0925eb37e4,
title = "Conditional mutual information-based generalization bound for meta learning",
abstract = "Meta-learning optimizes an inductive bias—typically in the form of the hyperparameters of a base-learning algorithm—by observing data from a finite number of related tasks. This paper presents an information-theoretic bound on the generalization performance of any given meta-learner, which builds on the conditional mutual information (CMI) framework of Steinke and Zakynthinou (2020). In the proposed extension to meta-learning, the CMI bound involves a training meta-supersample obtained by first sampling 2N independent tasks from the task environment, and then drawing 2M independent training samples for each sampled task. The meta-training data fed to the meta-learner is modelled as being obtained by randomly selecting N tasks from the available 2N tasks and M training samples per task from the available 2M training samples per task. The resulting bound is explicit in two CMI terms, which measure the information that the meta-learner output and the base-learner output provide about which training data are selected, given the entire meta-supersample. Finally, we present a numerical example that illustrates the merits of the proposed bound in comparison to prior information-theoretic bounds for meta-learning.",
keywords = "Training, Deep learning, Adaptation models, Training data, Data models, Task analysis, Mutual information",
author = "Arezou Rezazadeh and Sharu Jose and Giuseppe Durisi and Osvaldo Simeone",
year = "2021",
month = sep,
day = "1",
doi = "10.1109/ISIT45174.2021.9518020",
language = "English",
series = "IEEE International Symposium on Information Theory proceedings",
publisher = "IEEE",
pages = "1176--1181",
booktitle = "2021 IEEE International Symposium on Information Theory (ISIT)",
}