Meta-Learning to Improve Pre-Training

6. Prosinec 2021

Řečníci

O prezentaci

Pre-training (PT) followed by fine-tuning (FT) is an effective method for training neural network models, and has led to significant performance improvements in various domains. PT can incorporate task and data reweighting strategies, augmentation policies, and noise models, all of which can significantly impact the quality of representations learned and so the hyperparameters controlling such components must be set appropriately. However, setting the values of these hyperparameters is challenging—existing methods to automatically learn hyperparameters struggle to scale to high-dimensions or to adapt to the two-stage PT FT process. In this work, we propose an efficient, gradient-based algorithm to meta-learn PT hyperparameters. We formalize the PT hyperparameter optimization problem and propose a novel solution combining both direct and implicit differentiation. In experiments, we demonstrate that our method improves predictive performance on two real-world domains. First, we optimize high-dimensional task weighting hyperparameters for multitask pre-training on protein-protein interaction graphs and improve AUROC by up to 3.9

Organizátor

O organizátorovi (NeurIPS 2021)

Neural Information Processing Systems (NeurIPS) is a multi-track machine learning and computational neuroscience conference that includes invited talks, demonstrations, symposia and oral and poster presentations of refereed papers. Following the conference, there are workshops which provide a less formal setting.

Uložení prezentace

Měla by být tato prezentace uložena po dobu 1000 let?

Jak ukládáme prezentace

Pro uložení prezentace do věčného trezoru hlasovalo 0 diváků, což je 0.0 %

Sdílení

Doporučená videa

Prezentace na podobné téma, kategorii nebo přednášejícího

Zajímají Vás podobná videa? Sledujte NeurIPS 2021