Open access
Date
2023Type
- Journal Article
ETH Bibliography
yes
Altmetrics
Abstract
We introduce meta-learning algorithms that perform zero-shot weight-space adaptation of neural network models to unseen tasks. Our methods repurpose the popular generative image synthesis techniques of natural language guidance and diffusion models to generate neural network weights adapted for tasks. We first train an unconditional generative hypernetwork model to produce neural network weights; then we train a second "guidance" model that, given a natural language task description, traverses the hypernetwork latent space to find high-performance task-adapted weights in a zero-shot manner. We explore two alternative approaches for latent space guidance: "HyperCLIP"-based classifier guidance and a conditional Hypernetwork Latent Diffusion Model ("HyperLDM"), which we show to benefit from the classifier-free guidance technique common in image generation. Finally, we demonstrate that our approaches outperform existing multi-task and meta-learning methods in a series of zero-shot learning experiments on our Meta-VQA dataset. Show more
Permanent link
https://doi.org/10.3929/ethz-b-000653575Publication status
publishedExternal links
Journal / series
Transactions on Machine Learning ResearchVolume
Publisher
OpenReviewSubject
Machine Learning (cs.LG); FOS: Computer and information sciencesOrganisational unit
09689 - Katzschmann, Robert / Katzschmann, Robert
Funding
173721 - Temporal Information Integration in Neural Networks (SNF)
189251 - Ultra compact miniaturized microscopes to image meso-scale brain activity (SNF)
ETH-20 19-1 - A cross‐disciplinary, data‐driven approach to predict stress resilience from large‐scale behavioral, molecular and neural activity data (ETHZ)
Related publications and datasets
Is new version of: https://doi.org/10.48550/arXiv.2210.08942
More
Show all metadata
ETH Bibliography
yes
Altmetrics