The list of questions from the panel discussion are here.
Recent years have seen rapid progress in meta-learning methods, which learn (and optimize) the performance of learning methods based on data, generate new learning methods from scratch, or learn to transfer knowledge across tasks and domains. Meta-learning can be seen as the logical conclusion of the arc that machine learning has undergone in the last decade, from learning classifiers, to learning representations, and finally to learning algorithms that themselves acquire representations and classifiers. The ability to improve one’s own learning capabilities through experience can also be viewed as a hallmark of intelligent beings, and there are strong connections with work on human learning in neuroscience.
Meta-learning methods are also of substantial practical interest, since they have, e.g., been shown to yield new state-of-the-art automated machine learning methods, novel deep learning architectures, and substantially improved one-shot learning systems.
Some of the fundamental questions that this workshop aims to address are:
- What are the fundamental differences in the learning “task” compared to traditional “non-meta” learners?
- Is there a practical limit to the number of meta-learning layers (e.g., would a meta-meta-meta-learning algorithm be of practical use)?
- How can we design more sample-efficient meta-learning methods?
- How can we exploit our domain knowledge to effectively guide the meta-learning process?
- What are the meta-learning processes in nature (e.g, in humans), and how can we take inspiration from them?
- Which ML approaches are best suited for meta-learning, in which circumstances, and why?
- What principles can we learn from meta-learning to help us design the next generation of learning systems?
The goal of this workshop is to bring together researchers from all the different communities and topics that fall under the umbrella of meta-learning. We expect that the presence of these different communities will result in a fruitful exchange of ideas and stimulate an open discussion about the current challenges in meta-learning, as well as possible solutions.
Invited Speakers
- Nando de Freitas (DeepMind)
- Lise Getoor (UC Santa Cruz)
- Hugo Larochelle (Google Brain)
- Sergey Levine (UC Berkeley)
- Michèle Sebag (Paris-Saclay)
Invited Panelists
- Roberto Calandra (Facebook AI Research)
- Nando de Freitas (DeepMind)
- Hugo Larochelle (Google Brain)
- Jürgen Schmidhuber (IDSIA)
- Michèle Sebag (Paris-Saclay)
Organizers
- Frank Hutter (University of Freiburg)
- Joaquin Vanschoren (Eindhoven University of Technology)
- Erin Grant (UC Berkeley)
- Jane Wang (DeepMind)
- Sachin Ravi (Princeton)
Important dates
Submission deadline: 17 October 2018 (11:59 PM anywhere on Earth)Notification: 09 November 2018Camera ready: 03 December 2018- Workshop: 08 December 2018
Schedule
09:00 | Introduction and opening remarks |
09:10 | Invited talk 1: Lise Getoor, “Exploiting structure for meta-learning” |
09:40 | Poster spotlights 1 |
10:00 | Poster session 1 |
10:30 | Coffee break |
11:00 | Invited talk 2: Sergey Levine, “What’s wrong with meta-learning (and how we can fix it)” |
11:30 | Poster session 2 |
12:00 | Lunch break |
13:30 | Invited talk 3: Hugo Larochelle, “Thoughts on progress made and challenges ahead in few-shot learning” |
14:00 | Invited talk 4: Michèle Sebag, “Monte Carlo tree search for algorithm configuration: MOSAIC” |
14:30 | Poster spotlights 2 |
14:50 | Poster session 3 |
15:00 | Coffee break |
15:30 | Poster session 4 |
16:00 | Invited talk 5: Nando de Freitas, “Tools that learn” |
16:30 | Contributed talk 1: JD Co-Reyes, “Guiding policies with language via meta-learning” |
16:45 | Contributed talk 2: Arthur Pesah & Antoine Wehenkel, “Recurrent machines for likelihood-free inference” |
17:00 | Panel discussion |
18:00 | End |
Invited Talks
Lise Getoor (UC Santa Cruz), “Exploiting structure for meta-learning”
Many machine learning problems exhibit rich structural dependencies. We need meta-learning algorithms which can represent, discover and exploit them, and we can use structured models to express the dependencies inherent in meta-learning. In this talk, I’ll introduce some common structural dependencies, show their power and how they can be represented, and discuss how we can make use of them for meta-learning.
Sergey Levine (UC Berkeley), “What’s wrong with meta-learning (and how we can fix it)”
Meta-learning, or learning to learn, offers an appealing framework for training deep neural networks to adapt quickly and efficiently to new tasks. Indeed, the framework of meta-learning holds the promise of resolving the long-standing challenge of sample complexity in deep learning: by learning to learn efficiently, deep models can be meta-trained to adapt quickly to classify new image classes from a couple of examples, or learn new skills with reinforcement learning from just a few trials.
However, although the framework of meta-learning and few-shot learning is exceedingly appealing, it carries with it a number of major challenges. First, designing neural network models for meta-learning is quite difficult, since meta-learning models must be able to ingest entire datasets to adapt effectively. I will discuss how this challenge can be addressed by describing a model-agnostic meta-learning algorithm: a meta-learning algorithm that can use any model architecture, training that architecture to adapt efficiently via simple finetuning.
The second challenge is that meta-learning trades off the challenge of algorithm design (by learning the algorithm) for the challenge of task design: the performance of meta-learning algorithms depends critically on the ability of the user to manually design large sets of diverse meta-training tasks. In practice, this often ends up being an enormous barrier to widespread adoption of meta-learning methods. I will describe our recent work on unsupervised meta-learning, where tasks are proposed automatically from unlabeled data, and discuss how unsupervised meta-learning can exceed the performance of standard unsupervised learning methods while removing the manual task design requirement inherent in standard meta-learning methods.
Hugo Larochelle (Google Brain), “Thoughts on progress made and challenges ahead in few-shot learning”
A lot of the recent progress on many AI tasks were enabled in part by the availability of large quantities of labeled data. Yet, humans are able to learn concepts from as little as a handful of examples. Meta-learning has been a very promising framework for addressing the problem of generalizing from small amounts of data, known as few-shot learning. In this talk, I’ll present an overview of the recent research that has made exciting progress on this topic. I will also share my thoughts on the challenges and research opportunities that remain in few-shot learning, including a proposal for a new benchmark.
Michèle Sebag (Paris-Saclay), “Monte Carlo tree search for algorithm configuration: MOSAIC”
The sensitivity of algorithms (related to machine learning, combinatorial optimization, constraint satisfaction) w.r.t. their hyperparameters, and the difficulty of finding the algorithm and its hyperparameter setting best suited to the problem instance at hand, has led to the rapidly developing field of algorithm selection and calibration, and, focusing on machine learning, to AutoML.
Several international AutoML challenges have been organized since 2013, motivating the development of the Bayesian optimization-based approach Auto-sklearn, the randomized search approach Hyperband, and others. This talk will present a new approach, called Monte Carlo Tree Search for Algorithm Configuration (MOSAIC), fully exploiting the tree structure of the algorithm portfolio-hyperparameter search space.
It is shown that MOSAIC outperforms the current AutoML winner Auto-Sklearn on both the AutoML challenge 2015, and the MNIST dataset.
Joint work: Heri Rakotoarison
Nando de Freitas (DeepMind), “Tools that learn”
Spotlights 1 (and Poster Sessions 1 & 2)
- Meta-Learner with Linear Nulling
- OBOE: Collaborative Filtering for AutoML Initialization
- Backpropamine: Meta-Training Self-Modifying Neural Networks with Gradient Descent
- Hyperparameter Learning via Distributional Transfer
- Toward Multimodal Model-Agnostic Meta-Learning
- Fast Neural Architecture Construction Using EnvelopeNets
- Meta-Dataset: A Dataset of Datasets for Learning to Learn from Few Examples
- Macro Neural Architecture Search Revisited
- AutoDL Challenge Design and Beta Tests
- Modular Meta-Learning in Abstract Graph Networks for Combinatorial Generalization
- Cross-Modulation Networks for Few-Shot Learning
- Large Margin Meta-Learning for Few-Shot Classification
- Amortized Bayesian Meta-Learning
- The Effects of Negative Adaptation in Model-Agnostic Meta-Learning
- Mitigating Architectural Mismatch During the Evolutionary Synthesis of Deep Neural Networks
- Evolvability ES: Scalable Evolutionary Meta-Learning
- Consolidating the Meta-Learning Zoo: A Unifying Perspective as Posterior Predictive Inference
- Deep Online Learning via Meta-Learning: Continual Adaptation for Model-Based RL
Spotlights 2 (and Poster Sessions 3 & 4)
- Incremental Few-Shot Learning with Attention Attractor Networks
- Auto-Meta: Automated Gradient-Based Meta-Learner Search
- Transferring Knowledge Across Learning Processes
- Few-Shot Learning for Free by Modeling Global Class Structure
- TAEML: Task-Adaptive Ensemble of Meta-Learners
- A Simple Transfer-Learning Extension of Hyperband
- Learned Optimizers That Outperform SGD on Wall-Clock and Validation Loss
- Learning to Learn with Conditional Class Dependencies
- Unsupervised Learning via Meta-Learning
- Control Adaptation via Meta-Learning Dynamics
- Learning to Adapt in Dynamic, Real-World Environments via Meta-Reinforcement Learning
- Learning to Design RNA
- Graph Hypernetworks for Neural Architecture Search
- Meta-Learning with Latent Embedding Optimization
- ProMP: Proximal Meta-Policy Search
- Attentive Task-Agnostic Meta-Learning for Few-Shot Text Classification
- Variadic Learning by Bayesian Nonparametric Deep Embedding
- from Nodes to Networks: Evolving Recurrent Neural Networks
- Meta Learning for Defaults: Symbolic Defaults
Submission Instructions
The submission window for this workshop is now closed. Decision notifications were sent out November 9, 2018. Thank you to all who submitted!
We have provided a modified .sty
file here that appropriately lists the name of the workshop when \neuripsfinal
is enabled. Please use this style files in conjunction with corresponding LaTeX .tex
template from the NeurIPS website to submit a final camera-ready copy.
Accepted papers and supplementary material are available on the workshop website. However, these do not constitute archival publications and no formal workshop proceedings will be made available, meaning contributors are free to publish their work in archival journals or conferences.
FAQ
-
Can supplementary material be added beyond the 4-page limit and are there any restrictions on it?
Yes, you may include additional supplementary material, but we ask that it be limited to a reasonable amount (max 10 pages in addition to the main submission) and that it follow the same NeurIPS format as the paper.
-
Can a submission to this workshop be submitted to another NeurIPS workshop in parallel?
We discourage this, as it leads to more work for reviewers across multiple workshops. Our suggestion is to pick one workshop to submit to.
-
If a submission is accepted, is it possible for all authors of the accepted paper to receive a chance to register?
We cannot confirm this yet, but it is most likely that we will have at most one registration to offer per accepted paper.
-
Can a paper be submitted to the workshop that has already appeared at a previous conference with published proceedings?
We won’t be accepting such submissions unless they have been adapted to contain significantly new results (where novelty is one of the qualities reviewers will be asked to evaluate).
Accepted Abstracts
- A Simple Transfer-Learning Extension of Hyperband [supp] Lazar Valkov, Rodolphe Jenatton, Fela Winkelmolen, Cedric Archambeau
- Amortized Bayesian Meta-Learning Sachin Ravi, Alex Beatson
- Attentive Task-Agnostic Meta-Learning for Few-Shot Text Classification Xiang Jiang, Mohammad Havaei, Gabriel Chartrand, Hassan Chouaib, Thomas Vincent, Andrew D. Jesson, Nicolas Chapados, Stan Matwin
- Auto-Meta: Automated Gradient-Based Meta-Learner Search Sangyeul Lee, Jaehong Kim, Sungwan Kim, Moonsu Cha, Jung Kwon Lee, Yongseok Choi, Dong-Yeon Cho, Jiwon Kim, Youngduck Choi
- AutoDL Challenge Design and Beta Tests Zhengying Liu, Isabelle Guyon, Olivier Bousquet, Andre Elisseeff, Sergio Escalera, Sebastien Treger, Julio C.S. Jacques Junior, Danny Silver, Adrien Pavao, Wei Wei Tu, Lisheng Sun, Jingsong Wang, Quanming Yao
- Backpropamine: Meta-Training Self-Modifying Neural Networks with Gradient Descent Thomas Miconi, Kenneth O. Stanley, Jeff Clune
- Consolidating the Meta-Learning Zoo: A Unifying Perspective as Posterior Predictive Inference Jonathan Gordon, John Bronskill, Matthias Bauer, Richard Turner, Sebastian Nowozin
- Control Adaptation via Meta-Learning Dynamics James Harrison, Apoorva Sharma, Roberto Calandra, Marco Pavone
- Cross-Modulation Networks for Few-Shot Learning Hugo Prol Pereira, Vincent Dumoulin, Luis Herranz
- Deep Online Learning via Meta-Learning: Continual Adaptation for Model-Based RL Anusha Nagabandi, Sergey Levine, Chelsea Finn
- Evolvability ES: Scalable Evolutionary Meta-Learning Alexander Gajewski, Jeff Clune, Kenneth O. Stanley, Joel Lehman
- Fast Neural Architecture Construction Using EnvelopeNets Purushotham Kamath, Abhishek Singh, Debo Dutta
- Few-Shot Learning for Free by Modeling Global Class Structure Will S. Grathwohl, Xuechen Li, Eleni Triantafillou, Richard Zemel, David Duvenaud
- From Nodes to Networks: Evolving Recurrent Neural Networks [arXiv] Aditya Rawal, Risto Miikkulainen
- Gradient Agreement as an Optimization Objective for Meta-Learning [arXiv] Amir Erfan Eshratifar, David Eigen, Massoud Pedram
- Graph HyperNetworks for Neural Architecture Search Chris Zhang, Mengye Ren, Raquel Urtasun
- Hyperparameter Learning via Distributional Transfer Ho Chung Leon Law, Peilin Zhao, Junzhou Huang, Dino Sejdinovic
- Incremental Few-Shot Learning with Attention Attractor Networks Mengye Ren, Renjie Liao, Ethan Fetaya, Richard Zemel
- Large Margin Meta-Learning for Few-Shot Classification Yong Wang, Xiao-Ming Wu, Qimai Li, Jiatao Gu, Wangmeng Xiang, Lei Zhang, Victor O.K. Li
- Learned Optimizers That Outperform SGD on Wall-Clock and Validation Loss Luke Metz, Niru Maheswaranathan, Jeremy Nixon, Daniel Freeman, Jascha Sohl-Dickstein
- Learning to Adapt in Dynamic, Real-World Environments via Meta-Reinforcement Learning Anusha Nagabandi, Ignasi Clavera, Sergey Levine, Ronald Fearing, Chelsea Finn, Simin Liu, Pieter Abbeel
- Learning to Design RNA Frederic Runge, Danny Stoll, Stefan Falkner, Frank Hutter
- Learning to Learn with Conditional Class Dependencies Xiang Jiang, Mohammad Havaei, Farshid Varno, Gabriel Chartrand, Nicolas Chapados, Stan Matwin
- Macro Neural Architecture Search Revisited [supp] Hanzhang Hu, John Langford, Rich Caruana, Eric Horvitz, Martial Hebert, J. Andrew Bagnell, Debadeepta Dey
- Meta Learning for Defaults: Symbolic Defaults Jan N. Van Rijn, Florian Pfisterer, Janek Thomas, Bernd Bischl, Andreas Mueller, Joaquin Vanschoren
- Meta-Dataset: A Dataset of Datasets for Learning to Learn from Few Examples [supp] Eleni Triantafillou, Tyler Zhu, Vincent Dumoulin, Pascal Lamblin, Kelvin Xu, Ross Goroshin, Carles Gelada, Kevin Swersky, Pierre-Antoine Manzagol, Hugo Larochelle
- Meta-Learner with Linear Nulling [supp] Sung Whan Yoon, Jun Seo, Jaekyun Moon
- Guiding Policies with Language via Meta-Learning John Co-Reyes, Abhishek Gupta, Suvansh Sanjeev, Nick Altieri, John DeNero, Pieter Abbeel, Sergey Levine
- Meta-Learning with Latent Embedding Optimization Andrei A. Rusu, Dushyant Rao, Jakub Sygnowski, Oriol Vinyals, Razvan Pascanu, Simon Osindero, Raia Hadsell
- Mitigating Architectural Mismatch During the Evolutionary Synthesis of Deep Neural Networks Audrey Chung, Paul Fieguth, Alexander Wong
- Modular Meta-Learning in Abstract Graph Networks for Combinatorial Generalization Ferran Alet, Maria Bauza Villalonga, Alberto Rodriguez, Tomas Lozano-Perez, Leslie Kaelbling
- OBOE: Collaborative Filtering for AutoML Initialization Chengrun Yang, Yuji Akimoto, Dae Won Kim, Madeleine Udell
- ProMP: Proximal Meta-Policy Search Ignasi Clavera, Jonas Rothfuss, Dennis Lee, Tamim Asfour, Pieter Abbeel
- Recurrent Machines for Likelihood-Free Inference Arthur Pesah, Antoine Wehenkel, Gilles Louppe
- TAEML: Task-Adaptive Ensemble of Meta-Learners Minseop Park, Jungtaek Kim, Saehoon Kim, Yanbin Liu, Seungjin Choi
- The Effects of Negative Adaptation in Model-Agnostic Meta-Learning [arXiv] Tristan Deleu, Yoshua Bengio
- Toward Multimodal Model-Agnostic Meta-Learning [supp] Risto Vuorio, Shao-Hua Sun, Hexiang Hu, Joseph Lim*
- Transferring Knowledge Across Learning Processes [supp] Sebastian Flennerhag, Andreas Damianou, Pablo Moreno, Neil Lawrence
- Unsupervised Learning via Meta-Learning [arXiv] Kyle Hsu, Sergey Levine, Chelsea Finn
- Variadic Learning by Bayesian Nonparametric Deep Embedding Kelsey Allen, Hanul Shin, Evan Shelhamer, Joshua Tenenbaum
Program Committee
We thank the program committee for shaping the excellent technical program (in alphabetical order):
Aaron Klein, Abhishek Gupta, Alexandre Lacoste, Andre Carvalho, Andrew Brock, Anusha Nagabandi, Aravind Srinivas, Balazs Kegl, Benjamin Letham, Brandon Schoenfeld, Chelsea Finn, Daniel Hernandez-Lobato, Dumitru Erhan, Eleni Triantafillou, Eytan Bakshy, Ghassen Jerfel, Hugo Larochelle, Hugo Jair Escalante, Ignasi Clavera, Igor Mordatch, Jake Snell, Jan van Rijn, Jasper Snoek, Jürgen Schmidhuber, Ke Li, Lars Kotthoff, Marius Lindauer, Matt Hoffman, Mengye Ren, Michael Chang, Misha Denil, Parminder Bhatia, Pavel Brazdil, Pieter Gijsbers, Rafael Mantovani, Razvan Pascanu, Ricardo Prudencio, Roberto Calandra, Rodolphe Jenatton, Roger Grosse, Roman Garnett, Sayna Ebrahimi, Sergio Escalera, Stephen Roberts, Thanard Kurutach, Thomas Elsken, Tin Ho, Udayan Khurana
Past workshops
Workshop on Meta-Learning (MetaLearn 2017) @ NeurIPS 2017
Contacts
For any further questions, you can contact us at info@metalearning.ml.
Sponsors
We are very thankful to our corporate sponsors!