metafun meta learning with iterative functional updates
play

MetaFun: Meta-Learning with Iterative Functional Updates Jin Xu, - PowerPoint PPT Presentation

MetaFun: Meta-Learning with Iterative Functional Updates Jin Xu, Jean-Francois Ton, Hyunjik Kim, Adam R. Kosiorek, Yee Whye Teh 37th International Conference on Machine Learning Supervised Meta-Learning Supervised Meta-Learning Supervised


  1. MetaFun: Meta-Learning with Iterative Functional Updates Jin Xu, Jean-Francois Ton, Hyunjik Kim, Adam R. Kosiorek, Yee Whye Teh 37th International Conference on Machine Learning

  2. Supervised Meta-Learning

  3. Supervised Meta-Learning

  4. Supervised Meta-Learning

  5. Encoder-Decoder Approaches to Supervised Meta-Learning What is learning ?

  6. Encoder-Decoder Approaches to Supervised Meta-Learning What is learning ?

  7. Encoder-Decoder Approaches to Supervised Meta-Learning What is learning ?

  8. Encoder-Decoder Approaches to Supervised Meta-Learning What is learning ? What is meta-learning? (in encoder-decoder approaches like CNP [1] )

  9. Encoder-Decoder Approaches to Supervised Meta-Learning What is learning ? What is meta-learning? (in encoder-decoder approaches like CNP [1] )

  10. Encoder-Decoder Approaches to Supervised Meta-Learning What is learning ? What is meta-learning? (in encoder-decoder approaches like CNP [1] ) “A model of learning”

  11. Encoder-Decoder Approaches to Supervised Meta-Learning Summarises Predicts conditioned the context on the representation.

  12. Encoder-Decoder Approaches to Supervised Meta-Learning Summarises Predicts conditioned the context on the representation. Both parameterised by NNs

  13. Incorporating Inductive Biases into Deep Learning Models Convolutional structure as inductive bias. Classifier Dog

  14. Incorporating Inductive Biases into Deep Learning Models Convolutional structure as inductive bias. Classifier Dog What are good inductive biases for “a model of learning”?

  15. MetaFun Overview What is a better form of set representation?

  16. MetaFun Overview What are good inductive biases/structures for the encoder? What is a better form of set representation?

  17. MetaFun Overview Euclidean Space

  18. MetaFun Overview Functional Representation Euclidean Space Function Space (e.g. Hilbert Space)

  19. MetaFun Overview Functional Representation Euclidean Space Function Space Encoders with Iterative Structure (e.g. Hilbert Space)

  20. MetaFun Overview Functional Representation Euclidean Space (permutation of data points should not change set representation) Encoders with Iterative Structure

  21. MetaFun Overview Functional Representation Euclidean Space (permutation of data points should not change set representation) [1][2][7] Encoders with Iterative Structure

  22. MetaFun Overview Functional Representation Euclidean Space (permutation of data points should not change set representation) [1][2][7] Encoders with Iterative Structure Fixed dimensional representation can be limiting for large set size [4] , and often lead to underfitting [3] .

  23. MetaFun Overview Functional Representation Euclidean Space Permutation invariance Flexible capacity (permutation of data points should not change set representation) [1][2][7] Encoders with Iterative Structure Fixed dimensional representation can be limiting for large set size [4] , and often lead to underfitting [3] .

  24. MetaFun Overview Functional Representation Euclidean Space Permutation invariance Flexible capacity (permutation of data points should not change set representation) [1][2][7] Encoders with Iterative Structure Fixed dimensional representation can be limiting for large set size [4] , and often lead to underfitting [3] . Self-attention modules [6] or relation network [9] can model interaction within the context, but not context-target interaction

  25. MetaFun Overview Functional Representation Euclidean Space Permutation invariance Flexible capacity (permutation of data points should not change set representation) Within-context and context-target interaction [1][2][7] Encoders with Iterative Structure Fixed dimensional representation can be limiting for large set size [4] , and often lead to underfitting [3] . Self-attention modules [6] or relation network [9] can model interaction within the context, but not context-target interaction

  26. MetaFun Overview Functional Representation Euclidean Space Permutation invariance Flexible capacity Within-context and context-target interaction Function Space Encoders with Iterative Structure (e.g. Hilbert Space)

  27. MetaFun Overview Functional Representation Euclidean Space Permutation invariance Flexible capacity Within-context and context-target interaction Function Space Encoders with Iterative Structure (e.g. Hilbert Space) Learning to update representation with feedback is easier than learning representation directly

  28. MetaFun Overview Functional Representation Euclidean Space Permutation invariance Flexible capacity Within-context and context-target interaction Function Space Encoders with Iterative Structure (e.g. Hilbert Space) Learning to update representation with feedback is easier than learning representation directly Iterative structure may be a good inductive bias for “the model of learning”. (Learning algorithms are often iterative, such as gradient descent)

  29. MetaFun

  30. MetaFun and Functional Gradient Descent Gradient Descent solve by iterative optimisation

  31. MetaFun and Functional Gradient Descent Gradient Descent Functional Gradient Descent solve solve For supervised learning problems, the objective function often has this form: by iterative optimisation by iterative optimisation

  32. MetaFun and Functional Gradient Descent Gradient Descent Functional Gradient Descent solve solve For supervised learning problems, the objective function often has this form: by iterative optimisation by iterative optimisation

  33. MetaFun and Functional Gradient Descent ?

  34. MetaFun and Functional Gradient Descent ?

  35. MetaFun and Functional Gradient Descent ?

  36. MetaFun and Functional Gradient Descent ?

  37. MetaFun and Functional Gradient Descent ? Evaluate functional representation at context:

  38. MetaFun and Functional Gradient Descent ? Evaluate functional representation at context: Local update funcion:

  39. MetaFun and Functional Gradient Descent ? Evaluate functional representation at context: Local update funcion: Functional pooling:

  40. MetaFun and Functional Gradient Descent ? Evaluate functional representation at context: Local update funcion: Functional pooling:

  41. MetaFun and Functional Gradient Descent ? Evaluate functional representation at context: Local update funcion: Functional pooling:

  42. MetaFun MetaFun Iteration Local update funcion: Functional pooling: Apply functional updates: will be the final representation after iterations

  43. MetaFun Functional Representation MetaFun Iteration Local update funcion: Permutation invariance ✔ Flexible capacity ✔ Functional pooling: Within-context and context-target interaction Apply functional updates:

  44. MetaFun Functional Representation MetaFun Iteration Local update funcion: Permutation invariance ✔ Flexible capacity ✔ Functional pooling: Within-context and context-target interaction ✔ Apply functional updates: Both the within-context interaction and the interaction between context and target are considered when updating the representation at each iteration.

  45. MetaFun MetaFun Iteration Local update funcion: Functional pooling: Apply functional updates:

  46. MetaFun for Classification MetaFun Iteration Local update funcion: Functional pooling: Apply functional updates: Deep kernels or attention modules

  47. MetaFun for Classification MetaFun Iteration Regression: MLP on concatenation of inputs Local update funcion: Classification: Functional pooling: ? Apply functional updates: Deep kernels or attention modules

  48. MetaFun for Classification ? Evaluate functional representation at context: Local update funcion: Functional pooling:

  49. MetaFun for Classification Local update funcion:

  50. MetaFun for Classification Local update funcion:

  51. MetaFun for Classification MetaFun Iteration Regression: MLP on concatenation of inputs Local update funcion: Classification: Functional pooling: With structure similar to Apply functional updates: Deep kernels or attention modules

  52. MetaFun for Classification MetaFun Iteration Regression: MLP on concatenation of inputs Local update funcion: Classification: Functional pooling: With structure similar to Apply functional updates: Incorporate label information into the network structure rather than concatenating the label to the inputs Deep kernels or attention modules

  53. MetaFun for Classification MetaFun Iteration Regression: MLP on concatenation of inputs Local update funcion: Classification: Functional pooling: With structure similar to Apply functional updates: Incorporate label information into the network structure rather than concatenating the label to the inputs Deep kernels or attention modules Naturally integrate within-class and between-class interaction

  54. MetaFun and Gradient-Based Meta-Learning Model Agnostic Meta-Learning (MAML) [8] MetaFun Iteration Local update funcion: Functional pooling: Apply functional updates:

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend