deep multi task learning with evolving weights
play

Deep multi-task learning with evolving weights Machine learning - - PowerPoint PPT Presentation

Deep multi-task learning with evolving weights Machine learning - computer vision published in European Symposium on Artificial Neural Networks (ESANN 2016) Soufiane Belharbi Romain Hrault Clment Chatelain Sbastien Adam


  1. Deep multi-task learning with evolving weights Machine learning - computer vision published in European Symposium on Artificial Neural Networks (ESANN 2016) Soufiane Belharbi Romain Hérault Clément Chatelain Sébastien Adam soufiane.belharbi@insa-rouen.fr LITIS lab., Apprentissage team - INSA de Rouen, France images/logos JDD, Le Havre. 14 June, 2016 LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights

  2. Introduction Machine learning What is machine learning (ML)? ML is programming computers (algorithms) to optimize a performance criterion using example data or past experience . Learning a task Learn general models from data to perform a specific task f . f w : x − → y x : input y : output (target, label) w : parameters of f f ( x ; w ) = y From training to predicting the future: Learn to predict Train the model using data examples ( x , y ) 1 Predict the y new for the new coming x new images/logos 2 LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 1/29

  3. Introduction Machine learning What is machine learning (ML)? ML is programming computers (algorithms) to optimize a performance criterion using example data or past experience . Learning a task Learn general models from data to perform a specific task f . f w : x − → y x : input y : output (target, label) w : parameters of f f ( x ; w ) = y From training to predicting the future: Learn to predict Train the model using data examples ( x , y ) 1 Predict the y new for the new coming x new images/logos 2 LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 1/29

  4. Introduction Machine learning What is machine learning (ML)? ML is programming computers (algorithms) to optimize a performance criterion using example data or past experience . Learning a task Learn general models from data to perform a specific task f . f w : x − → y x : input y : output (target, label) w : parameters of f f ( x ; w ) = y From training to predicting the future: Learn to predict Train the model using data examples ( x , y ) 1 Predict the y new for the new coming x new images/logos 2 LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 1/29

  5. Introduction Machine learning applications Face detection/recognition Image classification Handwriting recognition(postal address recognition, signature verification, writer verification, historical document analysis (DocExplore http://www.docexplore.eu )) Speech recognition, Voice synthesizing Natural language processing (sentiment/intent analysis, statistical machine translation, Question answering (Watson), Text understanding/summarizing, text generation) Anti-virus, anti-spam Weather forecast Fraud detection at banks Mail targeting/advertising Pricing insurance premiums Predicting house prices in real estate companies Win-tasting ratings Self-driving cars, Autonomous robots Factory Maintenance diagnostics Developing pharmaceutical drugs (combinatorial chemistry) Predicting tastes in music (Pandora) Predicting tastes in movies/shows (Netflix) Search engines (Google) Predicting interests (Facebook) Web exploring (sites like this one) Biometrics (finger prints, iris) Medical analysis (image segmentation, disease detection from symptoms) Advertisements/Recommendations engines, predicting other books/products you may like (Amazon) Computational neuroscience, bioinformatics/computational biology, genetics Content (image, video, text) categorization Suspicious activity detection images/logos Frequent pattern mining (super-market) Satellite/astronomical image analysis LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 2/29

  6. Introduction ML in physics Event detection at CERN ( The European Organization for Nuclear Research ) ⇒ Use ML models to determine the probability of the event being of interest. ⇒ Higgs Boson Machine Learning Challenge ( https://www.kaggle.com/c/higgs-boson ) images/logos LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 3/29

  7. Introduction ML in quantum chemistry Computing the electronic density of a molecule ⇒ Instead of using physics laws, use ML ( FAST ). See Stéphane Mallat et al. work: https://matthewhirn. images/logos files.wordpress.com/2016/01/hirn_pasc15.pdf LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 4/29

  8. Function estimation How to estimate f w ? Models Parametric ( w ) vs. non-parametric Estimate f w = train the model using data Training: supervised (use ( x , y )) vs. unsupervised (use only x ) Training = optimizing an objective cost Different models to learn f w Kernel models (support vector machine (SVM)) Decision tree Random forest Linear regression K-nearest neighbor Graphical models Bayesian networks Hidden Markov Models (HMM) Conditional Random Fields (CRF) Neural networks (Deep learning): DNN, CNN, RBM, DBN, RNN. images/logos LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 5/29

  9. Function estimation How to estimate f w ? Models Parametric ( w ) vs. non-parametric Estimate f w = train the model using data Training: supervised (use ( x , y )) vs. unsupervised (use only x ) Training = optimizing an objective cost Different models to learn f w Kernel models (support vector machine (SVM)) Decision tree Random forest Linear regression K-nearest neighbor Graphical models Bayesian networks Hidden Markov Models (HMM) Conditional Random Fields (CRF) Neural networks (Deep learning): DNN, CNN, RBM, DBN, RNN. images/logos LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 5/29

  10. Deep neural network Deep neural networks (DNN) x 1 x 2 ˆ y 1 x 3 y 2 ˆ x 4 x 5 State of the art in many task: computer vision, natual language processing. Training requires large data To speed up the training: use GPUs cards Training deep neural networks is difficult ⇒ Vanishing gradient ⇒ More parameters ⇒ Need more data Some solutions: ⇒ Pre-training technique [ Y.Bengio et al. 06, G.E.Hinton et al. 06 ] images/logos ⇒ Use unlabeled data LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 6/29

  11. Deep neural network Deep neural networks (DNN) x 1 x 2 ˆ y 1 x 3 y 2 ˆ x 4 x 5 State of the art in many task: computer vision, natual language processing. Training requires large data To speed up the training: use GPUs cards Training deep neural networks is difficult ⇒ Vanishing gradient ⇒ More parameters ⇒ Need more data Some solutions: ⇒ Pre-training technique [ Y.Bengio et al. 06, G.E.Hinton et al. 06 ] images/logos ⇒ Use unlabeled data LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 6/29

  12. Deep neural network Deep neural networks (DNN) x 1 x 2 ˆ y 1 x 3 y 2 ˆ x 4 x 5 State of the art in many task: computer vision, natual language processing. Training requires large data To speed up the training: use GPUs cards Training deep neural networks is difficult ⇒ Vanishing gradient ⇒ More parameters ⇒ Need more data Some solutions: ⇒ Pre-training technique [ Y.Bengio et al. 06, G.E.Hinton et al. 06 ] images/logos ⇒ Use unlabeled data LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 6/29

  13. Context Semi-supervised learning General case: Data = { labeled data , unlabeled data } � �� � � �� � expensive (money, time), few cheap, abundant E.g: medical images ⇒ semi-supervised learning: Exploit unlabeled data to improve the generalization images/logos LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 7/29

  14. Context Pre-training and semi-supervised learning The pre-training technique can exploit the unlabeled data A sequential transfer learning performed in 2 steps: Unsupervised task ( x labeled and unlabeled data) 1 Supervised task ( ( x , y ) labeled data) 2 images/logos LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 8/29

  15. Pre-training technique and semi-supervised learning Layer-wise pre-training: auto-encoders x 1 x 2 ˆ y 1 x 3 ˆ y 2 x 4 x 5 A DNN to train images/logos LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 9/29

  16. Pre-training technique and semi-supervised learning Layer-wise pre-training: auto-encoders 1) Step 1: Unsupervised layer-wise training Train layer by layer sequentially using only x (labeled or unlabeled) x 1 ˆ x 1 x 2 ˆ x 2 x 3 ˆ x 3 x 4 ˆ x 4 x 5 ˆ x 5 images/logos LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 10/29

  17. Pre-training technique and semi-supervised learning Layer-wise pre-training: auto-encoders 1) Step 1: Unsupervised layer-wise training Train layer by layer sequentially using only x (labeled or unlabeled) x 1 h 1 , 1 x 2 h 1 , 2 x 3 h 1 , 3 x 4 h 1 , 4 x 5 h 1 , 5 images/logos LITIS lab., Apprentissage team - INSA de Rouen, France Deep multi-task learning with evolving weights 10/29

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend