why deep learning is more
play

Why Deep Learning Is More Natural Questions Efficient than Support - PowerPoint PPT Presentation

Main Objectives of . . . Need for Machine . . . Support Vector . . . Deep Learning: a Brief . . . Why Deep Learning Is More Natural Questions Efficient than Support Support Vector . . . Support Vector . . . Vector Machines, and How What


  1. Main Objectives of . . . Need for Machine . . . Support Vector . . . Deep Learning: a Brief . . . Why Deep Learning Is More Natural Questions Efficient than Support Support Vector . . . Support Vector . . . Vector Machines, and How What Are Sparsity . . . Our Explanation It Is Related to Sparsity Home Page Techniques in Signal Title Page Processing ◭◭ ◮◮ ◭ ◮ Laxman Bokati 1 , Vladik Kreinovich 1 , Olga Kosheleva 1 , and Page 1 of 33 Anibal Sosa 2 Go Back 1 University of Texas at El Paso, USA lbokati@miners.utep.edu, olgak@utep.edu, vladik@utep.edu Full Screen 2 Universidad Icesi, Cali, Colombia, hannibals76@gmail.com Close Quit

  2. Main Objectives of . . . Need for Machine . . . 1. Main Objectives of Science and Engineering Support Vector . . . • We want to make our lives better, we want to select Deep Learning: a Brief . . . actions and designs that will make us happier. Natural Questions Support Vector . . . • We want to improve the world so as to increase our Support Vector . . . happiness level. What Are Sparsity . . . • To do that, we need to know: Our Explanation Home Page – what is the current state of the world, and – what changes will occur if we perform different ac- Title Page tions. ◭◭ ◮◮ • Crudely speaking: ◭ ◮ – learning the state of the world and learning what Page 2 of 33 changes will happen is science, while Go Back – using this knowledge to come up with the best ac- Full Screen tions and best designs is engineering. Close Quit

  3. Main Objectives of . . . Need for Machine . . . 2. Need for Machine Learning Support Vector . . . • In some cases, we already know how the world oper- Deep Learning: a Brief . . . ates. Natural Questions Support Vector . . . • E.g., we know that the movement of the celestial bodies Support Vector . . . is well described by Newton’s equations. What Are Sparsity . . . • It is described so well that we can predict, e.g., Solar Our Explanation eclipses centuries ahead. Home Page • In many other cases, however, we do not have such a Title Page good knowledge. ◭◭ ◮◮ • We need to extract the corresponding laws of nature ◭ ◮ from the observations. Page 3 of 33 Go Back Full Screen Close Quit

  4. Main Objectives of . . . Need for Machine . . . 3. Need for Machine Learning (cont-d) Support Vector . . . • In general, prediction means that: Deep Learning: a Brief . . . Natural Questions – we can predict the future value y of the physical Support Vector . . . quantity of interest Support Vector . . . – based on the current and past values x 1 , . . . , x n of What Are Sparsity . . . related quantities. Our Explanation • To be able to do that, we need to have an algorithm Home Page that: Title Page – given the values x 1 , . . . , x n , ◭◭ ◮◮ – computes a reasonable estimate for the desired fu- ◭ ◮ ture value y . Page 4 of 33 Go Back Full Screen Close Quit

  5. Main Objectives of . . . Need for Machine . . . 4. Need for Machine Learning (cont-d) Support Vector . . . • In the past, designing such algorithms was done by Deep Learning: a Brief . . . geniuses: Natural Questions Support Vector . . . – Newton described how to predict the motion of ce- Support Vector . . . lestial bodies, What Are Sparsity . . . – Einstein provided more accurate algorithms, Our Explanation – Schroedinger, in effect, described how to predict Home Page probabilities of different quantum states, etc. Title Page • This still largely remains the domain of geniuses, Nobel ◭◭ ◮◮ prizes are awarded every year for these discoveries. ◭ ◮ • However, now that the computers has become very ef- Page 5 of 33 ficient, they are often used to help. Go Back Full Screen Close Quit

  6. Main Objectives of . . . Need for Machine . . . 5. Need for Machine Learning (cont-d) Support Vector . . . • This use of computers is known as machine learning : Deep Learning: a Brief . . . Natural Questions – we know, in several cases c = 1 , . . . , C , which values y ( c ) corresponded to appropriate values x ( c ) Support Vector . . . 1 , . . . , x ( c ) n ; Support Vector . . . – we want to find an algorithm f ( x 1 , . . . , x n ) for which, What Are Sparsity . . . for all these cases c , we have y ( c ) ≈ f ( x ( c ) 1 , . . . , x ( c ) n ). Our Explanation • The value y may be tomorrow’s temperature in a given Home Page area. Title Page • It may be a binary (0-1) variable deciding, e.g., whether ◭◭ ◮◮ a given email is legitimate or a spam. ◭ ◮ Page 6 of 33 Go Back Full Screen Close Quit

  7. Main Objectives of . . . Need for Machine . . . 6. Machine Learning: a Brief History Support Vector . . . • One of the first successful general machine learning Deep Learning: a Brief . . . techniques was the technique of neural networks . Natural Questions Support Vector . . . • In this technique, we look for algorithms of the type Support Vector . . . � n K � What Are Sparsity . . . � � f ( x 1 , . . . , x n ) = W k · s w ki · x i − w k 0 − W 0 . Our Explanation i =1 k =1 Home Page • Here, a non-linear function s ( z ) is called an activation Title Page function , and values w ki and W k knows as weights . ◭◭ ◮◮ • As the function s ( z ), researchers usually selected the ◭ ◮ so-called sigmoid function Page 7 of 33 1 s ( z ) = 1 + exp( − z ) . Go Back Full Screen • This algorithm emulates a 3-layer network of biological neurons – the main brain cells doing data processing. Close Quit

  8. Main Objectives of . . . Need for Machine . . . 7. Machine Learning: a Brief History (cont-d) Support Vector . . . • In the first layer, we have input neurons that read the Deep Learning: a Brief . . . inputs x 1 , . . . , x n . Natural Questions Support Vector . . . • In the second layer – called a hidden layer – we have Support Vector . . . K neurons each of which: What Are Sparsity . . . – first generates a linear combination of the input Our Explanation n signals: z k = � w ki · x i − w k 0 Home Page i =1 – and then applies an appropriate nonlinear function Title Page s ( z ) to z k , resulting in a signal y k = s ( z k ). ◭◭ ◮◮ • The processing by biological neurons is well described ◭ ◮ by the sigmoid activation function. Page 8 of 33 • This is the reason why this function was selected for Go Back artificial neural networks in the first place. Full Screen • After that, in the final output layer, the signals y k from the neurons in the hidden layer are combined. Close Quit

  9. Main Objectives of . . . Need for Machine . . . 8. Machine Learning: a Brief History (cont-d) Support Vector . . . K Deep Learning: a Brief . . . • The linear combination � W k · y k − W 0 is returned as Natural Questions k =1 the output. Support Vector . . . Support Vector . . . • A special efficient algorithm – backpropagation – was developed to train the corresponding neural network. What Are Sparsity . . . Our Explanation • This algorithm finds the values of the weights that pro- Home Page vide the best fit for the observation results x ( c ) 1 , . . . , x ( c ) n , y ( c ) . Title Page ◭◭ ◮◮ ◭ ◮ Page 9 of 33 Go Back Full Screen Close Quit

  10. Main Objectives of . . . Need for Machine . . . 9. Support Vector Machines (SVM): in Brief Support Vector . . . • Later, in many practical problem, a different technique Deep Learning: a Brief . . . became more efficient: the SVM technique. Natural Questions Support Vector . . . • Let us explain this technique on the example of a binary Support Vector . . . classification problem, i.e., a problem in which: What Are Sparsity . . . – we need to classify all objects (or events) into one Our Explanation of two classes, Home Page – based on the values x 1 , . . . , x n of the corresponding Title Page parameters ◭◭ ◮◮ • In such problems, the desired output y has only two ◭ ◮ possible values; this means that: Page 10 of 33 – the set of all possible values of the tuple x = ( x 1 , . . . , x n ) Go Back – is divided into two non-intersecting sets S 1 and S 2 corresponding to each of the two classes. Full Screen Close Quit

  11. Main Objectives of . . . Need for Machine . . . 10. Support Vector Machines (cont-d) Support Vector . . . • We can thus come up with a continuous f-n f ( x 1 , . . . , x n ) Deep Learning: a Brief . . . such that f ( x ) ≥ 0 for x ∈ S 1 and f ( x ) ≤ 0 for x ∈ S 2 . Natural Questions Support Vector . . . • As an example of such a function, we can take f ( x ) = Support Vector . . . def d ( x, S 2 ) − d ( x, S 1 ), where d ( x, S ) = inf s ∈ S d ( x, s ). What Are Sparsity . . . • If x ∈ S , then d ( x, s ) = 0 for s = x thus d ( x, S ) = 0. Our Explanation Home Page • For points x ∈ S 1 , we have d ( x, S 1 ) = 0 but usually d ( x, S 2 ) > 0, thus f ( x ) = d ( x, S 2 ) − d ( x, S 1 ) > 0. Title Page ◭◭ ◮◮ • For points x ∈ S 2 , we have d ( x, S 2 ) = 0 while, in gen- eral, d ( x, S 1 ) > 0, thus f ( x ) = d ( x, S 2 ) − d ( x, S 1 ) < 0 . ◭ ◮ • In some cases, there exists a linear function that sepa- Page 11 of 33 n rates the classes: f ( x 1 , . . . , x n ) = a 0 + � a i · x i . Go Back i =1 Full Screen • In this case, there exist efficient algorithms for finding the corresponding coefficients a i . Close Quit

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend