regularization effect of large initial learning rate
play

Regularization Effect of Large Initial Learning Rate Yuanzhi Li* - PowerPoint PPT Presentation

Regularization Effect of Large Initial Learning Rate Yuanzhi Li* Colin Wei* Tengyu Ma Carnegie Mellon University Stanford University Stanford University Large Initial Learning Rate is Crucial for Generalization Large Initial Learning Rate is


  1. Regularization Effect of Large Initial Learning Rate Yuanzhi Li* Colin Wei* Tengyu Ma Carnegie Mellon University Stanford University Stanford University

  2. Large Initial Learning Rate is Crucial for Generalization

  3. Large Initial Learning Rate is Crucial for Generalization • Common schedule: large initial learning rate + annealing

  4. Large Initial Learning Rate is Crucial for Generalization • Common schedule: large initial learning rate + annealing • … But small learning rate: better train and test performance up until annealing Train Accuracy Val Accuracy annealing

  5. Large Initial Learning Rate is Crucial for Generalization • Common schedule: large initial learning rate + annealing • … But small learning rate: better train and test performance up until annealing Train Accuracy Val Accuracy annealing • Large LR outperforms small LR after annealing!

  6. LR schedule changes order of learning patterns => generalization

  7. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures”

  8. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures” • Ignores other patterns, harming generalization

  9. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures” • Ignores other patterns, harming generalization • Large initial LR + annealing learns easy-to-fit patterns first

  10. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures” • Ignores other patterns, harming generalization • Large initial LR + annealing learns easy-to-fit patterns first • Only memorizes hard-to-fit patterns after annealing

  11. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures” • Ignores other patterns, harming generalization • Large initial LR + annealing learns easy-to-fit patterns first • Only memorizes hard-to-fit patterns after annealing • => learns to use all patterns, helping generalization!

  12. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures” • Ignores other patterns, harming generalization • Large initial LR + annealing learns easy-to-fit patterns first • Only memorizes hard-to-fit patterns after annealing • => learns to use all patterns, helping generalization! • Intuition: larger LR • ⇒ larger noise in activations

  13. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures” • Ignores other patterns, harming generalization • Large initial LR + annealing learns easy-to-fit patterns first • Only memorizes hard-to-fit patterns after annealing • => learns to use all patterns, helping generalization! • Intuition: larger LR • ⇒ larger noise in activations • ⇒ effectively weaker representational power

  14. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures” • Ignores other patterns, harming generalization • Large initial LR + annealing learns easy-to-fit patterns first • Only memorizes hard-to-fit patterns after annealing • => learns to use all patterns, helping generalization! • Intuition: larger LR • ⇒ larger noise in activations • ⇒ effectively weaker representational power • ⇒ won’t overfit to “signatures”

  15. LR schedule changes order of learning patterns => generalization • Small LR quickly memorizes hard-to-fit “class signatures” • Ignores other patterns, harming generalization • Large initial LR + annealing learns easy-to-fit patterns first • Only memorizes hard-to-fit patterns after annealing • => learns to use all patterns, helping generalization! • Intuition: larger LR • ⇒ larger noise in activations • ⇒ effectively weaker representational power • ⇒ won’t overfit to “signatures” • Non-convexity is crucial: different LR schedules find different solutions

  16. Demonstration on Modified CIFAR10

  17. Demonstration on Modified CIFAR10 Group 1: 20% examples with hard-to-generalize, easy-to- fit patterns original image

  18. Demonstration on Modified CIFAR10 Group 1: 20% examples with Group 2: 20% examples with hard-to-generalize, easy-to- easy-to-generalize, hard-to- fit patterns fit patterns original image hard-to-fit patch indicating class

  19. Demonstration on Modified CIFAR10 Group 1: 20% examples with Group 2: 20% examples with Group 3: 60% examples hard-to-generalize, easy-to- easy-to-generalize, hard-to- with both patterns fit patterns fit patterns original image hard-to-fit patch indicating class

  20. Demonstration on Modified CIFAR10 Group 1: 20% examples with Group 2: 20% examples with Group 3: 60% examples hard-to-generalize, easy-to- easy-to-generalize, hard-to- with both patterns fit patterns fit patterns original image hard-to-fit patch indicating class • Small LR memorizes patch, ignores rest of the image • ⇒ learns image from 20% examples

  21. Demonstration on Modified CIFAR10 Group 1: 20% examples with Group 2: 20% examples with Group 3: 60% examples hard-to-generalize, easy-to- easy-to-generalize, hard-to- with both patterns fit patterns fit patterns original image hard-to-fit patch indicating class • Small LR memorizes patch, ignores rest of the image • ⇒ learns image from 20% examples • Large initial LR initially ignores patch, only learns it after annealing • ⇒ learns image from 80% examples

  22. Theoretical Setting Group 1: 20% examples with Group 2: 20% examples with Group 3: 60% examples hard-to-generalize, easy-to- easy-to-generalize, hard-to- with both patterns fit patterns fit patterns linearly classifiable patterns

  23. Theoretical Setting Group 1: 20% examples with Group 2: 20% examples with Group 3: 60% examples hard-to-generalize, easy-to- easy-to-generalize, hard-to- with both patterns fit patterns fit patterns clustered but not linearly classifiable patterns linearly separable

  24. Theoretical Setting Group 1: 20% examples with Group 2: 20% examples with Group 3: 60% examples hard-to-generalize, easy-to- easy-to-generalize, hard-to- with both patterns fit patterns fit patterns Contains both clustered but not linearly classifiable patterns patterns linearly separable

  25. Conclusion

  26. Conclusion • Small LR optimizes faster, but generalizes worse than large initial LR + annealing

  27. Conclusion • Small LR optimizes faster, but generalizes worse than large initial LR + annealing • Explanation: order of learning pattern types • Easy-to-generalize, hard-to-fit patterns • Hard-to-generalize, easy-to-fit patterns

  28. Conclusion • Small LR optimizes faster, but generalizes worse than large initial LR + annealing • Explanation: order of learning pattern types • Easy-to-generalize, hard-to-fit patterns • Hard-to-generalize, easy-to-fit patterns • SGD noise from large LR is mechanism for regularization

  29. Conclusion • Small LR optimizes faster, but generalizes worse than large initial LR + annealing • Explanation: order of learning pattern types • Easy-to-generalize, hard-to-fit patterns • Hard-to-generalize, easy-to-fit patterns • SGD noise from large LR is mechanism for regularization Come find our poster: 10:45 AM -- 12:45 PM @ East Exhibition Hall B + C #144!

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend