deep reinforcement learning
play

Deep Reinforcement Learning Olexandr Isayev, Ph.D. University of - PowerPoint PPT Presentation

De Novo molecular design with Deep Reinforcement Learning Olexandr Isayev, Ph.D. University of North Carolina at Chapel Hill @olexandr olexandr@unc.edu http://olexandrisayev.com About me Ph.D. in Chemistry (computational) Minor in CS/ML


  1. De Novo molecular design with Deep Reinforcement Learning Olexandr Isayev, Ph.D. University of North Carolina at Chapel Hill @olexandr olexandr@unc.edu http://olexandrisayev.com

  2. About me Ph.D. in Chemistry (computational) Minor in CS/ML Worked in Federal research lab on HPC & GPU computing to solve chemical problems Now I am faculty at the University of North Carolina, Chapel Hill We use ML & AI to solve challenging problems in chemistry http://olexandrisayev.com Twitter: @olexandr olexandr@unc.edu

  3. A public-private partnership that supports the discovery of new medicines through open access research www.thesgc.org

  4. The Long and Winding Road to Drug Discovery Data Science approaches useful across the pipeline, but very different techniques aim for success, but if not: fail early, fail cheap

  5. internal rate of return (IRR) Source: Endpoints News https://endpts.com

  6. Drowning in Data …but starving for Knowledge

  7. The growing appreciation of molecular modeling and informatics 7

  8. “Behold the rise of the machines”

  9. Summary of recent AI-based studies on chemical library design Molecular Generative models Method of biasing representations generated compounds • Autoencoders • None • Fingerprints • Generative • Latent space • SMILES adversarial optimization • Fine-tuning on small models (GANs) • Graphs • Recurrent neural subset of molecules networks (RNNs) with the desired • Convolutional property • Reinforcement neural networks (CNNs) Learning

  10. De Novo molecular design with Deep Reinforcement Learning General Approach Application to Molecular design Tm; LogP; pIC50; etc Predictive Deep Network Molecules Generative Deep Network Patent pending arXiv:1711.10907

  11. Drug discovery pipeline PREDICTIVE CHEMICAL CHEMICAL PROPERTY/ QSAR MODELS STRUCTURES DESCRIPTORS ACTIVITY QSAR MAGIC CHEMICAL DATABASE HITS VIRTUAL (confirmed SCREENING actives) ~10 6 – 10 9 molecules INACTIVES (confirmed inactives)

  12. Design of the ReLeaSE* method Challenges: • Generate chemically feasible SMILES • Develop SMILES- based QSAR model • Employ Predictive ML model to bias library generation *Popova, Mariya, Olexandr Isayev, and Alexander Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  13. Language of SMILEs

  14. Generative model 1.5M molecules from <START>c1ccc(O)cc1<END> c 1 c c c c 1 <END> ChEMBL ) ( F ) c1ccc(O)cc1 c 1 c c c ( O ) c c 1 <START> NO + loss + loss Did the YES Softmax training loss converge ?

  15. Reinforcement learning for chemical design Generative model FC(F)COc1ccc2c(Nc3ccc(Cl)c(Cl)c3)ncnc2c1 Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  16. Reinforcement learning for chemical design Generative model FC(F)COc1ccc2c(Nc3ccc(Cl)c(Cl)c3)ncnc2c1 Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  17. Reinforcement learning for chemical design Generative model INACTIVE! Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  18. Reinforcement learning for chemical design Generative model INACTIVE! Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  19. Reinforcement learning for chemical design Generative model INACTIVE! Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  20. Reinforcement learning for chemical design Generative model Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  21. Reinforcement learning for chemical design Generative model Fc1ccc2c(Nc3ccc(F)c(F)c3)ncnc2c1 Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  22. Reinforcement learning for chemical design Generative model Fc1ccc2c(Nc3ccc(F)c(F)c3)ncnc2c1 Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  23. Reinforcement learning for chemical design Generative model ACTIVE! Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  24. Reinforcement learning for chemical design Generative model ACTIVE! Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  25. Reinforcement learning for chemical design Generative model ACTIVE! Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  26. Reinforcement learning for chemical design Generative model Predictive model M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  27. Technical details • Models were trained on Nvidia Titan X and Titan V GPUs • Training generative model on ChEMBL took ~ 25 days • Training predictive models took ~ 2 hours • Biasing generative model with reinforcement learning for one property ~ 1 day • Generative model produces 1000s compounds per minute

  28. Results: Biasing target properties in the designed libraries Optimized Baseline * -2 0 2 4 6 8 10 12 Partition coefficient (logP) JAK2 Inhibition (pIC50) M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  29. Train data distribution JAK2 (Kinase) inhibition Maximized property distribution Minimized property distribution NEW CHEMOTYPE CAS 236-084-2 ZINC37859566 (buffer reagent) New molecule SIMILAR SCAFFOLDS arXiv:1711.10907

  30. Results: analysis of similarity Distribution of Tanimoto similarity to the nearest neighbor in training dataset for compounds predicted to be active for EGFR by consensus of QSAR models: Similarity= 0.69 Similarity= 0.57 Similarity = 0.86 0.5 0.6 0.7 0.8 0.9 1.0 Tanimoto similarity

  31. Results: Synthetic accessibility score* of the designed libraries *Ertl, Peter, and Ansgar Schuffenhauer. "Estimation of synthetic accessibility score of drug-like molecules based on molecular complexity and fragment contributions." Journal of cheminformatics 1.1 (2009): 8.

  32. Target predictions for generated compounds using SEA* *Keiser MJ, Roth BL, Armbruster BN, Ernsberger P, Irwin JJ, Shoichet BK. Relating protein pharmacology by ligand chemistry. Nat Biotech 25 (2), 197-206 (2007).

  33. Target predictions for generated compounds using SEA* *Keiser MJ, Roth BL, Armbruster BN, Ernsberger P, Irwin JJ, Shoichet BK. Relating protein pharmacology by ligand chemistry. Nat Biotech 25 (2), 197-206 (2007).

  34. Model visualization for JAK2 (projection using t-SNE) 10 ZINC469992 9 pIC50 = 8.23 8 ZINC19982368 ZINC2876515 pIC50 = 8.64 7 pIC50 = 8.39 6 5 4 pIC50 = 0.63 ZINC66347860 3 pIC50 = 3.31 2 ZINC3549031 1 pIC50 = 10.37 pIC50 = 3.76 M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  35. Examples of Stack-RNN cells with interpretable gate activations M. Popova, O. Isayev, A. Tropsha. "Deep reinforcement learning for de-novo drug design." arXiv preprint arXiv:1711.10907 (2017).

  36. Summary • AI methods coupled with SMILES representation afford biased libraries generation • The system naturally embeds reinforcement to produce novel structure with the desired property • The system can be tuned to bias libraries towards specific property ranges • Next phase is experimental validation of hits by UNC SGC team

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend