differentiable mutual information and matrix exponential
play

Differentiable Mutual Information and Matrix Exponential for - PowerPoint PPT Presentation

Differentiable Mutual Information and Matrix Exponential for Multi-Resolution Image Registration Abhishek Nan, Matthew Tennant, Uriel Rubin, Nilanjan Ray Medical Imaging with Deep Learning, 2020 https://github.com/abnan/DRMIME Registration


  1. Differentiable Mutual Information and Matrix Exponential for Multi-Resolution Image Registration Abhishek Nan, Matthew Tennant, Uriel Rubin, Nilanjan Ray Medical Imaging with Deep Learning, 2020 https://github.com/abnan/DRMIME

  2. Registration

  3. With MSE Source: https://anhir.grand-challenge.org/

  4. Problems? Multi-modal images ● MSE won’t work ● Source: https://www.mathworks.com/discovery/image-registration.html

  5. Solution? Mutual Information ●

  6. Issues? Mutual Information for images is computed using joint histograms. ● Histograms are not differentiable. ● No gradient descent? ●

  7. Differentiable mutual information The function T is realized by a neural ● network with parameter θ . V( θ ) is differentiable and can be used as a ● objective function in place of MI.

  8. MINE for images

  9. Currently

  10. Matrix exponential Matrix exponential of a square matrix A is given by the following: ● Geometric transformation matrices can be obtained by exponential of a linear combination of basis ● matrices.

  11. Matrix Exponential (Examples)

  12. Why matrix exponentials?

  13. So far...

  14. More problems? Medical/Microscopy images often are extremely high resolution. So gradient descent can be ● extremely slow. Optimization for neural networks is non-convex. ●

  15. Solution? Gaussian Pyramids ● Source: https://en.wikipedia.org/wiki/Pyramid_(image_processing)

  16. Can we do better? ● What if we did simultaneous optimization for all levels? ○ Each level of optimization is for different images. So MI between them changes as well. Solution? ● A single MINE can be trained for all of these! ○ How? ● Mini-batches can be constructed by sampling from all levels ○

  17. What about the loss? Since we are doing simultaneous optimization, with modern deep learning frameworks, it’s very ● easy to combine the loss from each level and perform joint optimization. For eg, for just 1 level: ● Loss = MI(F, G v (M)) ○ For 4 levels: ● Loss = (¼) * [MI(F 1 , G v (M 1 )) + MI(F 2 , G v (M 2 )) + MI(F 3 , G v (M 3 )) + MI(F 4 , G v (M 4 ))] ○

  18. Evaluation Public datasets ● Available ground truth ●

  19. Results

  20. Source: https://projects.ics.forth.gr/cvrl/fire/

  21. Thank you!

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend