Title: "Deep learning and Bayesian inference for medical imaging" (Grant No. 702666)
The project was carried out from 2016/10/1 to 2018/09/30 within the Marie Skłodowska-Curie Action (Individual Fellowship), financed by the European Comission.
The aim of the project is to apply current state-of-the-art techniques of deep learning to medical imaging. Additionally, we aim at enriching deep learning models with Bayesian inference to better reflect uncertainty in medical data.
Volume-preserving Normalizing Flows
One approach is to apply a generative modelling to medical data. For this purpose we used variational auto-encoders (VAE) that utilize neural networks to model probability distributions between visible and latent variables. Further, in order to increase flexibility of the VAE we proposed a new volume-preserving flow using a series of Householder transformations. In other words, we transform a vector of latent variables to new vector of latent variables using the Householder transformation. As a result, we obtain a series (levels) of latent variables and the final one follows more flexible posterior distribution. The proposed approach was applied to grayscale image patches of histopathological data.
Another idea was to improve on the linear inverse auto-regressive flow (LinIAF). The LinIAF applies a linear transformation to latent variables with a lower-triangular matrix with ones on the diagonal. In order to increase the flexibility of this normalizing flow we proposed to use a convex combination of lower-triangular matrices. As a result, we obtained two levels of latent variables and the final one follows more flexible posterior distribution. The proposed approach was also applied to grayscale image patches of histopathological data.
See publications for details (NIPS Workshop on Bayesian Deep Learning 2016 & Benelearn 2017).
Sylvester Normalizing Flows
Volume-preserving normalizing flows allow to obtain a richer family of variational posteriors, however, these are limited to linear transformations. In order to overcome this problem, we proposed a non-linear transformation by utilizing the Sylvester's determinant identity. As a result, we obtained a class of non-linear inversible transformations. The main idea of the approach is to parameterize weights of the transformations using orthogonal matrices. In order to allow efficient calculations, we proposed to use Householder matrices and a numerical procedure to parameterize orthogonal matrices. The proposed approach was applied to different image benchmark datasets.
See publications for details (UAI 2018).
New prior for VAE
We proposed to extend the variational auto-encoder (VAE) framework with a new type of prior ("Variational Mixture of Posteriors" prior, or VampPrior for short). The VampPrior consists of a mixture distribution (e.g., a mixture of Gaussians) with components given by variational posteriors conditioned on learnable pseudo-inputs. We further extended this prior to a two layer hierarchical model and showed that this architecture, where prior and posterior are coupled, learns significantly better models. The model also avoids the usual local optima issues that plague VAEs related to useless latent dimensions. We provided empirical studies on six datasets, namely, static and dynamic MNIST, OMNIGLOT, Caltech 101 Silhouettes, Frey Faces and Histopathology patches, and show that applying the hierarchical VampPrior delivers state-of-the-art results on all datasets in the unsupervised permutation invariant setting and the best or close to SOTA methods in the invariant setting.
Further, we utilized the VampPrior in the fair classification setting. Fairness is very statistical important property in many practical applications including medicine. We proposed a two-level hierarchical VAE with a class label and a sensitive variable. In order to enforce fairness, we used a regularization term based on the mutual information measure.
See publications for details (AISTATS 2018, ICML Workshop 2018).
Medical imaging using Deep Multiple Instance Learning
Training a whole slide imaging tool requires relatively large amount of computational resources. Similarly, providing pixel level annotations interferes pathologists daily work. In order to overcome these issues, we proposed to apply multi-instance learning combined with deep learning to histopathology classification. Our goal is to utilize weakly-labaled data to train deep learning models in an end-to-end fashion. In the research we discussed different permutation-invariant operators and proposed a new one basing on the attention mechanism that is fully trainable.
See publications for details (NIPS Workshop 2017, MIDL 2018, ICML 2018).