Improving the diagnosis of retinal diseases through machine learning OCT image generation.
- CRIStAL (Lille Computer Science, Signal and Automation Research Center)
- Temeoo
- The Pitié Salpêtrière University Hospital
The main objective of this project is to develop and evaluate an AI method to generate synthetic retinal OCT images from existing data. This will overcome limitations related to real OCT images acquisition and contribute to advances in ophthalmology research.
The specific objectives are as follows:
- Design and implement a deep learning model capable of generating high quality synthetic retinal OCT images.
- Collect and preprocess an existing retinal OCT dataset, ensuring that it is representative of different retinal conditions.
- Train the AI model to generate synthetic OCT images using the set of preprocessed data.
- Evaluate the performance of the model by comparing the generated synthetic images to real OCT images from patients.
This project will assist in overcoming the lack of real retinal OCT images by providing a powerful tool for generating high-quality synthetic images, therefore contributing to achieving progress in the field of retinopathy and medical imaging.
The goal is to produce DRUSEN-affected OCT images. I recommend you check out the notebooks in the directory which are also available on kaggle and colab. They're quite detailed. The images looks pretty good. I bet training to 512 (original size) would looks good as well.
Examples of generated images at a definition of 256 by 256 pixels :
Examples of real images :
Quick exploration of the paper model architecture. Implementation at 256x256 pixels. Mixed results after training for a few epochs. Maybe need deeper training ? I'll explore pix2pix in the meantime to try to reproduce results.
Back to cycleGAN : I trained for 10 epochs on 2000 images and got way better results (learning rate = 1e-5 different than in the paper = 2e-4) DFE to OCT looks better than the opposite.
I saved the weights but I may not reuse them. Let's allow the model to run on two GPUs and train on kaggle with the paper learning rate. Note : global loss was still decreasing (not identity mapping loss but we set lambda_identity to 0). After training for several more epochs on 20000 samples, results do not look very promising. Maybe I should use the identity loss.
The goal is to map dilated fundus examination (DFE) images to OCT images in order to enhance the diagnosis from a DFE image. We use a training dataset where DFEs are mapped to corresponding OCTs. We'll try to get coherent images that are at least 256px. This is the resolution the Centrale paper achieved. It is yet left to be seen whether using a more advanced network could yoeld coherent results. There is more fundus data available but OCT is more accurate with predicting pathologies. The two may be complementary as well.
I'll implement various knowledge distillation techniques where the teacher model will be trained on OCT images and the student on fundus images.
Relevant paper : Wang, Lehan, Weihang Dai, Mei Jin, Chubin Ou, and Xiaomeng Li. “Fundus-Enhanced Disease-Aware Distillation Model for Retinal Disease Classification from OCT Images.” arXiv, August 1, 2023.
This paper implements the reverse of what we propose to do. It uses DFE images to complement diagnoses from OCT scans. Moreover, DFE and OCT are not paired during training which lessens the data requirements to train the model. We'll study the feasibility of reversing the model to complement DFE images with OCT scans.
This paper implements very interesting loss functions which use the KL divergence. The concepts they develop are class similarity alignment and class prototype matching. OCT models are generally better than DFE models at predicting retina pathologies. However, there is less OCT data available as the imagery technique is costly, which makes predictions from DFE images very interesting.
A striking aspect of the paper is that knowledge distillation techniques are usually implemented from a large cumbersome model to a smaller model. The goal is to get the smaller model to match the performance of the cumbersome one. The paper does the opposite, which they justifiy by arguing that OCT and DFE are complementary, which we'll have to test.
Maybe StyleGAN could help explore the intra-pathology distribution.






