Abstract
To generate and evaluate synthesized postoperative OCT images of epiretinal membrane (ERM) based on preoperative OCT images using deep learning methodology. This study included a total 500 pairs of preoperative and postoperative optical coherence tomography (OCT) images for training a neural network. 60 preoperative OCT images were used to test the neural networks performance, and the corresponding postoperative OCT images were used to evaluate the synthesized images in terms of structural similarity index measure (SSIM). The SSIM was used to quantify how similar the synthesized postoperative OCT image was to the actual postoperative OCT image. The Pix2Pix GAN model was used to generate synthesized postoperative OCT images. Total 60 synthesized OCT images were generated with training values at 800 epochs. The mean SSIM of synthesized postoperative OCT to the actual postoperative OCT was 0.913. Pix2Pix GAN model has a possibility to generate predictive postoperative OCT images following ERM removal surgery.
Similar content being viewed by others
Introduction
Epiretinal membrane (ERM) is an avascular fibrocellular proliferative layer that develops on the surface of the inner retina. In the previous studies of ERM, the prevalence has been reported to be 6.0–11.8% for westerners, and 2.2–7.9% for Asians, and ERM is more common in people over 50 years of age1,2,3,4,5,6. When ERM causes structural changes on the macula, it causes symptoms such as monocular diplopia, metamorphopsia, or reduced visual acuity. Surgical ERM removal is a well-established treatment of choice to improve visual quality in symptomatic patients with ERM7,8,9.
Recently, optical coherence tomography (OCT) has been used to quantify the structural changes of retina10. Many studies have reported the correlations between postoperative visual improvement and retinal microstructure in OCT images in ERM patients. These include preoperative photoreceptor integrity11,12, presence of ectopic inner foveal layer13, the ratio of foveal outer nuclear layer (ONL) thickness to the juxtafoveal ONL plus outer plexiform layer thickness14, preoperative inner nuclear layer thickness15,16, inner retinal irregularity index17, photoreceptor outer segment length18, central foveal thickness and central inner retinal layer thickness19.
Regarding postoperative OCT parameters related to visual prognosis in ERM, Kim et al. showed a significant correlation between final BCVA and early postoperative central macular thickness (CMT). The CMT at 1 month after surgery was more related to the final visual outcome than CMT at 3 and 6 months20. Romano et al. reported that an increase in retinal nerve fiber layer thickness at 1 month and a decrease in ONL thickness at 3 and 6 months follow-up were associated with a good final BCVA21. Mahmoudzadeh et al. demonstrated that postoperative inner retinal microcysts or ellipsoid zone disruption at 3 month was related to poor final visual outcome22.
However, despite successful ERM surgery, some patients are dissatisfied with their postoperative visual acuity. This may be because the patient’s expectations are too high or the patient’s surgical outcome is really lower than the surgeon’s empirical expectations.
Many clinicians’ efforts and their studies related to the various prognostic factors of ERM surgery have enabled surgeons to explain the expected visual outcome of surgery with greater predictive power to the patients before the surgery. However, since no one can time travel, the postoperative OCT biomarkers cannot be used to predict surgical outcome in advance of the surgery. If the postoperative OCT image can be specified in details before surgery, it is expected to help improve the prognosis prediction of the surgery.
Recently, with the rapid growth of researches on artificial intelligence, efforts are being made to use it in various ways in the field of medicine23. In particularly, deep neural network has made a remarkable progression. Generative adversarial network (GAN), a representative deep neural network, was firstly developed by Ian Goodfellow24,25. GAN is a model that generates a variety of realistic fake images originated from the data distribution in datasets. GAN consists of two learning models. One is a model that generates a false synthetic image, called a generator (G), and it is trained to generate a fake image as real as possible. The other model is called a discriminator (D) and it is trained to determine whether the synthesized image generated by G has been manipulated. If these two models are trained as if they are competing with each other, ideally G can generate realistic fake images25. GAN has been widely used for medical image processing, especially in the field of radiology. Previous studies have shown that GAN can be used for image reconstruction, denoising, data augmentation, domain transfer between modalities, segmentation, classification and detection etc.24,26,27. Shitrit et al. introduced a method that can secure good quality images while reducing MRI scan time by generating estimated images using GAN28. Wang et al. have published several papers using GAN which include a study on segmenting the basal membrane in histopathology images of microinvasive cervix carcinoma, a study on removing metal artifacts from CT images, and converting low-quality PET brain images into high-quality images29,30,31. Ben-Cohen et al. demonstrated domain translation, which converts a CT image into a PET image32. In a study on liver lesion classification, Frid-Adar et al showed that synthetic data augmentation using GAN can improve the performance of the classifier33.
In the field of ophthalmology, several studies have been attempted to predict intervention results by generating predictive post interventional images using GAN34,35,36.
Therefore, the author would like to confirm whether it is possible to predict the postoperative structural changes of the patient in ERM using deep learning methodology known as GAN. The results of this study are expected to serve as a stepping stone for further researches to analyze the relationship between the predicted retinal structure and visual prognosis of ERM surgery in the future.
Materials and methods
This study adhered to the tenets of the Declaration of Helsinki and received approval from the Nune Eye Hospital Institutional Review Board/Ethics Committee (N-2304-001-999). The requirement for informed consent was waived by the Nune Eye Hospital Institutional Review Board/Ethics Committee because of the retrospective nature of this study.
Study design and dataset creation
The author reviewed the medical records of patients who were diagnosed with idiopathic ERM and underwent ERM removal surgery at Nune Eye Hospital in Seoul, South Korea from January 1, 2011 to December 31, 2022. Patients with secondary ERM, myopia of more than 26mm of axial length, presence of any other retinal disease were excluded.
All surgeries were performed by 12 retinal surgeons, and 25-gauge pars plana vitrectomy (PPV) was performed under retrobulbar anesthesia using the Constellation Vision System (Alcon, Fort Worth, TX, USA) and a high refractive index contact lens. After core vitrectomy, triamcinolone acetonide or 0.25% indocyanine green (ICG) was used to visualize the ERM, followed by removal of the epiretinal and internal limiting membranes using intraocular forceps. According to the doctor's decision, phacoemulsification and intraocular lens implantation were performed together with PPV for patients who needed cataract surgery.
A pair of horizontal preoperative and postoperative OCT (Spectralis, Heidelberg Engineering, Heidelberg, Germany) scan image of each patient’s macular center was obtained. The authors manually selected images passing through the fovea in the OCT volume scan. The author used the OCT image examined at 1 month after surgery as the postoperative OCT image. Pre and postoperative OCT image pairs were divided into two sets; a training dataset for model training and a test dataset for model evaluation.
All Macular OCT images were cropped from their original resolution of 1008 × 596 pixels to 496 × 496 pixels in order to exclude en face image and resized to 256 × 256 pixels for deep learning (Fig. 1, Supplementary Fig. 1).
Synthesizing postoperative OCT image
Conditional generative adversarial network
Conditional generative adversarial network (cGAN) is a conditional version of GAN. Since GAN is an unconditioned generative model, realistic images are randomly generated and there is no option to control the synthesized image. Mirza et al. proposed a cGAN in which conditional data y can be added to G and D37.
The objective function of cGAN is defined as follows:
D: discriminator, G: generator, Pdata: probability distribution of dataset, Pz: noise distribution to be used to sample a random noise vector, x: image in dataset, z: random noise vector, y: a condition.
In cGAN, it can be modulated to generate images corresponding to a condition y, such as class labels.
Pix2Pix (image to image translation with cGAN)
In 2017, Isola et al. proposed Pix2Pix, a universal image to image translation with cGAN, that can effectively perform various image translation tasks without changing the framework to be fit to specific applications38.
In GAN, G is trained to learn a mapping to the actual data space from random noise. On the other hand, in cGAN, G is learning a mapping to the actual data space from random noise plus a specific condition. Meanwhile, in Pix2Pix, G is trained to learn a mapping from the input image, which is a condition, to the actual ground truth.
The objective function of Pix2Pix is defined as follows38:
D: discriminator, G: generator, x: image in dataset, z: random noise vector, y: ground truth image, L: loss function.
Pix2Pix generates data that seems like real, such as the equation of (3). As defined in the equation of (4), the generated image is synthesized such that the difference from the actual ground truth is minimized. Therefore, Pix2Pix generates specific deterministic images corresponding to the ground truth images, rather than generating a variety of realistic images that are characteristic of previous GAN or cGAN.
We use Pix2Pix model to generate predictive postoperative OCT images. The Pix2Pix network architecture was implemented using Python (V.3.6) programming language in PyCharm with Anaconda virtual environment as described in the original article38. The G architectures are as follows.
Encoder: C64-C128-C256-C512-C512-C512-C512-C512,
U-Net decoder: CD512-CD1024-CD1024-C1024-C1024-C512-C256-C128,
The D architectures are as follows:
C64-C128-C256-C512,
Ck: a Convolution-BatchNorm-ReLU layer with k filters. CDk: a Convolution-BatchNorm- Dropout-ReLU layer with a dropout rate of 50%. Convolutions: 4 × 4 spatial filters applied with stride 2. Convolutions in the encoder, and in the discriminator, downsample by a factor of 2, whereas in the decoder they upsample by a factor of 2. All ReLUs in the encoder and the discriminator are leaky, with slope 0.2, while ReLUs in the decoder are not leaky38.
In Pix2Pix, G was trained to produce a synthesized postoperative OCT image with preoperative OCT image as an input. D was trained to determine whether the synthesized postoperative image was a real postoperative OCT or a fake image created by the G (Fig. 2).
All training and testing were performed using Google Colaboratory, which can run a .py file.
Evaluation of synthesized postoperative OCT image
Structural similarity index measure
The structural similarity index measure (SSIM) was used to evaluate how similar the synthesized postoperative OCT image was to the real postoperative OCT image. The SSIM is a well-known image quality assessment metrics to measure the similarity between two given images based on statistical changes in image luminance, contrast and structure. The value of SSIM is between 0 and 1. the closer to 1, the more similar the two images are39.
Image binarization using adaptive thresholding
The author binarized images to minimize the effects of differences in luminance and contrast of the corresponding retinal layers between the synthesized image and the real OCT image when SSIM evaluation. Binarization is to classify image pixels as either black and white. The author binarized the images using an adaptive binarization method known to provide qualified segmentation despite strong illumination changes in the images40.
Region of interest
We set the region of interest (ROI) from the innermost surface of retina to the boundary of choroid on binarized images using the AssistedFreehand function (hand-drawn ROI, where the line drawn automatically follows edges in the underlying image) in MATLAB and then calculated the each SSIM. The author named the SSIM of the preoperative image to the postoperative image as SSIM A, and the SSIM of the synthesized image to the postoperative image as SSIM B. All these assessments were performed using MATLAB R2021a software (Fig. 3).
Statistical analysis
Statistical analysis was performed using the R software. To compare baseline characteristics between each dataset, the Wilcoxon rank sum test for the age, the chi-square test for the sex and the combined cataract surgery, the Cochran-Armitage test for ERM stage, and the t test for the axial length were used. P < 0.05 was considered statistically significant.
Results
Baseline characteristics
The training dataset consisted of a total of five hundred pairs of OCT images and the test dataset consisted of sixty pairs of OCTs. Therefore, total sixty synthesized postoperative OCT images were generated based on the preoperative OCT images in the test dataset. In the training dataset, the mean age of patients was 63.70 ± 8.00 years, and 59.8% were female. In the test dataset, the mean age of patients was 64.27 ± 6.39 years, and 70.0% were female. All ERM stages were stage 2 or higher in both datasets. A total of 313 (62.6%) cases in the training data set and 40 (66.7%) cases in the test data set had received combined cataract surgery with PPV. Axial length was 23.35 ± 0.75 in the training dataset and 23.42 ± 0.89 in the testing dataset (Table 1).
Pix2Pix model training and synthesized OCT images from the test dataset
The model was trained with 500 pairs of OCT images for 800 epochs, batch size 10, Adam optimizer, learning rate of 2 × 10–4, with random jitter and mirroring. The synthesized OCT images were generated using training values for 800 epochs. A total of sixty synthesized postoperative OCT images were generated (Fig. 4, Supplementary Fig. 2).
Similarity evaluation of synthesized OCT images
The SSIM A and B were calculated for sixty image pairs, respectively. For 58 out of 60 images, the SSIM of the synthesized images was higher than that of the preoperative images (Fig. 5). The two images where SSIM B is lower than SSIM A are shown in (A) and (B) of Fig. 6.
Representative images for SSIM A and SSIM B. SSIM structural similarity index measure, SSIM A SSIM for preoperative to real postoperative OCT, SSIM B SSIM for synthesized postoperative to real postoperative OCT, Pre preoperative OCT, Post real an actual postoperative OCT, Post synth synthesized postoperative OCT.
Discussion
In this study, the author generated and evaluated synthesized postoperative OCT images to predict structural changes that occur after ERM removal surgery.
Similarity evaluation
The similarity of OCT images between synthesized and real postoperative was evaluated using SSIM. SSIM is one of the image quality assessment (IQA) methods.
Traditional method of IQA is subjective human evaluation to determine how good the quality of an image is compared to a reference image. This is usually too inconvenient, time-consuming and expensive to acquire reliable conclusions. To overcome these drawbacks, methods for conducting objective IQA have been developed. There are several popular IQAs. The most popular and widely used IQA is mean squared error, calculated by averaging the squared intensity differences between image pixels. But these kind of IQAs are not very well matched to the image quality perceived by human. SSIM was developed in an effort to evaluate image quality in the way that humans perceive images. It is based on the hypothesis that the human visual system is highly adapted with the perception of structural information. SSIM is expressed in luminance, contrast, and structure information. After grouping the pixels of the image in small window units and calculating each SSIM, the mean SSIM is obtained through the average of the values, and a value between [0, 1] is presented as an evaluation index. The closer the value is from 0 to 1, the more similar the images are. If 1, both images are identical39,41.
Many researches on image-to-image translation are being actively conducted in the field of medicine. Several previous studies using SSIM as an image quality assessment presented various SSIM values ranging from 0.570 to 0.92041,42,43,44,45,46.
Our results show 0.913 of mean SSIM value for the synthesized images to real postoperative images. The SSIM value of each synthesized image is higher than SSIM of the preoperative OCT to real postoperative image except for two cases (Figs. 5, 6). Of the two cases, one was not synthesized properly, and the other case had the mildest preoperative retinal distortion (Fig. 6A,B). Therefore, it can be reasonably concluded that the author generated predictive postoperative OCT images by using deep learning methodology known as Pix2Pix. This implies that a kind of cGAN methodology has the potential to be used to synthesize predictive postoperative OCT images of ERM patients.
There were five images with SSIM B values less than 0.9 (Fig. 6B–F). Three of these were ERM stage 4 and the other two were ERM stage 3. Among the test dataset images, there were a total of three of ERM stage 4, and all of them have relatively low SSIM values. Therefore, the more deformed the retina, the more difficult it was to predict postoperative OCT images in our model. Other characteristics of them, such as axial length and degree of myopia, were not different from the other patients. Meanwhile, the lowest SSIM value of 0.87044 for image shown in (B) of Fig. 6 seems to be too high considering that image (B) was not generated satisfactorily. Therefore, there remains room for complementing image quality assessment to evaluate the synthesized images.
Post-intervention predictions using GAN in the field of ophthalmology
Three studies reported post therapeutic predictions using GAN in the field of ophthalmology47. To our knowledge, Yoo et al. proposed a concept to generate the post therapeutic predictive images using GAN at the first time34. He presented a synthesized postoperative appearance after orbital decompression surgery. Yoo et al. trained a GAN model with109 image pairs of pre and postoperative facial images of patient who underwent orbital decompression surgery due to thyroid-associated ophthalmopathy with data augmentation. They explained that the predicted postoperative images were similar to the actual postoperative images and that images generated using GANs would help patients predict the impact of orbital decompression surgery in advance. Although the resolution of the synthesized image was relatively low (64 × 64 pixels), the result showed the applicability of GAN to oculoplastic surgery. The other two studies were on treatment prediction for patients with exudative age-related macular degeneration (ARMD). Liu et al. trained a GAN model with 479 image pairs of pre and post therapeutic OCT images (256 × 256 pixels) of patient who received intravitreal anti vascular endothelial growth factor (VEFG) injection, and then the performance of the GAN model was evaluated using 50 OCT image pairs in the test sets. Liu et al. showed that 92% of the synthesized images had acceptable quality with an 85% accuracy to predict the post treatment macular status in their study. They insisted that GAN has great potential to synthesize post interventional OCT images with good quality and high accuracy in ARMD treatment35. Lee et al. proposed a conditional GAN with a multi-channel input for the post interventional prediction for anti-VEGF injection treatment. They trained the GAN model through data augmentation using 723 OCT image pairs (256 × 256 pixels) combined with fluorescein angiography or indocyanine green angiography, and then the performance of the GAN was evaluated using 150 OCT image pairs in the test sets. They showed that a conditional GAN was able to generate post therapeutic OCT images using baseline OCT, fluorescein angiography and indocyanine green angiography images in ARMD treatment36. However, these studies had several limitations in terms of short-term follow-up periods, selection bias of excluding poorly synthesized images.
Limitations
Our study also has limitations that were suggested in the previous studies.
First, we had to downsize the OCT images to train and test the model. This inherently has a possibility of losing the original information for images. The next limitation is short term follow-up period. The shape of the retina gradually changes over time after surgery. Therefore, the longer the follow-up period, the more difficult it is to predict the shape of the retina after ERM surgery. Therefore, a subsequent study on long term follow-up period is needed to confirm the performance of the model. Third, the author evaluated the quality of the synthesized sample images by regularly saving network weights and sample images during training to determine when to stop training. Since the optimal performance point of the generator was not evaluated with the validation dataset, it cannot be free from the possibility of overfitting. Fourth, alignment issues in paired images may affect the accuracy of SSIM. Finally, this study included multiple surgeons (a total of 12 retina specialist) who had different surgical experience. Depending on the surgeon, the extent of removal of the epiretinal membrane and the internal limiting membrane could be different, and the degree of microtrauma to the retina during surgery might not be identical. In other words, each surgeons have different skills. Therefore, this kind of surgeon’s factor has the possibility to affect the structural recovery as a confounding factor.
Conclusion
In conclusion, the author generated postoperative OCT images and demonstrated that Pix2Pix GAN model has the possibility to generate predictive postoperative OCT images following ERM removal surgery. This result is expected to serve as a stepping stone for further researches to analyze the relationship between the predicted retinal structure and visual prognosis of ERM surgery. The authors anticipate that this study may also stimulate further researches on the application of GANs in the field of interventional or therapeutic image prediction in ophthalmology.
Data availability
All relevant data that support the findings of this work are available in this manuscript and supplementary material. The raw datasets are available from the corresponding author upon reasonable request only.
Change history
23 November 2023
A Correction to this paper has been published: https://doi.org/10.1038/s41598-023-47940-w
References
Wise, G. N. Clinical features of idiopathic preretinal macular fibrosis. Schoenberg lecture. Am. J. Ophthalmol. 79(3), 349–7 (1975).
Sidd, R. J., Fine, S. L., Owens, S. L. & Patz, A. Idiopathic preretinal gliosis. Am. J. Ophthalmol. 94(1), 44–8 (1982).
McCarty, D. J. et al. Prevalence and associations of epiretinal membranes in the visual impairment project. Am. J. Ophthalmol. 140(2), 288–94 (2005).
You, Q., Xu, L. & Jonas, J. B. Prevalence and associations of epiretinal membranes in adult Chinese: The Beijing eye study. Eye (Lond.) 22(7), 874–9 (2008).
Kawasaki, R. et al. Prevalence and associations of epiretinal membranes in an adult Japanese population: The Funagata study. Eye (Lond.) 23(5), 1045–51 (2009).
Okamoto, F., Sugiura, Y., Okamoto, Y., Hiraoka, T. & Oshika, T. Associations between metamorphopsia and foveal microstructure in patients with epiretinal membrane. Investig. Ophthalmol. Vis. Sci. 53(11), 6770–5 (2012).
Machemer, R. The surgical removal of epiretinal macular membranes (macular puckers) (author’s transl). Klin. Monbl. Augenheilkd. 173(1), 36–42 (1978).
Margherio, R. R. et al. Removal of epimacular membranes. Ophthalmology 92(8), 1075–83 (1985).
McDonald, H. R., Verre, W. P. & Aaberg, T. M. Surgical management of idiopathic epiretinal membranes. Ophthalmology 93(7), 978–83 (1986).
Schmidt-Erfurth, U. & Waldstein, S. M. A paradigm shift in imaging biomarkers in neovascular age-related macular degeneration. Prog. Retin Eye Res. 50, 1–24 (2016).
Kim, J. H., Kim, Y. M., Chung, E. J., Lee, S. Y. & Koh, H. J. Structural and functional predictors of visual outcome of epiretinal membrane surgery. Am. J. Ophthalmol. 153(1), 103–10.e1 (2012).
Suh, M. H., Seo, J. M., Park, K. H. & Yu, H. G. Associations between macular findings by optical coherence tomography and visual outcomes after epiretinal membrane removal. Am. J. Ophthalmol. 147(3), 473–80.e3 (2009).
González-Saldivar, G., Berger, A., Wong, D., Juncal, V. & Chow, D. R. Ectopic inner foveal layer classification scheme predicts visual outcomes after epiretinal membrane surgery. Retina 40(4), 710–7 (2020).
Hosoda, Y., Ooto, S., Hangai, M., Oishi, A. & Yoshimura, N. Foveal photoreceptor deformation as a significant predictor of postoperative visual outcome in idiopathic epiretinal membrane surgery. Investig. Ophthalmol. Vis. Sci. 56(11), 6387–93 (2015).
Takabatake, M., Higashide, T., Udagawa, S. & Sugiyama, K. Postoperative changes and prognostic factors of visual acuity, metamorphopsia, and aniseikonia after vitrectomy for epiretinal membrane. Retina 38(11), 2118–27 (2018).
Kim, J. H., Kang, S. W., Kong, M. G. & Ha, H. S. Assessment of retinal layers and visual rehabilitation after epiretinal membrane removal. Graefe’s Arch. Clin. Exp. Ophthalmol. 251, 1055–64 (2013).
Cho, K. H., Park, S. J., Cho, J. H., Woo, S. J. & Park, K. H. Inner-retinal irregularity index predicts postoperative visual prognosis in idiopathic epiretinal membrane. Am. J. Ophthalmol. 168, 139–49 (2016).
Shiono, A. et al. Photoreceptor outer segment length: A prognostic factor for idiopathic epiretinal membrane surgery. Ophthalmology 120(4), 788–94 (2013).
Kim, J. Y. et al. Visual prognostic factors of epiretinal membrane surgery in patients with pseudophakia. Ophthalmologica 243(1), 43–50 (2020).
Kim, J. et al. Long-term temporal changes of macular thickness and visual outcome after vitrectomy for idiopathic epiretinal membrane. Am. J. Ophthalmol. 150(5), 701-9e1 (2010).
Romano, M. et al. Variability of visual recovery with time in epiretinal membrane surgery: A predictive analysis based on retinal layer OCT thickness changes. J. Clin. Med. 12(6), 2107 (2023).
Mahmoudzadeh, R. et al. Pars plana vitrectomy for idiopathic epiretinal membrane: OCT biomarkers of visual outcomes in 322 eyes. Ophthalmol. Retina 6(4), 308–17 (2022).
Hamet, P. & Tremblay, J. Artificial intelligence in medicine. Metabolism 69, S36–S40 (2017).
Zhao, J., Hou, X., Pan, M. & Zhang, H. Attention-based generative adversarial network in medical imaging: A narrative review. Comput. Biol. Med. 149, 105948 (2022).
Goodfellow, I.J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y. Generative adversarial nets. In NIPS’2014 (2014).
Yi, X., Walia, E. & Babyn, P. Generative adversarial network in medical imaging: A review. Med. Image Anal. 58, 101552 (2019).
Sorin, V., Barash, Y., Konen, E. & Klang, E. Creating artificial images for radiology applications using generative adversarial networks (GANs)–A systematic review. Acad. Radiol. 27(8), 1175–85 (2020).
Shitrit, O., Riklin Raviv, T. Accelerated magnetic resonance imaging by adversarial neural network. Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: Third International Workshop, DLMIA 2017, and 7th International Workshop, ML-CDS 2017, Held in Conjunction with MICCAI 2017, Québec City, QC, Canada, September 14, Proceedings 3 (Springer, 2017).
Wang, D., Gu, C., Wu, K., Guan, X. Adversarial neural networks for basal membrane segmentation of microinvasive cervix carcinoma in histopathology images. In 2017 International conference on machine learning and cybernetics (ICMLC) (IEEE, 2017).
Wang, J., Zhao, Y., Noble, J. H. & Dawant, B. M. Conditional generative adversarial networks for metal artifact reduction in CT images of the ear. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16–20, 2018, Proceedings, Part I (eds Frangi, A. F. et al.) (Springer, 2018).
Wang, Y. et al. 3D conditional generative adversarial networks for high-quality PET image estimation at low dose. Neuroimage 174, 550–62 (2018).
Ben-Cohen, A. et al. Cross-modality synthesis from CT to PET using FCN and GAN networks for improved automated lesion detection. Eng. Appl. Artif. Intell. 78, 186–94 (2019).
Frid-Adar, M. et al. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing 321, 321–31 (2018).
Yoo, T. K., Choi, J. Y. & Kim, H. K. A generative adversarial network approach to predicting postoperative appearance after orbital decompression surgery for thyroid eye disease. Comput. Biol. Med. 118, 103628 (2020).
Liu, Y. et al. Prediction of OCT images of short-term response to anti-VEGF treatment for neovascular age-related macular degeneration using generative adversarial network. Br. J. Ophthalmol. 104(12), 1735–40 (2020).
Lee, H., Kim, S., Kim, M. A., Chung, H. & Kim, H. C. Post-treatment prediction of optical coherence tomography using a conditional generative adversarial network in age-related macular degeneration. Retina 41(3), 572–80 (2021).
Mirza M, Osindero S. Conditional generative adversarial nets. Preprint at https://arXiv.org/arXiv:14111784.2014.
Isola P, Zhu J-Y, Zhou T, Efros AA. Image-to-image translation with conditional adversarial networks. Preprint at https://arXiv.org/arXiv:161107004.2017 (2016).
Wang, Z., Bovik, A. C., Sheikh, H. R. & Simoncelli, E. P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–12 (2004).
Bradley, D. & Roth, G. Adaptive thresholding using the integral image. J. Graph. Tools 12(2), 13–21 (2007).
Mudeng, V., Kim, M. & Choe, S.-W. Prospects of structural similarity index for medical image analysis. Appl. Sci. 12(8), 3754 (2022).
Mason, A. et al. Comparison of objective image quality metrics to expert radiologists’ scoring of diagnostic quality of MR images. IEEE Trans. Med. Imaging 39(4), 1064–72 (2019).
Zhu, Z., Wahid, K., Babyn, P. & Yang, R. Compressed sensing-based MRI reconstruction using complex double-density dual-tree DWT. J. Biomed. Imaging 2013, 1–12 (2013).
Kim W, Woo S-K, Park J, Sheen H, Lim I, Lim SM, et al. Contrast CT image generation model using CT image of PET/CT. In 2018 IEEE Nuclear Science Symposium and Medical Imaging Conference Proceedings (NSS/MIC), (IEEE, 2018).
Duan, C. et al. Fast and accurate reconstruction of human lung gas MRI with deep learning. Magn. Reson. Med. 82(6), 2273–85 (2019).
Zhang, H., Shinomiya, Y. & Yoshida, S. 3D MRI reconstruction based on 2D generative adversarial network super-resolution. Sensors 21(9), 2978 (2021).
You, A., Kim, J. K., Ryu, I. H. & Yoo, T. K. Application of generative adversarial networks (GAN) for ophthalmology image domains: A survey. Eye Vis. 9(1), 1–19 (2022).
Author information
Authors and Affiliations
Contributions
J.K. designed and conducted the study, analyzed the data, and was major contributor to writing the manuscript. H.S.C. supervised the overall work.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The original online version of this Article was revised: The original version of this Article erroneously included the Acknowledgements. It has consequently been removed.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kim, J., Chin, H.S. Deep learning-based prediction of the retinal structural alterations after epiretinal membrane surgery. Sci Rep 13, 19275 (2023). https://doi.org/10.1038/s41598-023-46063-6
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-023-46063-6








