Abstract
Deep convolutional generative adversarial networks (GAN) allow for creating images from existing databases. We applied a modified light-weight GAN (FastGAN) algorithm to cerebral blood flow SPECTs and aimed to evaluate whether this technology can generate created images close to real patients. Investigating three anatomical levels (cerebellum, CER; basal ganglia, BG; cortex, COR), 551 normal (248 CER, 174 BG, 129 COR) and 387 pathological brain SPECTs using N-isopropyl p-I-123-iodoamphetamine (123I-IMP) were included. For the latter scans, cerebral ischemic disease comprised 291 uni- (66 CER, 116 BG, 109 COR) and 96 bilateral defect patterns (44 BG, 52 COR). Our model was trained using a three-compartment anatomical input (dataset ‘A’; including CER, BG, and COR), while for dataset ‘B’, only one anatomical region (COR) was included. Quantitative analyses provided mean counts (MC) and left/right (LR) hemisphere ratios, which were then compared to quantification from real images. For MC, ‘B’ was significantly different for normal and bilateral defect patterns (P < 0.0001, respectively), but not for unilateral ischemia (P = 0.77). Comparable results were recorded for LR, as normal and ischemia scans were significantly different relative to images acquired from real patients (P ≤ 0.01, respectively). Images provided by ‘A’, however, revealed comparable quantitative results when compared to real images, including normal (P = 0.8) and pathological scans (unilateral, P = 0.99; bilateral, P = 0.68) for MC. For LR, only uni- (P = 0.03), but not normal or bilateral defect scans (P ≥ 0.08) reached significance relative to images of real patients. With a minimum of only three anatomical compartments serving as stimuli, created cerebral SPECTs are indistinguishable to images from real patients. The applied FastGAN algorithm may allow to provide sufficient scan numbers in various clinical scenarios, e.g., for “data-hungry” deep learning technologies or in the context of orphan diseases.
Similar content being viewed by others
Introduction
In recent years, the use of artificial intelligence (AI) based on neural networks in medical imaging has been rapidly expanding. Improved diagnostic accuracy of this technology is tightly linked to enlarged network models, which are also associated with higher costs, computing to train AI and increasing efforts of data annotation, e.g., by human experts1,2. As such, data augmentation based on processing techniques provides images similar to supervised data, e.g., by applying geometric deformation, brightness, saturation changes, random cropping, or mix-ins to natural images3. Of note, such conventional techniques of data expansion may be limited for molecular or conventional imaging, mainly caused by asymmetry of organ shapes and locations, less standardized protocols for patient’s orientation, constraints of quantifications, or varying receptor density, glucose consumption or blood flow on a subcellular level3,4.
Relative to the afore-mentioned methods of data augmentation applying direct alterations to existing images, generative adversarial networks (GANs) are based on artificial neural networks to create images closely following the feature distributions of a set of supervised images. In brief, those neural networks consist of generator and discriminator: a generator produces images with features resembling real-world images, and a discriminator segregates between real and the generated images5. GAN uses a simple training strategy of competing generator and discriminator against each other to synthesize images closely resembling real ones. In this regard, GAN is a promising technology for medical imaging, and has been actively studied for various purposes such as data augmentation, modality conversion, segmentation, super-resolution, denoising and reduction of radiation exposure for medical imaging4,6,7,8,9,10,11. Focusing on data augmentation, prior studies of GAN in magnetic resonance imaging have reported on the generation of 2D images with lesions12, generation of 3D images13, conversion of abnormal images to normal images14, and synthesis of brain images that reflect age-related changes15. In addition, previous studies on positron emission tomography images have also demonstrated that image generation by independently learning images of different stages of cognitive decline are feasible (including normal cases, mild cognitive impairment, and Alzheimer’s disease)16. Moreover, in the field of neuroimaging, previous investigators focused on conversion of 11C Pittsburgh compound B images17, or 18F-florbetapir18, e.g., to obtain sufficient number of training cases for computer-aided diagnosis. Taken together, in most of those studies, GAN-generated images were then applied to augment imbalanced datasets or “data-hungry” deep learning technologies, without the need of labeling by expert readers. Prior to a more widespread adoption of this technology, however, such generated scans should be validated, e.g., by comparing with real images, preferably among a broad spectrum of different disease conditions19. In this regard, the number of needed scans increases relative to the number of different conditions attributed to the underlying disease. For instance, in patients with cerebral ischemia using N-isopropyl p-I-123-iodoamphetamine (123I-IMP) SPECT, various defect patterns can be recorded, e.g., affecting only one hemisphere or global reduced blood flow20. Minimizing the number of training data fed into GAN, however, would be desirable, as it would enable for an increased use of this application even if only a small sample size of supervised images is available.
To address this issue, a light-weight GAN (FastGAN) has been recently proposed to enable learning with a smaller set of supervised real data, thereby allowing to reduce the number of initially provided items serving as stimuli21. Incorporating a conditioning mechanism into FastGAN, we aimed to generate brain images of uni- and bilateral cerebral ischemia using 123I-IMP SPECT. Created scans were then validated by quantitative comparison with images of real patients, which allowed to determine whether FastGAN-based scans of reduced cerebral blood flow resemble their real equivalents.
Material and methods
SPECT procedures and training data
250 patients (age, 61.0 ± 16.4 years, 96/250 (38.4%) female), which had undergone 123I-IMP brain SPECTs to assess cerebral ischemia were included in this retrospective analysis. Given the retrospective nature of this study, informed consent was waived by the institutional review board at Saitama Medical University International Medical Center (#2022-016), which also approved the study. All procedures were carried out following current guidelines22. Images were performed under rest and stress condition at one day and thus, a total of 500 scans were available for analyses. We used a Siemens Symbia 16 SPECT/CT system (Siemens Healthineers, Erlangen, Germany), equipped with the quantitative SPECT (QSPECT) reconstruction program and split-dose autoradiographic (ARG) method. SPECT scanning was performed using a low-energy high-resolution collimator with a SPECT condition of continuous rotation at 90 views, 2 min/rotation, 2 cycles of 7 repeats. Data were reconstructed according to the quantitative assessment of rest- and acetazolamide-CBF QSPECT/DTARG protocol23,24. Matrix size of the slices was 64 × 64 and each SPECT image was reoriented perpendicular to the anterior–posterior commissure line. Detailed image reconstruction is also provided in25. Slices at three anatomical levels including cerebellum (CER), basal ganglia (BG), and cortex (COR) were selected from whole brain images by a board-certified nuclear medicine physician (T.H.). The number of selected slices for defect patterns and slice levels are shown in Table 1. Two datasets were created. For dataset ‘A’, we used a three-compartment anatomical input, including CER, BG, and COR, while for dataset ‘B’, only one anatomical region (COR) was considered.
Network model
For our network model, we applied the previously published FastGAN21 to conditional GAN19 with modification for specifying defect pattern and adaptation to image matrix size. In brief, the quality of the generated images was improved by alternately training a generator that synthesizes images and a discriminator assessing the authenticity of images.
Generator
For the generator, latent and conditional vectors for specifying defect patterns were taken as input (Fig. 1). In this regard, the latent vector was used for ensuring variety of generated images (vector dimensions, 256). The conditional vector for the dataset ‘A’ included slice levels and cerebral patterns of radiotracer accumulation (as 1 or 0). In this regard, the first three elements corresponded to anatomical levels, while the patterns included normal, uni- and bilateral defects. The conditional vectors for the dataset ‘B’ included only the COR, along with the same defect patterns (Table 2). The generator consisted of blocks with four different roles and a skip-layer excitation module. The fully connected layer embedded the input vector into a 64 dimensional vector. This layer also adjusted the data dimension so that the input block is acceptable. Both the input and up-sampling blocks enlarged the feature maps progressively to produce more detailed image, while the output block generated a monochromatic brain SPECT image from the input feature maps. Moreover, the skip-layer excitation module enabled efficient training of the generator through gradient flow between distant layers and self-gating21. Different from the previously published FastGAN, our model used only one skip-layer excitation, as such an approach allowed for reducing the number of parameters needed to be learned along with a small matrix size of input images. Additionally, adaptive pooling layer is omitted as it is considered unnecessary for low-resolution images (Fig. 2).
Generator network in our model. The latent vector and the conditional vector for specifying patterns of radiotracer accumulation served as input to the generator to synthesize a two-dimensional brain SPECT. Symbols F, n, s and p denote channels of output feature maps, number of neurons, strides and padding, respectively. The “same” for padding indicates that padding is applied to the input feature map so that the height and width of the input and output feature maps are not changed. GLU is a gating unit proposed in42. Tanh is a hyperbolic tangent activation function. Loss function \({\mathcal{L}}_{G}\) is defined as Eq. (1).
Skip-layer excitation module used in the generator. Symbols H, W and F in feature maps denote height, width and channels, respectively. Symbols s, p and a denote strides, padding and slope of Leaky ReLU activation function, respectively. For padding, “none” indicates that no padding is applied to the input feature map.
Loss function of the generator \({\mathcal{L}}_{G}\) was given by21:
where \(z\) was a latent vector sampled from standard normal distribution, \(y\) was a conditional vector, \(G\left(z|y\right)\) was a synthesized image by the generator and \(D\left(\cdot \right)\) was real/fake logits for \(G\left(z|y\right)\) predicted by the discriminator.
Discriminator
For the discriminator, two types of images (one being real and one being generated) served as inputs which had to be discriminated (Fig. 3). For specifying the pattern of radiotracer accumulation, a conditional image was applied, in which number of channels were 6 and 3 for the dataset ‘A’ and ‘B’, respectively. Each channel corresponded to an element of the conditional vector in the generator, and all pixel values in a channel are 0 or 1. The input image was processed through input and down-sampling blocks to extract the features of the real image, while the output block processed the feature maps from the last down-sampling block, thereby assessing the probability that the input image was real as the following Equations21.
where \({\mathcal{L}}_{real}\) and \({\mathcal{L}}_{fake}\) were adversarial loss for real and generated images. \(x\) and \(\widehat{x}\) were sampled from real images \({I}_{real}\) and the generated images \(G\left(z|y\right)\), respectively. \(D\left(x\right)\) represented real/fake logits for the input \(x\).
Discriminator in our model. The discriminator uses as input the real or generated image with the conditional image representing the pattern of radiotracer accumulation. Symbols F, s, p and a denote channels of output feature maps, strides, padding and slope of Leaky ReLU activation function, respectively. For padding, the “same” indicates that padding is applied to the input feature map so that the height and width of the input and output feature maps are not changed, “none” indicates that no padding is applied to the input feature map. GLU is a gating unit proposed in42. Loss function \({\mathcal{L}}_{real}\), \({\mathcal{L}}_{fake}\) and \({\mathcal{L}}_{reocn}\) are defined as Eqs. (2, 3, 4), respectively.
To efficiently learn features of real images, self-supervised learning was employed with cropping and simple decoders. Briefly, regional feature maps with half height and half width were cropped at a random location of the feature map from the second down-sampling block. The feature map and the global feature map from the third down-sampling block were input to the simple decoders to reconstruct the regional and whole real image from these feature maps. The similarities between the reconstructed image and a real image at regional and global levels and the same location in the real image were evaluated by21:
The loss \({\mathcal{L}}_{recon}\) was evaluated on only real images. \({\mathcal{B}}_{1}\left(x\right)\) and \({\mathcal{B}}_{2}\left(x\right)\) was feature maps from the second and third down-sampling block, \({\mathcal{G}}_{1}\left(\cdot \right)\) was a function contained cropping and processing by the decoder on \({\mathcal{B}}_{1}\left(x\right)\), \(\mathcal{T}\left(x\right)\) was a function of cropping on sample \(x\), and \({\mathcal{G}}_{2}\left(\cdot \right)\) was a function by the decoder on \({\mathcal{B}}_{2}\left(x\right)\).
Total loss of the discriminator \({\mathcal{L}}_{D}\) was given by:
Training process
Each slice was normalized by the maximum count of the slice. To increase number of samples, the following data augmentation was performed: for weighted averaged slices, \(z\) was calculated using the target slice \({z}_{1}\) and a slice \({z}_{2}\),
where \({z}_{2}\) was randomly selected from adjacent in craniocaudal direction, while weight \(w\) was a random number ranging from 0 to 1. The weighted average slice was translated in anteroposterior direction by \(t\) pixel. The value of \(t\) was a random integer from -2 to 2. A horizontal flip was applied to real slices of normal and bilateral patterns as the discriminator inputs at random. The generator and discriminator were trained alternately in the following steps: (i) Synthesized images were outputted by the generator. (ii) The loss of the discriminator to the real image was calculated based on Eqs. (3,5). (iii) The loss of the discriminator to the generated image was calculated based on Eq. (4). (iv) The total loss of the discriminator was computed based on Eq. (2). (v) The loss of the generator was calculated based on Eq. (1). (vi) The parameters of the generators and discriminators were updated using the corresponding losses and Adam optimizers26 with \({\upbeta }_{1}=0.9\), \({\upbeta }_{2}=0.999\) and learning rates = \(2\times {10}^{-4}\). Our model was trained with 1000 epochs and batch size 4 using the dataset ‘A’ and ‘B’, independently on a single NVIDIA RTX 2080 GPU and TensorFlow 2.2.
Testing of both datasets for resembling real images
The epochs, where highest accuracy for real images were given, were determined by a board-certified nuclear medicine physician with 10 years of experience (T.H.) for the two trained models. Using the model parameters at the determined epochs, the same number of images as the real ones were generated for normal, uni- and bilateral conditions as described in Table 1. To evaluate the fidelity of the generated images by our GAN with parameters of the selected epoch, we calculated mean counts (MC), average of pixel-wise standard deviation (SD), and count ratios of left to right hemisphere (LR) for real and generated images trained with datasets ‘A’ and ‘B’. Inter-slice variation within each pattern of real and generated images was assessed using pixelwise SD map.
Statistical analysis
For comparing real images and images of both datasets ‘A’ and ‘B’, we applied one-way ANOVA with GraphPad Prism 9 (San Diego, CA, USA). A P < 0.05 was considered statistically significant.
Results
Epochs with highest accuracy were 826 and 651 for dataset ‘A’ and ‘B’, respectively. On a visual assessment, image quality of generated images with dataset ‘A’ was superior when compared to ‘B’ (Fig. 4).
Real images, generated images trained with dataset ‘A’ (with three-compartment levels serving as stimuli), and dataset ‘B’ (only providing one anatomical level as input). On a visual assessment, dataset ‘A’ including more anatomical information resembles real images more closely than generated images by dataset ‘B’.
For MC, ‘A’ revealed comparable findings when compared to real images, including normal (P = 0.8) and pathological scans (unilateral, P = 0.99; bilateral, P = 0.68). ‘B’ was significantly different for normal and bilateral defect patterns (P < 0.0001, respectively), but not for unilateral ischemic disease (P = 0.77). For LR, comparable results were recorded. For ‘A’, only uni- (P = 0.03), but not normal or bilateral defect scans (P ≥ 0.08) reached significance when compared to real images. For ‘B’, however, bilateral defects (P = 0.01), normal scans and unilateral ischemia (P < 0.0001, respectively) were significantly different (Fig. 5). As such, dataset ‘B’ was significantly different for virtually all investigated semi-quantitative parameters, while scans created by dataset ‘A’ were closely resembling to their real equivalents (except for unilateral disease on LR). Similar results were achieved for both pixel-wise average and SD. Both maps of normal and unilateral conditions for dataset ‘B’ were remarkably different from those calculated for dataset ‘A’ (Fig. 6).
Whisker Plots for comparing real images and generated images for datasets ‘A’ and ‘B’. First row: mean counts, second row: left to right hemisphere ratio (LR). Except for unilateral defect patterns on LR, all comparisons of ‘A’ with real images failed to reach significance. On the other hand, for dataset ‘B’, statistical significance was reached in almost all cases (expect for mean counts of unilateral ischemia), supporting the notion that ‘A’ (using more anatomical input) provides scans closely resembling real scans. *, ** and **** denote P < 0.05, P < 0.01 and P < 0.0001, respectively.
Discussion
Applying blood flow 123I-IMP SPECTs to our novel modified FastGAN, created scans were indistinguishable to acquired images from real patients, including normal studies and various degrees of ischemia. Although our developed neural network aimed at minimizing the number of training data, at least three anatomical compartments were still required to acquire images closely resembling scans of real patients. As such, if reasonable, but still rather limited amounts of supervised stimuli are provided, the applied FastGAN algorithm may allow to yield sufficient number of molecular brain scans for various clinical scenarios, e.g., for less balanced datasets in the context of orphan diseases or “data-hungry” deep learning technologies.
123I-IMP SPECTs have been frequently utilized to assess different degrees of cerebral ischemia, e.g., after head trauma27, stroke28, for identifying epileptogenic foci prior to surgical interventions or to differentiate between mild cognitive impairment and different types of dementia29. Of note, all of these studies enrolled a sufficiently large number of patients, while for other brain disorders, adequate patient recruitment may be challenging, e.g., to detect left or right hemispheric abnormalities in patients affected with Creutzfeldt-Jakob disease using 123I-IMP30. In this regard, augmenting cerebral blood flow scans of subjects with such an orphan disease may be helpful to test the clinical utility of this imaging modality, e.g., to differentiate between uni- or bilateral defect patterns. To the best of our knowledge, our modified FastGAN allowed for the first time to create artificial real equivalents using 123I-IMP SPECTs across a broad spectrum of disease patterns (Fig. 4). This technology is based on a neuronal network using both real images from actual patients fed to the GAN, a generator (trying to provide real images) and a discriminator (verifying whether the created scan is real or an imitation)31. The ongoing contest between both opponents along with a feedback loop will then help the discriminator to optimize its capability to determine which images should be classified as real, while the generator will learn creating scans more closely resembling real images31. As with every AI application, the number of initially provided scans serving as stimulus is of importance. As such, we aimed to reduce the number of needed real input images by applying only one skip-layer excitation, as such an approach allowed for minimizing the number of parameters needed to be learned along with a small matrix size of input images. Although dataset ‘A’ using CER, BG, and COR as input provided more realistic images than ‘B’ (only utilizing COR), we only applied a maximum of three anatomical compartments to create images that are indistinguishable to their real equivalents of patients (Fig. 5). Also partially explaining the superior performance of ‘A’ relative to ‘B’, the number of supervised data of normal, uni- and bilateral cerebral ischemia was rather imbalanced in the present study. For instance, representing diversity for each defect pattern, the radiotracer accumulation in pixel-wise SD maps of bilateral ischemia generated by dataset ‘A’ were lower than the real images, in particular for the frontal and occipital lobe (Fig. 6). This indicates that a lack of diversity within a specific pattern of cerebral ischemia may also lead to less realistic images. To overcome this issue, mini-batch standard deviation could be effective. In this regard, both the generator and discriminator develop consecutively, e.g., by adding more and more details during the training process, ultimately leading to further stabilization of the produced scans32. Taken together, the stimulus for the GAN should be carefully evaluated, e.g., for a bias of imbalanced supervised images for each defect pattern and anatomical location. Nonetheless, this may be challenging in a real-world scenario, where clinical cases provided to a GAN cannot always cover the entire spectrum of a certain disease31.
Moreover, as another limitation of this study, cerebral ischemia was restricted to three discrete patterns, ranging from normal scans, uni- and bilateral defects. Bigolin Lanfredi et al. have recently proposed a GAN model to visualize the progression of chronic obstructive pulmonary disease. The model incorporated a regression subnetwork to learn features in X-chest images for quantitative disease severity based on forced expiratory volume/forced lung capacity33 and such an adversarial regression training could also be incorporated on brain SPECTs. Another solution to reach the goal of quantitative severity in the generated images could be application of latent space. In this regard, a latent vector serving as input source of the desired image is searched. By editing that latent vector, variations of the desired image are generated. The use of latent space has been reported for natural images34,35,36, but it has also been used in the context of modality transformation for medical images37. As another limitation, our novel GAN was only applied to one specific disease using one single radiotracer and thus, our model should be validated among a broad spectrum of different radiopharmaceuticals for SPECT or positron emission tomography frequently applied in the clinic, e.g., 18F-labeled prostate-specific membrane antigen or somatostatin receptor directed PET38,39,40. Future studies should also address the aspect of three-dimensional images. A novel approach was recently proposed using 123I-ioflupane SPECT, which aimed to mimic characteristics of Parkinson's disease by integrating a transformer-based technique, which is based on a framework different from GAN41. In this regard, consecutive slices of a three-dimensional image were used and a neural network model was trained to generate slices for the rest of the region.
Conclusion
We developed a light-weight GAN model for brain SPECT imaging that allowed us to create normal scans, but also varying degrees of cerebral ischemia closely resembling realistic images. In this context, we successfully applied a limited number of supervised data serving as input with a maximum of three anatomical compartments. As such, if reasonable amounts of supervised stimuli are provided, the applied FastGAN algorithm may allow to yield sufficient number of molecular brain scans for various clinical scenarios, e.g., for imbalanced datasets in the context of orphan diseases or “data-hungry” deep learning technologies.
Data availability
The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.
References
Ching, T. et al. Opportunities and obstacles for deep learning in biology and medicine. J. R. Soc. Interface 15(141), 20170387 (2018).
Chartrand, G. et al. Deep learning: A primer for radiologists. Radiographics 37(7), 2113–2131 (2017).
Shorten, C. & Khoshgoftaar, T. M. A survey on image data augmentation for deep learning. J. Big Data 6(1), 60 (2019).
Yi, X., Walia, E. & Babyn, P. Generative adversarial network in medical imaging: A review. Med Image Anal. 58, 101552 (2019).
Goodfellow, I.J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., et al. Generative Adversarial Networks. ArXiv e-prints [Internet]. 2014 June 01, 2014. https://ui.adsabs.harvard.edu/#abs/2014arXiv1406.2661G.
Vey, B. L., Gichoya, J. W., Prater, A. & Hawkins, C. M. The role of generative adversarial networks in radiation reduction and artifact correction in medical imaging. J. Am. Coll. Radiol. 16(9), 1273–1278 (2019).
Koshino, K. et al. Narrative review of generative adversarial networks in medical and molecular imaging. Ann. Transl. Med. 9(9), 821 (2021).
Cheng, Z., Wen, J., Huang, G. & Yan, J. Applications of artificial intelligence in nuclear medicine image generation. Quant. Imaging Med. Surg. 11(6), 2792–2822 (2021).
Arabi, H., AkhavanAllaf, A., Sanaat, A., Shiri, I. & Zaidi, H. The promise of artificial intelligence and deep learning in PET and SPECT imaging. Phys. Med. 83, 122–137 (2021).
Apostolopoulos, I. D., Papathanasiou, N. D., Apostolopoulos, D. J. & Panayiotakis, G. S. Applications of generative adversarial networks (GANs) in positron emission tomography (PET) imaging: A review. Eur. J. Nucl. Med. Mol. Imaging https://doi.org/10.1007/s00259-022-05805-w (2022).
Matsubara, K., Ibaraki, M., Nemoto, M., Watabe, H. & Kimura, Y. A review on AI in PET imaging. Ann. Nucl. Med. 36(2), 133–143 (2022).
Kazuhiro, K. et al. Generative adversarial networks for the creation of realistic artificial brain magnetic resonance images. Tomography. 4(4), 159–163 (2018).
Kwon, G., Han, C., Kim, D-s. Generation of 3D Brain MRI Using Auto-Encoding Generative Adversarial Networks. arXiv e-prints [Internet]. 2019 August 01, 2019:[arXiv:1908.02498 p.]. https://ui.adsabs.harvard.edu/abs/2019arXiv190802498K.
Sun, L. et al. An adversarial learning approach to medical image synthesis for lesion detection. IEEE J. Biomed. Health Inform. 24(8), 2303–2314 (2020).
Xia, T. et al. (eds) Consistent Brain Ageing Synthesis (Springer International Publishing, 2019).
Islam, J. & Zhang, Y. GAN-based synthetic brain PET image generation. Brain Info. 7(1), 3 (2020).
Kimura, Y. et al. AI approach of cycle-consistent generative adversarial networks to synthesize PET images to train computer-aided diagnosis algorithm for dementia. Ann. Nucl. Med. 34(7), 512–515 (2020).
Kang, S. K., Choi, H. & Lee, J. S. Translating amyloid PET of different radiotracers by a deep generative model for interchangeability. Neuroimage 232, 117890 (2021).
Mirza, M., Osindero, S. Conditional Generative Adversarial Nets. arXiv e-prints [Internet]. 2014 November 01, 2014:[arXiv:1411.784 p.]. https://ui.adsabs.harvard.edu/abs/2014arXiv1411.1784M.
Nagamachi, S. et al. A comparative study of 123I-IMP SPET and CT in the investigation of chronic-stage head trauma patients. Nucl. Med. Commun. 16(1), 17–25 (1995).
Bingchen L, Yizhe Z, Kunpeng S, Ahmed E. Towards Faster and Stabilized GAN Training for High-fidelity Few-shot Image Synthesis. ICLR 2021 Conference. 2021.
Latchaw, R. E. et al. Guidelines and recommendations for perfusion imaging in cerebral ischemia: A scientific statement for healthcare professionals by the writing group on perfusion imaging, from the council on cardiovascular radiology of the American heart association. Stroke 34(4), 1084–1104 (2003).
Kim, K. et al. Quantitative mapping of basal and vasareactive cerebral blood flow using split-dose 123I-idoamphetamine and single photon emission computed tomography. Neuroimage https://doi.org/10.1016/j.neuroimage.2006.06.064 (2006).
Iida, H. et al. Multicenter evaluation of a standardized protocol for rest and acetazolamide cerebral blood flow assessment using a quantitative SPECT reconstruction program and split-dose 123I-iodoamphetamine. J. Nucl. Med. 51(10), 1624–1631 (2011).
Yamauchi, M. et al. Quantitative assessment of rest and acetazolamide CBF using quantitative SPECT reconstruction and sequential administration of 123I-iodoamphetamine: Comparison among data acquired at three institutions. Ann. Nucl. Med. 28(9), 836–850 (2014).
Kingma, D.P., Ba, J. Adam: A Method for Stochastic Optimization. arXiv e-prints [Internet]. 2014 December 01, 2014. https://ui.adsabs.harvard.edu/#abs/2014arXiv1412.6980K.
Ito, H., Ishii, K., Onuma, T., Kawashima, R. & Fukuda, H. Cerebral perfusion changes in traumatic diffuse brain injury IMP SPECT studies. Ann. Nucl. Med. 11(2), 167–172 (1997).
Hayashida, K. et al. Change of accumulation and filling pattern in evolution of cerebral infarction with I-123 IMP brain SPECT. Neuroradiology 33(1), 9–14 (1991).
Kanetaka, H. et al. Differentiating mild cognitive impairment, alzheimer’s disease, and dementia with lewy bodies using cingulate island sign on perfusion IMP-SPECT. Front. Neurol. https://doi.org/10.3389/fneur.2020.568438 (2020).
Jibiki, I. et al. Utility of 123I-IMP SPECT brain scans for the early detection of site-specific abnormalities in Creutzfeldt-Jakob disease (Heidenhain type): A case study. Neuropsychobiology 29(3), 117–119 (1994).
Erickson, B. J. & Cai, J. Magician’s corner: 5 generative adversarial networks. Radiology 2(2), e190215 (2020).
Karras, T., Aila, T., Laine, S., Lehtinen, J. Progressive Growing of GANs for Improved Quality, Stability, and Variation. arXiv e-prints [Internet]. 2017 October 01, 2017: [arXiv:1710.10196 p]. https://ui.adsabs.harvard.edu/abs/2017arXiv171010196K.
Bigolin Lanfredi, R., Schroeder, J.D., Vachet, C., Tasdizen, T. Adversarial regression training for visualizing the progression of chronic obstructive pulmonary disease with chest x-rays. arXiv e-prints [Internet]. 2019 August 01, 2019. https://ui.adsabs.harvard.edu/abs/2019arXiv190810468B.
Zhu, J-Y., Krähenbühl, P., Shechtman, E., Efros, A.A. Generative Visual Manipulation on the Natural Image Manifold2016 September 01, 2016: [arXiv:1609.03552 p.]. https://ui.adsabs.harvard.edu/abs/2016arXiv160903552Z.
Zhu, J., Shen, Y., Zhao, D., Zhou, B. In-Domain GAN Inversion for Real Image Editing2020 March 01, 2020:[arXiv:2004.00049 p.]. https://ui.adsabs.harvard.edu/abs/2020arXiv200400049Z.
Abdal, R., Qin, Y., Wonka, P., editors. Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space? In 2019 IEEE/CVF International Conference on Computer Vision (ICCV) (IEEE, Seoul, Korea, 2019).
Fetty, L. et al. Latent space manipulation for high-resolution medical image synthesis via the StyleGAN. Z. Med. Phys. 30(4), 305–314 (2020).
Werner, R. A. et al. (18)F-labeled, PSMA-targeted radiotracers: Leveraging the advantages of radiofluorination for prostate cancer molecular imaging. Theranostics 10(1), 1–16 (2020).
Werner, R. A. et al. The theranostic promise for Neuroendocrine Tumors in the late 2010s—Where do we stand, where do we go?. Theranostics 8(22), 6088–6100 (2018).
Yordanova, A. et al. Peptide receptor radionuclide therapy combined with chemotherapy in patients with neuroendocrine tumors. Clin. Nucl. Med. 44(5), e329–e335 (2019).
Watanabe, S., Ueno, T., Kimura, Y., Mishina, M. & Sugimoto, N. Generative image transformer (GIT): Unsupervised continuous image generative and transformable model for [123I] FP-CIT SPECT images. Ann. Nucl. Med. https://doi.org/10.1007/s12149-021-01661-0 (2021).
Dauphin YN, Fan A, Auli M, Grangier D. Language Modeling with Gated Convolutional Networks. In: Doina P, Yee Whye T, editors. Proceedings of the 34th International Conference on Machine Learning; Proceedings of Machine Learning Research: PMLR; 2017. 933--41.
Funding
Open Access funding enabled and organized by Projekt DEAL. This work was supported partially by Grant-in-Aid for Scientific Research (Grant Number 22H03027) of Japan Society for the Promotion of Science (JSPS) and was also supported by the “RECTOR” program at Okayama (TH). This project was also partially supported by the German Research Foundation (DFG, 453989101, TH, RAW; 507803309, RAW). This publication was supported by the Open Access Publication Fund of the University of Wuerzburg.
Author information
Authors and Affiliations
Contributions
K.K., R.A.W. and T.H. designed this study and drafted the manuscript. They also processed data and conducted analysis. I.K. provided data. N.N., Y.M and I.K. were involved in data creation. All authors revised the manuscript critically.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Werner, R.A., Higuchi, T., Nose, N. et al. Generative adversarial network-created brain SPECTs of cerebral ischemia are indistinguishable to scans from real patients. Sci Rep 12, 18787 (2022). https://doi.org/10.1038/s41598-022-23325-3
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-022-23325-3








