Abstract
Brain MR image transformation, which is the process of transforming MR images of one type to another, is critical to several downstream neuroimaging tasks that include brain tissue and lesion volume estimation and lesion detection. In recent years, several deep learning-based methods have been applied to address this task; however, for the most part, they have tended to be deterministic. These methods provide a single transformed output for a given input image with no accompanying measure of confidence in the transformation process. In contrast to this, in this study, we demonstrate how a class of probabilistic conditional generative algorithms can be applied to MR image transformation and quantify the performance of these algorithms. We also demonstrate that the ability to generate multiple transformed images for a given input image can be used to estimate the uncertainty in the output and to detect out-of-distribution (OOD) input images. In particular, we apply conditional Generative Adversarial Networks (cGAN), Noise Conditional Score Networks (NCSN), and Denoising Diffusion Probabilistic Models (DDPM) to transform T1, T2, FLAIR, and proton density (PD) MR images. Through extensive computational experiments, we conclude that the probabilistic algorithms are more accurate than other benchmark methods, and among these, the diffusion models yield the most accurate transformation results. Within the diffusion models, DDPM demonstrates higher performance in terms of similarity metrics, and NCSN exhibits accurate distributional measures and computationally favorable characteristics. We also demonstrate how the generative models can be used to assess the confidence in a given transformation and to detect input images that contain pathology and/or artifacts.
Introduction
Brain MR imaging is the primary method for diagnosing and monitoring patients with many neurological diseases and for performing brain-related research. Additionally, by tailoring MRI sequences, a variety of image types with different anatomical and pathological details can be acquired. However, in many instances, due to reasons that include the physical condition of a subject, the cost of additional acquisition time, and corruption of data, not all of the desired MR image types can be acquired. This can lead to undesired costs and burdens associated with additional scans and the exclusion of some subjects from a research study. This motivates the need to develop algorithms that can transform one MR image type into another.
Over the years, a wide range of methods have been developed for medical image transformation. More recently, Deep Learning (DL)-based methods have proven to be successful in this field1,2. This success stems from the inherent ability of these methods to represent the complex and highly nonlinear mappings that are found in many image transformation tasks. The reader is referred to survey articles on this topic3,4.
Medical image transformation can be categorized as inter-modality, which involves the transformation of images of one modality to another, or intra-modality, which involves the transformation from one image type or contrast to another one within the same modality. Examples of the former include transformation between MR, computed tomography (CT), and positron emission tomography images (PET) images5,6,7,8. Whereas examples of the latter include the transformation of MR images of one type to another. For example, the transformation of T2, T1, and FLAIR MR images. The focus of this work is intra-modality transformation of MR images, and in particular, on the conversion of structural MR images of one type to another.
There are two broad perspectives of viewing the image transformation problem. The first perspective, which is the one that is commonly used, is deterministic. In this case, the transformation itself is deterministic, and for one input, it generates a single instance of the output. Direct inference supervised methods9 can be considered the most straightforward DL-based algorithms based on a deterministic formulation. In these methods, the model receives the input image and transforms it using a DL-based network. The network is trained using a loss function that measures the difference between the predicted and the target image. Different choices of network architectures and loss functions yield different methods. For example, Dinkla et al.10 use a convolutional neural network (CNN) based encoder-decoder and the mean absolute error loss function to transform brain T1 MR images to CT images. Similarly, Xiang et al.11 use a network with multiple embedding blocks and a least squares error loss to transform MR images to CT images.
An alternative perspective of the image transformation problem is the probabilistic approach, which is the main focus of this work. Within this approach, the transformation that maps an input image to the output is assumed to be probabilistic. Therefore, for a single input, it yields multiple possible transformed images that are sampled from a probability distribution conditioned on the input image. A simple thought experiment illustrates why for the brain MR image transformation problem this perspective is more appropriate. It is well known that due to the underlying physics, MR images can display significant variability12. Consider an example in which multiple devices are used to capture T2 images for a single subject. All these images would differ from each other and will still be considered valid T2 images for that subject. In this scenario, if we develop a method to convert a T1 image for this subject into a T2 image, then it is clear that the appropriate output is not a single image; rather, it is a collection of images, all of which may be considered as valid T2 images. In other words, probabilistic models aim to capture the conditional distribution of plausible target outputs given an input image. Rather than being manually specified, this distribution is implicitly learned from the training data and reflects the inherent variability present in real MR scans, arising from factors such as subject positioning, scanner and acquisition protocols, anatomical and pathological differences, as well as variations in image quality and noise characteristics.
Generative models are DL-based models that are based on probabilistic principles and are, therefore, appropriate for a broad range of applications13, including medical image transformation. Among these, conditional generative adversarial networks (cGANs)14,15,16 have been successfully employed in image-to-image transformation tasks, including MR image transformation17,18. More recent studies have extended these approaches using hybrid adversarial architectures, such as MTSR-MRI19, which jointly performs modality translation and super-resolution. In general, GAN-based methods are difficult to work with since their training requires working with an adversarial loss, which leads to a min-max optimization problem that is harder to solve than the more common minimization problem. This difficulty is circumvented by diffusion models that generate multiple samples by utilizing the underlying score function of the probability density function and learn this score function by solving a minimization problem20,21,22. In recent years, diffusion models have found a wide range of applications in medical image transformation23,24,25, including the transformation of MR images26,27,28,29. One drawback of diffusion models is that they incur larger computational costs during the generation process, which happens over several hundred steps. Recently, several studies have focused on reducing computational time while aiming to preserve image quality, often by operating in latent space27. For instance, Ozbey et al. combine the diffusion process with an adversarial projection scheme to achieve faster image sampling during inference30. Also, Jiang et al. introduce Fast-DDPM, a denoising diffusion probabilistic model optimized for medical image-to-image generation31. Fast-DDPM reduces both training and inference time by aligning training and sampling steps and using fewer time steps.
Beyond GAN and diffusion-based models, several recent works explore alternative generative frameworks for medical image transformation. Chen et al.32 propose Diffusion Prior Synthesis and Optimization (DPSO), a source-free diffusion framework that leverages a probability flow ODE and target-domain priors to synthesize cross-modality MR images without access to source-domain data. UniSyn33 is a foundation-style generative model that synthesizes multiple imaging modalities conditioned on metadata, enabling flexible multi-modal image generation. Yazdani et al. propose a flow-matching–based generative framework for efficient and high-quality medical image synthesis34. The method learns continuous transport maps between noise and data distributions via flow matching, allowing faster one-step or few-step sampling. Applied to MRI synthesis tasks, the approach aims to achieve comparable or superior image quality and structural consistency relative to diffusion-based models while reducing inference time.
While the majority of the studies described above utilize conditional generative techniques, they often stop short of taking full advantage of the generative process. In most cases, for a given input image, they generate a single output image and treat this as the prediction. For example, the residual vision transformer (ResViT) model presented in35 is built based on the probabilistic formulation (GAN-based framework); however, in practice, it is used deterministically. We include this model in our comparisons as a benchmark due to the high performance of vision transformers reported in medical imaging tasks36. See Section ResViT for a brief explanation of this model. In contrast to this, in this work, we treat the generative algorithms as tools that allow us to sample from a conditional distribution and for a single input image generate an ensemble of output images. Thereafter, we use these images to generate our best guess for the output image and also quantify the pixel-wise uncertainty in this guess. We also demonstrate that a measure of this uncertainty can be used to determine whether the input image is tainted with artifacts or contains brain pathologies and is, therefore, out of distribution (OOD). We believe that the ability to detect anomalous input images without any supervised training will be an important component of a quality control process for image transformation. We also note that while a few prior works acknowledge the ability of generative models to produce multiple plausible outputs37,38, they typically do not analyze the resulting sample distributions or apply them to downstream tasks such as uncertainty quantification. Meanwhile, most existing approaches to uncertainty quantification in medical imaging rely on Bayesian inference or model ensembles39,40,41 and only a few rare studies, such as42, address uncertainty in the specific context of MR image transformation. In contrast, our work directly leverages the generative sample distribution to compute voxel-wise uncertainty and demonstrates its practical utility for detecting out-of-distribution inputs in brain MR images.
In the current study, we also address a critical gap in the current understanding of generative models employed for transforming brain MR images. We implement and test three promising models, namely the conditional GAN (cGAN)43,44,45,46, a diffusion model based on the conditional noise-conditional score network (NCSN)20,21,22, and the conditional denoising diffusion probabilistic model (DDPM)47,48, and systematically compare their performance on two benchmark datasets. Through this comparison, we note that the diffusion models yield the most accurate results, followed by a deterministic Direct model based on the U-Net architecture, cGAN model, and finally the ResViT model. We believe that these results will help guide future practitioners in selecting appropriate models for their tasks.
The remainder of the manuscript is organized as follows. In the following section, we present results for the three generative models and two benchmark models applied to brain MR image transformation tasks. Thereafter, we discuss the performance of these algorithms and explicitly demonstrate the benefits of treating the image transformation problem as a problem of conditional image generation. The benefits include more accurate results and the ability to detect OOD input images. Thereafter, we describe in detail the models implemented and tested in this study, the datasets used for training and testing the models, and the metrics used for evaluating their performance. We end the manuscript with conclusions and remarks for future work.
Results
MR image transformation for an axial slice image from the IXI dataset. Column 1: input image; column 2: target image; columns 3-4: predicted and error images for the direct model; columns 5-6: predicted and error images for the ResViT model; columns 7-9: predicted, error, and standard deviation images for cGAN; columns 10-12: predicted, error and standard deviation images for NCSN; columns: 13-15: predicted, error and standard deviation images for DDPM.
We implement a direct inference model based on a U-Net architecture (called the Direct model in this work) that solves the deterministic image transformation problem and is trained using a mean squared error loss. We also utilize a residual vision transformer (ResViT) model as a benchmark. Moreover, we implement a conditional Wasserstein GAN (cGAN), which comprises a generator and critic that are trained together using a Wasserstein loss. In addition, we implement two score-based diffusion models. These are the conditional versions of the noise-conditional score network (NCSN) model and the denoising diffusion probabilistic (DDPM) model. Both these models learn to transform samples generated from a zero-mean uncorrelated high-dimensional Gaussian distribution to samples from the desired conditional distribution of images by learning and using the score function for this transformation. They differ from each other in the variance of the Gaussian distribution. The NCSN model uses a distribution with a large variance, whereas the DDPM model uses a distribution with unit variance. All the models are trained and tested using the IXI49 (Information eXtraction from Images project) and OASIS50 (Open Access Series of Imaging Studies) datasets that include T1, T2, proton density (PD), and fluid-attenuated inversion recovery (FLAIR) images for healthy subjects and patients with Alzheimer’s disease. The probabilistic models generate multiple images for a single input image. For these models, the pixel-wise mean image obtained from these samples is used as the best guess, and the pixel-wise standard deviation image is used as a measure of uncertainty in the transformation. The performance of all models is evaluated against the ground truth by computing metrics like the structural similarity index (SSIM), peak signal-to-noise ratio (PSNR), learned perceptual image patch similarity (LPIPS)51, and boundary SSIM (BSSIM) between the ground truth and transformed images. Furthermore, the slice consistency SSIM (SCSSIM) metric is used to assess the consistency of the models in generating adjacent slices. These aspects of the work are described in detail in the Methods section.
MR image transformation for an axial slice image from the OASIS dataset. Column 1: input image; column 2: target image; columns 3-4: predicted and error images for the direct model; columns 5-6: predicted and error images for the ResViT model; columns 7-9: predicted, error, and standard deviation images for cGAN; columns 10-12: predicted, error and standard deviation images for NCSN; columns: 13-15: predicted, error and standard deviation images for DDPM.
In Figs. 1 and 2, we present typical results for all possible image transformation types within the IXI and OASIS datasets, respectively. Each dataset contains three different image types, and we transform images from one type to the other two. Consequently, in each figure, we present results that are organized in six rows based on the type of transformation. From left to right, in the first and second columns, we present the input and ground truth (labeled as target) images, respectively. In the remaining columns, we present model outputs. The third and fourth columns contain the predicted image and the error for the Direct model. Here, the error is defined as the absolute value of the pixel-wise difference between the target and the prediction. The fifth and sixth columns belong to the prediction and error of the ResViT model. The seventh to ninth columns contain the prediction, error, and pixel-wise standard deviation images for the cGAN model. Finally, columns 10-12 and 13-15 contain the corresponding results for the NCSN and DDPM models, respectively.
For each image in the test set, we evaluate the SSIM, PSNR, LPIPS, and BSSIM metrics between the predicted and target image and present the statistics in Fig. 3. For SSIM, PSNR, and BSSIM metrics, higher values indicate a closer match between the target and predicted images, while for LPIPS, lower values indicate better similarity. For SSIM and BSSIM, a value of 1, and for LPIPS, a value of 0, indicates a perfect match. Results for the IXI dataset are presented in the four left plots, while those for the OASIS dataset are presented in the right plots. Within each plot, we present results for all five models applied to six different transformation types in the form of a box plot. The rectangular extent of each box displays the interquartile range (IQR) where \(50\%\) of the values reside, and the whiskers depict the range of the values excluding the outliers. The outliers are displayed as markers in the plots and are defined as values that are smaller than 1.5 IQR below the \(25^{th}\) percentile and greater than 1.5 IQR above the \(75^{th}\) percentile.
In Table 1, we present the numerical values of the SSIM, PSNR, LPIPS, and BSSIM for test subjects from the IXI and OASIS datasets. Each row contains numerical values (the average value and the standard deviation in parentheses) of the metrics. We used pairwise one-sided Wilcoxon signed-rank tests (\(p < 0.05\)) to assess statistical significance in the reported metrics. A model is considered the best if it performs significantly better than all other models, allowing multiple models to be considered best when differences between them are not significant. The best-performing models are highlighted in bold in Table 1 (also in Tables 2, 4, and 5). For each row, we highlight the model or models with the highest value of each similarity metric. For SSIM, we observe that out of the twelve transformations, the DDPM model is the most accurate in six cases, the NCSN is the most accurate in four cases, and the Direct model is the most accurate in two cases. For the PSNR metric, the DDPM and Direct models outperform other models in ten and two cases, respectively. When focusing on the LPIPS metric, the DDPM method yields the highest values in eleven out of twelve cases, and NCSN is the best in one case. For the BSSIM metric, DDPM performs best in six cases, followed by NCSN in four, and Direct and ResViT models each leading in one case. The cGAN model never achieves the highest value for BSSIM in any transformation.
In addition to exploring the performance of MR image transformation models using inputs from intra-dataset test subjects, we discuss how different models respond to input images outside of the distribution for which they were trained. We refer to these input images as out-of-distribution (OOD) images and aim to explore how the models respond to these inputs and the possibility of detecting these inputs using the generative property of the probabilistic models. For this purpose, we utilize several datasets with OOD inputs. In the first experiment, we pass T1 images from the OASIS test subjects to the T1 to T2 transformation models trained on the IXI dataset. We compare the T2 output images against the available targets using SSIM, PSNR, LPIPS, and BSSIM metrics. We show the average results in Table 2, where each row belongs to a metric. The numbers in the parentheses indicate the changes compared to the results derived from a model trained and evaluated on the OASIS dataset. As expected, we observe that the majority of models demonstrate a decrease (indicated by a negative sign for SSIM, PSNR, and BSSIM and a positive sign for LPIPS) in their performance when transforming MR images using OOD inputs. We also observe that the highest decrease in the SSIM metric occurs in the NCSN, suggesting the possibility of a lower generalizability in these models. On the other hand, the ResViT model demonstrates the lowest changes in the SSIM and BSSIM metrics. This can stem from the fact that the ResViT model benefits from a sub-model that is pre-trained using a gigantic medical imaging dataset, facilitating higher generalizability across inputs. We further notice that LPIPS shows minimal changes compared to other metrics for most models.
In another experiment, we use sample input MR images with artifacts and brain pathologies (MS lesions and brain tumors) from three pubic datasets discussed in Datasets and Pre-processing. We show several examples of the outputs of the models using OOD inputs in Fig. 4. In this figure, from left to right, the first and second columns belong to the input and target images, where the target image is only available for images with pathology, as images with artifacts do not have T2 ground truth images. The third, fifth, and seventh columns represent the output images, and the fourth, sixth, and eighth columns show the standard deviation images, respectively, for the cGAN, NCSN, and DDPM models. Each row belongs to the input image with the indicated MR image anomaly. The first, second, and third rows correspond to images with artifacts: a strong RF zipper artifact (dashed-line shaped strong signals), an MR image reconstruction artifact, and a low SNR issue. We observe that all three models produce less reliable results when using images outside the distribution of the training dataset. We note that the Direct and ResViT models also yield inaccurate results, which are not shown here for brevity. We also observe that the probabilistic models exhibit responses with different characteristics to these inputs in their standard deviation images. We particularly see an increase in the standard deviation values, especially for the cGAN and NCSN models. While OOD inputs are more likely to result in inaccurate outputs, the information derived from the standard deviation images can be utilized to detect these potential inaccuracies, an option that is not available for deterministic models. The next row belongs to an image with a brain tumor marked as (t). As can be seen in the target column, the tumor area yields a hyperintense signal region in the target T2 image. However, the models reconstructed the tumor as a hypo-intense area in their outputs. This inaccuracy stems from the fact that the images with tumors are not included in the training data. In this example, the cGAN model does not demonstrate an increase in the standard deviation in the areas with the tumor. In contrast, the standard deviation is significantly higher in a relatively large area around the tumor for the NCSN model. The DDPM model shows an increase in the standard deviation for the areas close to the tumor. In the last row, the input image contains a white matter lesion marked as (msl). As can be seen, all three models have been relatively successful in reconstructing the lesion area and present slightly higher standard deviations almost globally.
The probabilistic models allow the evaluation of the pixel-wise standard deviation in the transformed images, which can be considered as a measure of uncertainty. It is worth considering whether this measure of uncertainty depends on whether the input image is OOD. From Fig. 4, we observe that this might be the case since the standard deviation in the transformed images for OOD input images is typically larger than the standard deviation for normal images.
Response of models to out-of-distribution (OOD) input images. Column 1: input image; column 2: target image (when available); column 3-4: output and standard deviation images for the cGAN model; columns 5-6: transformed image and standard deviation for NCSN; columns 7-8: transformed image and standard deviation for DDPM. Rows 1, 2, and 3: images with RF zipper, reconstruction artifact, and low SNR; rows 4-5: images with brain tumor and white matter MS lesion pathologies.
To explore the dependency of the standard deviation values on OOD inputs, we introduce a scalar score \(\hat{\sigma }\) calculated as the value of pixel-wise standard deviation averaged over all voxels of the brain for a given 3D image volume. The values of raw scores \(\hat{\sigma }\) are demonstrated as swarm plots in three sets shown in the first row of Fig. 5 for cGAN, NCSN, and DDPM models. In each plot, the left column (normal) contains results obtained by transforming normal images from 144 IXI test subjects and evaluating \(\hat{\sigma }\) for each subject using the T1 to T2 transformation model trained on the IXI training dataset. The left columns (OOD) contain calculated \(\hat{\sigma }\) for subjects with artifacts (shown in orange), those with MS lesions (in green), and those with brain tumors (in red), from which five examples are displayed in Fig. 4. This column also shows results for input images from the OASIS dataset, which are considered OOD, as the model is trained on the healthy subjects from the IXI dataset, but the OASIS dataset contains subjects with Alzheimer’s and those in the aging stages. From Fig. 5, we observe that a large value of \(\hat{\sigma }\) is often an indicator of OOD input images for the cGAN and NCSN models. We also observe that this conclusion can not be generalized to \(\hat{\sigma }\) values calculated for the DDPM model.
Distribution of \(\hat{\sigma }\), a measure calculated from the standard deviation images, for cGAN, NCSN, and DDPM models. Distributions are divided into left: normal input images and right: out-of-distribution (OOD) or images with pathology or artifact, as indicated with different colors in the legend. The first and second rows show the score values before and after calibration using the isotonic regression method.
These are further examined by using \(\hat{\sigma }\) as a score to classify the input images as either normal or OOD. We do this by varying the threshold for this classification to generate a receiver operating characteristic (ROC) curve and computing the area under this curve (AUC). This is shown in the left part of Fig. 6 for \(\hat{\sigma }\) values. Recall that for a perfect classifier, AUC is equal to 1, whereas an AUC of 0.5 corresponds to performance no better than a random guess. As shown in the plot legends, the AUC for the cGAN, NCSN, and DDPM models are respectively 0.976, 0.850, and 0.421. These results indicate that the scores calculated from the outputs of the cGAN and NCSN models are effective in classifying input images as normal or OOD. The right plot in Fig. 6 presents the F1 scores computed to evaluate the classification performance based on the normalized uncertainty values \(\hat{\sigma }\) for the cGAN, NCSN, and DDPM models, where the maximum achievable F1 scores, as indicated in the legend, are 0.905, 0.819, and 0.395. The F1 score serves as a balanced metric that combines both precision and recall (See Evaluation metrics and statistical measures for the definition).
Left: ROC curves comparing the classification performance of cGAN, NCSN, and DDPM models based on the raw calculated average standard deviation \(\hat{\sigma }\) for detecting normal from OOD images. The closer the curve is to the top-left corner, the better the model. Right: F1 score curve for different threshold values. A high F1 score indicates that the model has both high precision and high recall, meaning it makes few false positives and false negatives.
Despite the strong discriminative performance of \(\hat{\sigma }\) scores in separating normal and OOD inputs, further analyses revealed that these values were not well-calibrated for the binary classification task. This was reflected in higher error values for the expected calibration error (ECE), Brier score, and negative log-likelihood (NLL) shown in the first three columns of Table 3 (see Evaluation metrics and statistical measures for definitions). We applied isotonic regression calibration to the scalar OOD scores \(\hat{\sigma }\) produced by the cGAN, NCSN, and DDPM models. Calibration was performed on the validation set, using five-fold cross-validation on the in-distribution (ID) data combined with a randomly selected subset of OOD samples to ensure balanced coverage of both classes. This procedure learns a monotonic mapping that adjusts raw uncertainty estimates (\(\hat{\sigma }\)) so that their magnitudes better reflect empirical probabilities of belonging to the OOD class. The resulting calibration models were subsequently applied to rescale the OOD scores of the unseen test data, ensuring that all reported reliability metrics correspond to held-out samples. As shown in the second part of Table 3, the calibrated scores, denoted as (\(\bar{\sigma }\)), yielded substantially lower ECE, NLL, and Brier score values, with only a slight reduction in AUC. The calibrated \(\bar{\sigma }\) distributions, shown in the second row of Fig. 5, demonstrate probabilistically meaningful separations between in-distribution and OOD inputs, making calibrated scores more generalizable and interpretable.
Discussion
In the first part of this section, we investigate the discussed models in terms of their quantitative accuracy and the qualitative characteristics of their outputs. In the second part, we briefly provide examples of the neuroimaging implications and considerations that need to be taken into account when utilizing these models.
Images from the IXI dataset are of better quality and display less heterogeneity when compared with images from the OASIS dataset. Some images in the OASIS dataset suffer from low contrast and poor image quality. Consequently, on average, all image transformation methods perform better on the IXI dataset. This is seen clearly by computing the average SSIM, PSNR, LPIPS, and BSSIM values for each method over all six transformation types for the two datasets. These values are reported in Table 4, where we observe that for every method, all metrics are higher for the IXI dataset.
We note that while the outputs of the models are generated as batches of 2D slices, they are then stacked to reconstruct 3D volumes. All evaluation metrics (SSIM, PSNR, LPIPS, and BSSIM) in this work are computed based on normalized 3D volumes rather than slice-wise averages to ensure that the evaluation metrics reflect any potential inconsistencies across slices. To further assess volumetric coherence, we introduced a slice-consistency metric, SCSSIM, that measures the stability of inter-slice similarity across the reconstructed volume (see Evaluation metrics and statistical measures section for definition). The computed values are presented in Table 5. As shown, all methods achieved high slice-consistency scores (ranging from 97.20 to 99.25), indicating that the models preserve spatial continuity across slices.
From Table 4, we conclude that the cGAN model yields the lowest value for the PSNR, LPIPS, and BSSIM, and the second lowest for the SSIM metric. We speculate that the cause is the distortion observed at the boundaries of different types of brain tissues, as also reflected in low BSSIM values. These distortions are most clearly seen in the transformed T1 images. We show an example in Fig. 7, where for a test sample slice from the IXI dataset, we have plotted the target T1 image, a zoomed-in region of interest (ROI) in this image that encompasses the ventricles and basal ganglia, the corresponding transformed ROIs generated by the Direct, ResViT, cGAN, NCSN, and DDPM methods, and the ROI in the input T2 image. We will return to Fig. 7 in this section in order to illustrate qualitative trends observed in our study. As mentioned, in the cGAN output ROI of Fig. 7, we observe that the boundaries between subcortical structures like the thalamus and putamen (marked as (t) and (p) in the target image) and the white matter appear to be distorted in the transformed cGAN image. We note that these distortions are also observed in other images generated by the cGAN method. Our experience is consistent with other reports17 that claim that GAN-based models struggle to generate fine features and edges in transformed images. This is a significant drawback since it can lead to erroneous volumetric measurements in downstream tasks in MR imaging.
T2 to T1 transformation from the IXI test dataset. From left to right: target T1 image with a square-shaped region of interest (ROI), followed by images of the ROI for the target image, transformed images generated by the Direct, ResViT, cGAN, DDPM, and NCSN models, and the input image. In the ROI target image, the thalamus (t) and putamen (p), and in the DDPM image, the splenium of the corpus callosum (scc) and two hypo-intense pixels are marked by arrows.
The model with the lowest SSIM and slightly higher PSNR and BSSIM is ResViT. We observe from Fig. 7 that, unlike cGAN, this model is successful in generating fine features and tissue boundaries. This can also be observed as a higher value of BSSIM for this model compared to the cGAN model. However, a closer look at atypical slices reveals that the ResViT model occasionally produces areas of false positives (cloud shapes outside the brain) and false negatives (missing part of the brain in the output). We note that the former is resolvable using appropriate post-processing. However, the latter is unrecoverable as part of the brain will be excluded from the output. We speculate that in patch-based hybrid models such as ResViT, small tissue islands can fall across patch boundaries, fragmenting their representation and making them prone to suppression during reconstruction. Improving spatial resolution, using overlapping or hierarchical patch embeddings52,53, and proper preprocessing, such as applying skull stripping after transformation, can mitigate this effect. We provide an example for this issue in Fig. 8, where the lower parts of the temporal lobes, where the white arrows are pointed in the input image, are missing in the output of the ResViT model.
The calculated metrics for the Direct model are relatively acceptable. However, visual inspection of the outputs suggests that this method generally produces transformed images that are blurry. From Fig. 7, we can clearly see that the ROI generated by the Direct method is blurry when compared with the target ROI and the ROIs generated by most other methods. We observe this trend for the Direct method for most transformation types and test samples. One possible explanation for this trend could be the use of the mean-squared error as a loss function. However, we found that this trend persisted when we changed the loss function to the mean absolute error function (results not shown here).
The DDPM model has the highest overall SSIM, PSNR, LPIPS, and BSSIM values, whereas the NCSN model yields the second-highest SSIM and BSSIM values. A visual inspection of the transformed images produced by these models in Fig. 7 reveals that both generate qualitatively accurate images. However, a close look at the image generated by the DDPM method indicates the presence of two hypo-intense pixels posterior to the splenium of the corpus callosum (marked as “scc” in the figure). We find these artifacts to be an occasional occurrence in the images generated by the DDPM method. As discussed in Supplementary Notes online, we speculate that this model exhibits a numerical instability for the MR image transformation problem that results in producing outlier pixels with extremely large or small (negative) values. The outlier values can be corrected by employing a denoising step described in the Methods section. However, extra care is needed when working with this model. We note that the similarity metrics for the DDPM method reported in Tables 1 and 4 are obtained after applying this denoising step.
In summary, considering SSIM as the primary quantitative metric, the DDPM model shows higher accuracy compared to NCSN (by \(1.73\%\)) as the second-best model. However, both models generate transformed images with comparable quality. Additionally, the former exhibits numerical instability, whereas the latter does not.
As we mentioned, the probabilistic formulation in the generative models allows a series of additional analyses that are not possible when working with deterministic models. Specifically, for any given pixel in the input images, we have a distribution of output values in the generated samples and a single ground truth value. Performing various analyses on univariate pixel-wise data can provide insights into the characteristics of these models and the distribution of their outputs. One of these aspects is whether the distribution of the outputs aligns with the observed target data. One way to assess this is through regression calibration plots54, which visualize the relationship between predicted confidence levels and the empirical coverage, i.e., the actual fraction of targets whose true values fall within the corresponding predicted confidence intervals. For example, whether 10% of the target values lie within the 10% confidence intervals calculated based on the samples generated by the model, and so on.
We illustrate these regression calibration plots in Fig. 9 for the T1-to-T2 transformation of generative models on the IXI (left) and OASIS (right) datasets. A well-calibrated model produces a curve that closely follows the diagonal. Deviations below the diagonal indicate overconfident predictions (distributions of samples are too narrow), while curves above suggest underconfident predictions (distributions too wide). We note that the term “calibration” was also used in the context of binary classification of normal and OOD inputs from the standard deviation values. In the context of binary classification, the targets are labels with values of zero or one, and the scores form a distribution of values for the probability of belonging to the normal or OOD class. Here, in the context of regression calibration, the targets and outputs have continuous values. As shown, the orange curve, especially for the IXI dataset, is closest to the diagonal, indicating that the NCSN model is better calibrated compared to others. In contrast, the DDPM model tends to be underconfident, with true pixel values too often falling within the predicted intervals of samples. This suggests that the distributions of DDPM outputs are overly broad, reflecting standard deviations that are larger than what they should be. On the other hand, the cGAN model exhibits calibration curves that often lie below the diagonal, indicating overconfidence. This may arise from limited diversity in the generated outputs or from systematic bias in the predictions. We observed similar trends in transformations other than T1 to T2, which are not reported here for brevity.
The effectiveness of OOD detection based on the scores calculated from the standard deviation images is linked to model calibration. While NCSN uncertainty is well-calibrated and reliably captures ID variability, cGAN exhibits overconfident predictions for ID inputs but displays larger standard deviation image scores for OOD inputs. This behavior may arise from the adversarial training dynamics in cGAN, which amplify differences when encountering unfamiliar inputs. In contrast, DDPM produces broadly high uncertainty for all inputs, reducing the discriminative power of the scores for OOD detection. These observations highlight that both calibration quality and sensitivity to deviations determine how ensemble-derived uncertainties can be used to detect out-of-distribution images.
We provide a number of additional analyses based on the generative property of the cGAN, NCSN, and DDPM models in Supplementary Notes online. Specifically, we utilize the Shapiro-Wilk test to assess the normality of the distribution of the outputs. The results suggest that the outputs of the NCSN model demonstrate significantly higher normality compared to the other two models. We further investigate the impact of the number of samples on the accuracy of the outputs and the statistical measures, such as cumulants, in Supplementary Notes online.
In addition to quantitative evaluation and exploring the differences in the characteristics of the discussed models, recognizing the extent of their reliability is critical, especially in clinical contexts. For example, are models equally successful in all combinations of input and output contrasts, or are pathological features such as lesions preserved in the output images? To answer the first question, we note the differences when analyzing the metrics across all transformations in Table 1. For example, the DDPM model attains the two highest SSIM metrics when converting PD to T2 (97.91) and T2 to PD (97.13) images from the IXI dataset. This can be attributed to the inherent similarity in tissue contrast between these two MR image types, particularly in water-rich tissues. On the other hand, we observe the lowest SSIM when converting T2 to FLAIR (89.06) and FLAIR to T2 (91.50) images. One of the reasons for this drop in performance is the absence of certain features in the input image that are clearly observed in the target image. For example, when compared with the T2 images, FLAIR images are designed to demonstrate high sensitivity in highlighting white matter lesions and some other normal or pathological regions that appear as hyper-intense periventricular voxels. This is seen in Fig. 10, where in the target FLAIR image, we have marked several hyper-intense periventricular regions (pv) and white matter lesions (wm) that are very difficult to locate in the input T2 image. The transformed FLAIR image generated by the DDPM model is relatively successful in reconstructing periventricular hyper-intense regions, but fails to generate white matter lesions, as they are absent in the input image. We observe similar patterns in the output of other models.
We further conduct an experiment to investigate the ability of the models in preserving pathological features. We perform T1 to T2 image transformation using inputs that contain MS lesions from the MSSEG dataset. We illustrate an example of the outputs in Fig. 11. From left to right, the first and second images show the input image and the area within the highlighted ROI in the input image. The last and second-to-last images are the target images and the target image within the ROI. The remaining images (third to eighth) are the outputs of the models. The red curves indicate the MS lesion areas provided through manual segmentation by experts. As seen, most models have been relatively successful in reconstructing the lesions in their output. We quantify the ability of the models to preserve the lesions by calculating the Pearson correlation coefficient between the intensity of pixels within the lesion areas of the outputs and the target images. The results for the Direct, ResViT, cGAN, NCSN, and DDPM are respectively 0.748, 0.696, 0.669, 0.705, and 0.711. These numbers indicate a relatively strong correlation between the values of the output and target images within the lesion areas. This experiment presents an example of the successful preservation of pathologies in the output.
Another example is the effectiveness of the models in performing several image transformations on the OASIS test subjects that exhibit a range of pathologies, such as brain atrophy and Alzheimer’s disease lesions. However, we note that this is not always the case. We provided an example in Fig. 4, where the models fail to regenerate the brain tumor in their output. Therefore, beyond the extensive evaluations provided here, the selection of training data, the expectations from the outputs, and the evaluation strategies should be tailored based on the subsequent tasks. These are widely accepted points for any deep learning-based model in order to provide reliable outputs. However, the generative models hold the significance of providing tools that can warn the end-user when the outputs are more likely to be unreliable.
The results of T1 to T2 transformation of the discussed models for an input image with MS lesions. From left to right, the input image and the area within the highlighted ROI in the input image, the outputs of the indicated models, the target image within the ROI, and the target image are shown. The red curves indicate the manually segmented MS lesion areas.
Methods
Problem formulation
We assume that a dataset of N paired MR images of an input and an output type is available. This dataset is denoted by \(S=\{( \varvec{x}^{(i)}, \varvec{y}^{(i)})\}_{i=1}^N\), where \(\varvec{x}^{(i)} \in \mathbb {R}^{N_X}\) are images of the input type, \(\varvec{y}^{(i)} \in \mathbb {R}^{N_Y}\) are images of the output type, and \(N_X\) and \(N_Y\) are the number of pixels in the output and input images, respectively. In this study, \(N_X=N_Y\) and each image is a two-dimensional axial image of the brain. It is further assumed that each image pair corresponds to the same physical slice and belongs to the same subject.
Deterministic formulation (direct model)
In the deterministic formulation of the problem, given an input image, the aim is to solve a nonlinear regression problem to estimate the pixel intensity of the output image. To solve this problem, we assume that there exists a function, \(\varvec{f}:\mathbb {R}^{N_X} \longmapsto \mathbb {R}^{N_Y}\), that directly transforms the pixel intensity of the input image to the output image through a nonlinear transformation. We model \(\varvec{f}\) using a neural network and denote it by \(\tilde{\varvec{f}}\). This model is trained by minimizing a loss function that measures the difference between the output of the model and the target or ground truth for all images in the training dataset. This loss function is given by
The trained model is \(\tilde{\varvec{f}}^*=\mathop {\mathrm {arg\,min}}\limits _{\tilde{\varvec{f}}}\, \mathscr {L}(\tilde{\varvec{f}})\) and is referred as the direct model. We use this model as a simple baseline and compare the performance of the probabilistic models against it. In order to make a fair comparison, a similar U-Net neural network architecture is used for all models, i.e., the direct model, the generator in the cGAN, the score network in the NCSN, and the denoiser in the DDPM. (see Supplementary Notes online for details.)
Probabilistic formulation
In the probabilistic framework, we treat the input and output image types as random variables \(\varvec{X}\) and \(\varvec{Y}\), respectively. The dataset S represents samples drawn from the joint density function \(p_{\varvec{XY}}( \varvec{x}, \varvec{y})\). Using S, we aim to train deep generative models that receive a new input image \(\varvec{X} = \hat{\varvec{x}}\) and can generate an ensemble of likely transformed generated images sampled from the conditional distribution, \(\varvec{y}^g \sim p_{\varvec{Y}|\varvec{X}}(\varvec{y}|\hat{\varvec{x}})\). These generated samples can be used to compute any desired statistic, including the sample mean, which is treated as the best guess for the inferred transformed image, and the sample standard deviation image, which is interpreted as a measure of uncertainty in the transformed image. We solve the probabilistic problem using three distinct generative models: conditional Wasserstein generative adversarial networks (cGAN), noise conditional score networks (NCSN), and denoising diffusion probabilistic models (DDPM) methods. These are described next.
Conditional Wasserstein generative adversarial network (cGAN)
As shown in Fig. 12, the cGAN architecture consists of two deep neural networks, a generator \(\varvec{g}\) and a critic or discriminator d. The generator, \(\varvec{g}:\mathbb {R}^{N_X} \times \mathbb {R}^{N_Z} \longmapsto \mathbb {R}^{N_Y}\), receives an input image \(\varvec{x}\) and an instance of the random latent vector \(\varvec{z} \in \mathbb {R}^{N_Z}\) and generates an output image \(\varvec{y}^g = \varvec{g}( \varvec{x}, \varvec{z})\). The latent variable is sampled from a simple, known probability distribution, \(p_{\varvec{Z}}\). In this work, each component of the latent vector is independent and sampled from a standard normal distribution. By using a given input image \(\varvec{x}\) along with multiple instances of \(\varvec{z}\) drawn from \(p_{\varvec{Z}}\), the generator generates an ensemble of transformed images for a single input image. These images may be treated as samples drawn from a conditional distribution denoted by \(\varvec{y}^g \sim p^g_{\varvec{Y}|\varvec{X}}\).
The conditional Wasserstein generative adversarial network (cGAN) architecture used in this work. The generator \(\varvec{g}\) receives the input image \(\varvec{x}\), and generates a synthesized image \(\varvec{y}^{\varvec{g}}\) for any random latent vector \(\varvec{z}\). The critic d distinguishes between the generated synthesized image \(\varvec{y}^{\varvec{g}}\) and target image \(\varvec{y}\) paired with \(\varvec{x}\). The training process involves updating \(\varvec{g}\) such that its outputs are indistinguishable from real images by d and updating d to become better at distinguishing the generated images from real ones.
The critic \(d:\mathbb {R}^{N_X} \times \mathbb {R}^{N_Y} \longmapsto \mathbb {R}\) receives a pair of images and distinguishes between the pairs drawn from the true dataset, i.e., \(( \varvec{x}, \varvec{y}) \sim p_{\varvec{XY}}\), and pairs where the output image type is generated by the generator network, i.e., \(( \varvec{x}, \varvec{y}^{\varvec{g}})\), where \(\varvec{y}^{\varvec{g}} \sim p_{\varvec{Y}|\varvec{X}}^{\varvec{g}}\). This is accomplished by requiring the critic to yield larger values for images from the true dataset and smaller values for the images from the generated dataset. This leads to an adversarial loss function44 given by
and determining the optimal networks by solving a min-max optimization problem,
We note that the second term in Equation 3 is the gradient penalty term45, \(\lambda\) is the penalty parameter, and \(\hat{\mathfrak { \varvec{y}}} =\epsilon \varvec{y} +(1-\epsilon ) \varvec{y}^{\varvec{g}}\) where \(\epsilon \sim \mathscr {U}(0,1)\). It can be shown that the optimal critic (\(d^*\)) measures the Wasserstein-1 distance between the true joint distribution and its generated counterpart46, and therefore the optimal generator (\(\varvec{g}^*\)) aims to minimize this distance. This, in turn, implies that the generated conditional distribution \(p_{\varvec{Y}|\varvec{X}}^g\) is close to the true conditional distribution \(p_{\varvec{Y}|\varvec{X}}\).
We further note that the generator implemented in this work accepts \(\varvec{z}\) as input through a set of conditional instance normalization (CIN)55 blocks, as discussed in Supplementary Notes online. This is an example of the neural network considerations to incorporate stochasticity in generative models. In other words, a model can be constructed based on a probabilistic formulation while still lacking the appropriate design to generate multiple samples per input image. These models provide predictions in a deterministic fashion, a common practice in the literature, and therefore, lose any information derived from the distribution of generated images. See the ResViT model discussed next as an example.
Residual vision transformation (ResViT)
The ResViT35 model also uses a generative adversarial network architecture, where the generator consists of a hybrid architecture of encoder, information bottleneck, and decoder. The encoder and decoder contain deep convolutional layers, and the information bottleneck consists of several stacked residual transformer (ART) blocks. Each ART block contains a transformer module, a CNN module, and external skip connections. The discriminator model contains several convolutional layers. When converting a single MR image type to another, the original implementation is designed to provide a single output (See Supplementary Notes online for details).
Diffusion models
As illustrated in Fig. 13, the workflow of diffusion-based models (NCSN and DDPM) consists of a forward noising process (from right to left) and a backward denoising process (from left to right). In the unconditional version of a diffusion model, during the noising process, a sample is drawn from the true initial distribution \(\varvec{y}=\varvec{y}_0 \sim p_0(\varvec{y})\), i.e., an image from the training dataset, and an uncorrelated noise with variance \(\beta _t=\sigma ^2_t,t = 1 \dots T\) is added to it. This process is repeated until the image looks like pure noise, i.e., \(\varvec{y}_T \sim \mathscr {N}\left( \varvec{\mu }, \beta _T\mathbb {\varvec{I}}\right)\), where \(\mathbb {\varvec{I}}\in \mathbb {R}^{N_Y}\) denotes the identity matrix. The data from the forward noising process is used to train a denoising neural network that receives a noisy image \(\varvec{y}_t \sim p_{t}(\varvec{y})\) and generates an image \(\varvec{y}_{t-1} \sim p_{t-1}(\varvec{y})\) with slightly less noise. This network is applied to an image drawn from the distribution of pure noise \({\varvec{y}}_T \sim \mathscr {N}\left( \varvec{\mu }, \beta _T \mathbb {\varvec{I}} \right)\) in multiple denoising steps, \(t = T \dots 1\), to generate a sample from \(p_{0}(\varvec{y})\). In the conditional version of this algorithm, the noising and denoising steps are conditioned on the input image \(\varvec{x}\).
Forward noising and backward denoising processes in diffusion models. During training (from right to left) sample images go through several noising steps. This provides a series of images with different levels of noise. These images are used to train a model that receives a noisy image and time t and provides an output that is used to generate a less noisy image. After training, generating samples involves (from left to right) starting from a noise image and generating less noisy images, leading to the final generated image.
Noise conditional score network (NCSN)
The noising process in the NCSN method involves adding a noise to the initial image \(\varvec{y}_{0}\),
where \(\varvec{\epsilon } \sim \mathscr {N}(0, \mathbb {\varvec{I}}) \in \mathbb {R}^{N_Y}\), and the variances \(\beta _t\) are hyper-parameters that are selected according to a schedule21. The training process in the NCSN learns a neural network model \(\tilde{\varvec{s}}(\varvec{y},\varvec{x},t): \mathbb {R}^{N_Y} \times \mathbb {R}^{N_X} \times \mathbb {R} \longmapsto \mathbb {R}^{N_Y}\), that approximates the score function \(\nabla \log (p_{\varvec{Y}|\varvec{X}}(\varvec{y}|\varvec{x},t))\) of the time dependent conditional distribution. This is achieved by defining an objective function
where the optimal score network model is given by \(\tilde{\varvec{s}}^*= \mathop {\mathrm {arg\,min}}\limits _{\tilde{\varvec{s}}}\, \mathscr {L}(\tilde{\varvec{s}})\). The term within the square parentheses in Equation 6 is similar in form to a regression loss. Therefore, we recognize that the output of a fully trained score network is given by \(\tilde{\varvec{s}}^*(\varvec{y}_{t},\varvec{x},t) \approx -({\varvec{y}_{t}-\varvec{y}_0})/{\beta _t}\). By rearranging this equation, we note that \(\varvec{y}_0 \approx {\beta _t}\tilde{\varvec{s}}^*(\varvec{y}_{t},\varvec{x},t) + \varvec{y}_{t}\). That is, the value of the score function, when multiplied by the variance \(\beta _t\), is what should be added to the perturbed image \(\varvec{y}_{t}\) to generate the image without noise (\(\varvec{y}_0\)) at any step t. In practice, the denoising process in NCSN is performed iteratively through T steps of Langevin MC21, where at any step, a less noisy image \(\varvec{y}_{t-1}\) is generated using
Here \(\gamma _t = \lambda _\gamma \beta _t / \beta _{1}\) is the step size, \(\lambda _\gamma\) is a hyper-parameter, and \(\beta _1\) is the variance for the first time step (typically the smallest \(\beta _t\)). The sampling process begins by selecting \(\varvec{y}_{T} \sim \mathscr {N}(0, \beta _T \mathbb {\varvec{I}})\) and a given input image \(\varvec{x}\), and using these in the equation above. This results in \(\varvec{y}_{0}\), which is a single sample of the transformed image generated conditioned on \(\varvec{x}\). This process is repeated to generate multiple samples.
Denoising diffusion probabilistic models (DDPM)
The noising process in the DDPM method also involves adding a tailored noise to \(\varvec{y}_{0}\),
where \(\alpha _t = 1 - \beta _t\) and \(\bar{\alpha }_t = \prod _{s=1}^t\alpha _s\)56 and \(\varvec{\epsilon } \sim \mathscr {N}(0, \mathbb {\varvec{I}})\). In the DDPM method, unlike NCSN, the variances are limited to be between zero and one as \(\beta _t \in (0, 1), t=1 \dots T\), such that \(\beta _1< \beta _2< \dots < \beta _T\). The noisy image \(\varvec{y}_t\) generated through Equation 8 and its counterpart from the initial distribution \(\varvec{y}_0\) are used to train a denoiser function in the form of a neural network \(\tilde{\varvec{\eta }}: \mathbb {R}^{N_Y} \times \mathbb {R}^{N_X} \times \mathbb {R} \longmapsto \mathbb {R}^{N_Y}\). This network is trained using the loss function56,
where the optimal denoiser is given by \(\tilde{\varvec{\eta }}^*= \mathop {\mathrm {arg\,min}}\limits _{\tilde{\varvec{\eta }}}\, \mathscr {L}(\tilde{\varvec{\eta }})\). Similar to the NCSN, the loss function in Equation 9 resembles a regression loss. Further, by comparing the term within the bracket in Equation 9 and the noising process (Equation 8), we observe that the denoiser learns to estimate a measure of noise added to the clear image; that is, \(\tilde{\varvec{\eta }}^*(\varvec{y}_t, \varvec{x},t) \approx \varvec{\epsilon }\).
The DDPM model used in this study48 is an extension of an earlier model56. The training of this model involves minimization of a hybrid loss that contains an additional term so that the model also learns the sequence of noise scales \(\tilde{\beta } (t)\) for \(t=1 \dots T\), instead of accepting them as parameters defined by the user. This network is denoted by \(\tilde{\beta } (t)\), and it is trained concurrently with \(\tilde{\varvec{\eta }}\). The denoising process in the DDPM method is given by
The sampling process begins by sampling \(\varvec{y}_{T} \sim \mathscr {N}(0, \mathbb {\varvec{I}})\), selecting an input image \(\varvec{x}\), and using these in Equation 10. These iterations end with the \(\varvec{y}_0\), which is a sample of the transformed image conditioned on the input image \(\varvec{x}\).
The NCSN and DDPM methods share many common features, including beginning with a sample drawn from a simple Gaussian distribution, iteratively denoising this sample using Langevin dynamics, and using a neural network that is trained using a simple regression loss. The key difference between the two methods is in the initial distribution from which samples are generated. In the NCSN method, this is a multivariate Gaussian distribution with zero mean and a very large variance, whereas in the DDPM method, it is a multivariate Gaussian distribution with zero mean but with a variance of unity. The reader is referred to57 for a unified description of these methods and their connections with the theory of stochastic differential equations.
Evaluation of mean and standard deviation images
For all three generative models (cGAN, NCSN, and DDPM), the inference process begins by using a 2D axial image of the brain \(\varvec{x}\) as input. Thereafter, in the cGAN model, n samples of the latent vectors, \(\varvec{z}\) with dimension \(N_{Z}=128\) are used as input to the generator, which generates n images \(\varvec{y}^{(i)}, i = 1, \cdots , n\) from the desired conditional distribution. For the diffusion models, the process is similar; however, the dimension of \(\varvec{z}\) is the same as that of \(\varvec{y}\) (\(128^2\)). These samples are treated as the initial condition in the denoising process, which terminates in the generated images \(\varvec{y}^{(i)}, i = 1, \cdots , n\). We note that the sampling process in the diffusion models is computationally expensive since it involves \(T = 1000\) steps. We also note that we used \(n=20\) as the number of samples throughout this work. In Supplementary Notes online, we briefly discuss the impact of this number on the distributional characteristics and accuracy of the outputs.
All three generative models generate an ensemble of images from the conditional distribution, which is used to compute a single pixel-wise mean image, \(\bar{\varvec{y}}= \frac{1}{n} \sum _{i=1}^{n} \varvec{y}^{(i)}\) and a pixel-wise standard deviation image \(\varvec{y}' = (\frac{1}{n} \sum _{i=1}^{n}{( \varvec{y}^{(i)}-\bar{\varvec{y}})^2})^{{1}/{2}}\), where the operations are interpreted as being pixel-wise.
Transformed images from the DDPM model demonstrate isolated points with hyper or hypo-intense voxel intensities following the process described above. These voxels are identified by looking for outliers in the pixel-wise standard deviation image. They are eliminated by replacing the mean intensity value with its pixel-wise median. We note that the generative capability of the DDPM model allows us to compute the pixel-wise mean, median, and standard deviation values and, therefore, eliminate these outliers (see Supplementary Notes online for a brief note).
Evaluation metrics and statistical measures
Here, we briefly discuss the metrics used to measure the similarity between individual output images and their counterparts from the test dataset. We also briefly define the statistical measurements and metrics related to binary classification, uncertainty, and distributional behavior of the models discussed throughout this work.
Image to image similarity metrics
1) Structural similarity index measure (SSIM) is an image similarity metric that is used in this work to assess the similarity between the outputs of the models and the target images in the form of 3D volumes. SSIM measures the similarity based on local patterns of intensity and incorporates luminance, contrast, and structural information. SSIM values range from zero to one, where a SSIM value of one is indicative of a perfect match. In this work, calculated SSIM values are reported as a percentage. 2) Peak signal-to-noise ratio (PSNR) quantifies the fidelity of a reconstructed image by comparing the maximum signal intensity to the magnitude of the reconstruction error. While higher PSNR typically indicates better image quality, it does not always correlate well with perceived visual similarity58. 3) Learned perceptual image patch similarity (LPIPS)51 computes the distance between deep feature representations extracted from a pre-trained deep neural network (AlexNet in this work). The LPIPS distance between two arbitrary images \(\varvec{y}^{(1)}\) and \(\varvec{y}^{(2)}\) is computed as
where \(\phi _l(\varvec{y}^{(i)})_{h,w}\) is the latent representation of image \(\varvec{y}^{(i)}\) at location (h, w) in layer l of the pre-trained neural network. \(w_l\) are weights associate to layer l for calculating LPIPS, and \(\odot\) denotes element-wise multiplication. We note that LPIPS provides a measure of error (not similarity); therefore, lower values of LPIPS indicate better perceptual similarity between \(\varvec{x}\) and \(\varvec{y}\), with the value of zero indicating a perfect match. 4) The Boundary SSIM (BSSIM) metric is a variant of SSIM that measures similarity between edge maps of images rather than raw intensities of the whole images. By applying SSIM to edges extracted via the Sobel filter, BSSIM emphasizes structural fidelity at region boundaries, highlighting successful tissue reconstruction during transformation. The SCSSIM metric between two arbitrary 3D volumes \(\varvec{Y}^{(1)}\) and \(\varvec{Y}^{(2)}\) is defined as
where \(\varvec{Y}^{(1)}_i\) and \(\varvec{Y}^{(1)}_i\) denote the i-th out of N axial slices of the generated and target volumes, respectively. Any sudden changes between consecutive slices, captured by SSIM as a local structural deviation, result in a lower score, reflecting reduced volumetric smoothness. A score of one indicates a perfect match in inter-slice differences between the generated and target images. We also computed the Fréchet Inception Distance (FID) using several pre-trained feature extractors. However, the results did not exhibit consistent trends across models or modalities, in line with prior observations59. We speculate that FID quantifies distribution-level similarity and is more suitable for unconditional image generation. Therefore, we focused on LPIPS, which more directly reflects perceptual correspondence between paired MR images.
Classification evaluation
1) The F1 score is employed to evaluate the binary classification performance for distinguishing normal inputs from OOD inputs. It is defined as the harmonic mean of precision and recall, providing a single metric that balances false positives and false negatives. In this work, the predicted uncertainty values \(\hat{\sigma }\) are converted to binary labels \(\hat{y}_i\) using a threshold, and then compared against the true labels \(y_i \in \{0,1\}\) (normal or OOD) to determine the counts of true positives (TP), false positives (FP), and false negatives (FN). The F1 score is then computed as
A higher F1 score indicates better balance between correctly detecting OOD samples and minimizing false alarms, reflecting more reliable classification performance. 2) In binary classification, the Brier score quantifies the mean squared error between the predicted probabilities \(\hat{\sigma }_i\) and the true binary labels \(y_i \in \{0,1\}\), without converting predictions to discrete labels. It jointly assesses both the accuracy and calibration of probabilistic predictions, where lower Brier scores correspond to better predictive performance. Formally, the Brier score is defined as:
where N is the total number of samples. 3) In the context of binary classification, the Negative Log-Likelihood (NLL), also known as log loss, is a measure of how well a classifier predicts the true labels. It penalizes confident but incorrect predictions more heavily and is defined as:
where \(y_i\) is the true label and \(\hat{\sigma }_i\) is the predicted probability for class OOD. 4) Expected calibration error (ECE) measures the discrepancy between predicted confidence and empirical accuracy. Given M confidence bins, ECE is computed as
where \(n_m\) is the number of samples in bin m, and \(\text {acc}(m)\) and \(\text {conf}(m)\) are the empirical accuracy and average confidence in that bin, respectively. For example, we expect to observe \(10\%\) of values within a \(10\%\) confidence interval of the distribution of the outputs in a well-calibrated model. ECE provides a measure for calibration error, and a calibration method aims to reduce this error by updating the scores that better represent the true likelihood of outcomes60
Distributional measures
1) The Shapiro-Wilk test is a statistical normality check that evaluates whether a set of samples is drawn from a normal distribution. In the Shapiro–Wilk test, the null hypothesis is that the data are drawn from a normal distribution. Therefore, for samples that yield a p-value greater than the chosen significance level (\(p > 0.05\) in this work), we fail to reject the null hypothesis. This means there is no statistically significant evidence to conclude that the samples deviate from normality. 2) Cumulants are statistical descriptors that characterize a probability distribution. They are derived from the cumulant generating function
The first four cumulants are: the mean \(\kappa _1 = \mathbb {E}[X]=\mu\), the variance \(\kappa _2 = \text {Var}(X)=\sigma ^2\), the third-order cumulant \(\kappa _3 = \mathbb {E}[(X - \mu )^3]\) (skewness), and the fourth-order cumulant \(\kappa _4 = \mathbb {E}[(X - \mu )^4] - 3\sigma ^4\). We use the Shapiro-Wilk test and cumulants to analyze the distribution of the outputs in generative models
Datasets and pre-processing
We use two public datasets with multiple image types, IXI49 and OASIS50, for training and evaluation of the proposed methods. The IXI dataset consists of 600 MR images of three types: T1-weighted (T1), T2-weighted (T2), and proton density (PD). The scans are collected in three different hospitals in London from normal healthy subjects. Of these, 575 image sets passed QC (those with missing images and realignment issues were removed), from which 357, 74, and 144 images were used for the training, validation, and testing of the models. We balanced our datasets to have the appropriate number of scans from each hospital in the training, validation, and test datasets.
OASIS-3 is a public longitudinal dataset for normal aging and Alzheimer’s disease, consisting of T1, T2, and fluid-attenuated inversion recovery (FLAIR) images. Within this dataset, we used images from the participants that contain all image types scanned in the same MRI session. Out of 243 images, 156, 31, and 56 are used for training, validation, and testing, respectively.
For images in the IXI dataset, we use the FLIRT module of FSL61 to realign the 3D T1 images to the spaces of the T2 images, as they have the narrowest axial field of view. This is done by first using T1 images as the reference, as they have the highest resolution, and then using the inverse transformation to provide T1 images in T2 space. For the OASIS dataset, we align T2 and FLAIR images to a reference T1 image. Thereafter, we extract brains from all images by using the T1 image in a brain extraction algorithm62 to generate a mask that is applied to all images. Finally, we perform min-max normalization for all images in both datasets.
In addition to the above datasets used for training and evaluating models, we use a set of sample images from three publicly available datasets as out-of-distribution (OOD) inputs. The first set contains seven images with common artifacts from the FSL example box of MR images63. This set includes images with a strong RF zipper, two images with motion artifact, significant rotation, reconstruction artifacts, low SNR, and a strong bias field. These are T1 images without any artifact-free or target images from other contrasts available. The second set contains 14 images from patients with multiple sclerosis selected from the MICCAI MS segmentation challenge64 (MSSEG). Finally, the last set consists of 14 images with brain tumors taken from the Multimodal Brain Tumor Segmentation Challenge (BraTS) validation dataset65,66,67.
Conclusions
We have investigated the effectiveness of several deterministic and probabilistic deep learning methods in transforming MR images. To that end, we implemented and evaluated a U-Net-based direct inference and a residual vision transformer model as benchmarks. The former is based on a deterministic formulation, while the latter has a probabilistic formulation but is implemented to generate outputs deterministically. We implemented and extensively evaluated the conditional version of three major generative models, namely, the Conditional Generative Adversarial Network (cGAN), the Denoising Diffusion Probabilistic Model (DDPM), and the Noise Conditional Score Network (NCSN). We trained all these models on the publicly available IXI and OASIS datasets and then applied the trained models to transform T1, T2, FLAIR, and PD MR images. We quantified this performance by computing the SSIM, PSNR, LPIPS, and BSSIM image similarity metrics. We concluded that the two diffusion models consistently yielded more accurate transformations. When comparing two diffusion models, the DDPM model yields higher accuracy, especially with larger samples \((n \ge 5)\) and remains more robust when encountering inputs from slightly different distributions. However, the DDPM model exhibits certain issues. Especially, it produces isolated pixels with outlier values. On the other hand, the NCSN model yields slightly lower accuracy. However, the NCSN model demonstrates several favorable characteristics. For example, it is faster and achieves higher accuracy with a lower number of samples (\(n < 5\)). Also, the outputs of this model are artifact-free and do not require any additional post-processing.
We also investigated the utility of the generative property of the models, i.e., their ability to produce an ensemble of outputs for a given input image. In particular, in addition to the standard MR images, we incorporated several out-of-distribution (OOD) sets of inputs, including images with artifacts and pathologies, such as Alzheimer’s disease, multiple sclerosis (MS) lesions, and brain tumors. Thereafter, for each input image, we computed the average pixel-wise standard deviation using the three probabilistic models. We concluded that this quantity effectively distinguished OOD images from their regular counterparts in cGAN and NCSN models.
In clinical research, the proposed image transformation models can generate image contrasts that are otherwise unavailable, thereby improving data completeness for downstream analyses. The accompanying uncertainty-based OOD detection provides an automated quality control mechanism that enhances analytical reliability by excluding low-quality data and reducing the time and cost of manual inspection. From a clinical standpoint, this framework could be integrated into image reconstruction or transformation pipelines to automatically flag uncertain outputs for radiologist review. Although deep learning–based models are not yet routinely embedded in clinical workflows, the proposed approach offers a practical and transparent pathway toward uncertainty-aware, quality-controlled image generation.
While the diffusion models (NCSN and DDPM) achieved higher accuracy than the other models, they were also more computationally demanding, particularly during inference. Training times were approximately 42 h for DDPM, 29 h for NCSN, 25 h for cGAN, and 47 h for ResViT (9 h + 38 h for sequential phases). The DDPM and NCSN models required about 15 s and 13 s, respectively, to generate a single slice on an NVIDIA A40 GPU (48 GB), whereas the cGAN, ResViT, and Direct models completed the task in under 0.1 s. Training with a batch size of 16 required approximately at least 16–24 GB of GPU memory, while inference could be performed on less advanced GPUs (e.g., NVIDIA GeForce RTX 2080). The cGAN model required additional memory during training due to its dual-network architecture and the gradient penalty term. Sampling for diffusion models can be further accelerated using batched inputs, resulting in approximately 4 and 2.5 seconds of generation time per image for the DDPM and NCSN models, respectively.
We also conducted several analyses to explore the characteristics of the distribution of the samples produced by the generative models. We used the Shapiro-Wilk test to investigate the normality of the distribution of values generated for individual pixels. We concluded that approximately \(89\%\) of distributions in NCSN outputs demonstrate normality. This value is about \(51\%\) and \(41\%\) for the cGAN and DDPM models, respectively. We further investigated the calibration errors for the outputs of these models. Again, the NCSN model exhibited the best calibration properties among the generative models. In contrast, the cGAN and DDPM model outputs showed overconfidence and underconfidence, respectively.
In this work, we have focused on implementing and quantifying the performance of the proposed models in terms of image similarity metrics and exploring their distributional measures and characteristics for uncertainty quantification. A potential follow-up study could explore the use of transformed images in downstream tasks. For instance, most structural analysis software, like FreeSurfer, primarily rely on T1 MR scans. In the follow-up study, one could consider the use of native T1 images and T1 images obtained from transforming T2 images in FreeSurfer and compare the resulting outputs from the two sets. The outputs would include quantities of clinical interest, such as the volume of brain tissues. An analysis of this type would also provide the groundwork for determining whether MR images of a given type, such as T2 images, can effectively replace T1 images in clinical applications.
Data availability
The data generated and/or analyzed during the current study are available from the corresponding author upon reasonable request.
References
Litjens, G. et al. A survey on deep learning in medical image analysis. Med. Image Anal. 42, 60–88 (2017).
Chen, X. et al. Recent advances and clinical applications of deep learning in medical image analysis. Med. Image Anal. 79, 102444 (2022).
Wang, T. et al. A review on medical imaging synthesis using deep learning and its clinical applications. J. Appl. Clin. Med. Phys. 22, 11–36 (2021).
Dayarathna, S. et al. Deep learning based synthesis of mri, ct and pet: Review and analysis. Med. Image Anal. 92, 103046 (2023).
Zhao, S., Geng, C., Guo, C., Tian, F. & Tang, X. Saru: A self-attention resunet to generate synthetic ct images for mr-only bnct treatment planning. Med. Phys. 50, 117–127 (2023).
Sikka, A. et al. Mri to pet cross-modality translation using globally and locally aware gan (gla-gan) for multi-modal diagnosis of alzheimer’s disease. arXiv preprint arXiv:2108.02160 (2021).
Li, Q. et al. Eliminating ct radiation for clinical pet examination using deep learning. Eur. J. Radiol. 154, 110422 (2022).
Li, W. et al. Magnetic resonance image (mri) synthesis from brain computed tomography (ct) images based on deep learning methods for magnetic resonance (mr)-guided radiotherapy. Quant. Imaging Med. Surg. 10, 1223 (2020).
Aljuaid, A. & Anwar, M. Survey of supervised learning for medical image processing. SN Comput. Sci. 3, 292 (2022).
Dinkla, A.M. et al. Mr-only brain radiation therapy: dosimetric evaluation of synthetic cts generated by a dilated convolutional neural network. Int. J. Radiat. Oncol. Biol. Phys. 102, 801–812 (2018).
Xiang, L. et al. Deep embedding convolutional neural network for synthesizing ct image from t1-weighted mr image. Med. Image Anal. 47, 31–44 (2018).
Hu, F. et al. Image harmonization: A review of statistical and deep learning methods for removing batch effects and evaluation metrics for effective harmonization. NeuroImage 274, 120125 (2023).
Vajjhala, N.R., Roy, S.S., Taşcı, B. & Chowdhury, M.E.H. Foundations and emerging trends in generative artificial intelligence (ai) for industrial applications. Gener. Artif. Intell. (AI) Approaches for Ind. Appl. 1–18 (2025).
Goodfellow, I. et al. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27, 2672–2680 (2014).
Isola, P., Zhu, J.-Y., Zhou, T. & Efros, A. A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, 1125–1134 (2017).
Chen, X. & Jia, C. An overview of image-to-image translation using generative adversarial networks. In Pattern Recognition. ICPR International Workshops and Challenges: Virtual Event, January 10–15, 2021, Proceedings, Part VI, 366–380 (Springer, 2021).
Yang, Q. et al. Mri cross-modality image-to-image translation. Sci. Rep. 10, 3753 (2020).
Dar, S. U. et al. Image synthesis in multi-contrast mri with conditional generative adversarial networks. IEEE Trans. Med. Imaging 38, 2375–2388 (2019).
Dey, A. & Ebrahimi, M. Mtsr-mri: Combined modality translation and super-resolution of magnetic resonance images. In Medical Imaging with Deep Learning, 743–757 (PMLR, 2024).
Song, Y. & Ermon, S. Generative modeling by estimating gradients of the data distribution. Adv. Neural Inf. Process. Syst. 32, 11918–11930 (2019).
Song, Y. & Ermon, S. Improved techniques for training score-based generative models. Adv. Neural Inf. Process. Syst. 33, 12438–12448 (2020).
Batzolis, G., Stanczuk, J., Schönlieb, C.-B. & Etmann, C. Conditional image generation with score-based diffusion models. arXiv preprint arXiv:2111.13606 (2021).
Kazerouni, A. et al. Diffusion models in medical imaging: A comprehensive survey. Med. Image Anal. 88, 102846 (2023).
Yang, L. et al. Diffusion models: A comprehensive survey of methods and applications. ACM Comput. Surv. 56, 1–39 (2023).
Zhao, H. et al. Diffusion models for computational neuroimaging: A survey. arXiv preprint arXiv:2502.06552 (2025).
Zhang, J., Yan, R., Perelli, A., Chen, X. & Li, C. Phy-diff: Physics-guided hourglass diffusion model for diffusion mri synthesis. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 345–355 (Springer, 2024).
Kim, J. & Park, H. Adaptive latent diffusion model for 3d medical image to image translation: Multi-modal magnetic resonance imaging study. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 7604–7613 (2024).
Zhu, L. et al. Make-a-volume: Leveraging latent diffusion models for cross-modality 3d brain mri synthesis. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 592–601 (Springer, 2023).
Li, Y. et al. Zero-shot medical image translation via frequency-guided diffusion models. IEEE Trans. Med. Imaging 43, 980–993 (2023).
Özbey, M. et al. Unsupervised medical image translation with adversarial diffusion models. IEEE Trans. Med. Imaging. 42, 3524–3539 (2023).
Jiang, H. et al. Fast-ddpm: Fast denoising diffusion probabilistic models for medical image-to-image generation. IEEE J. Biomed. Health Inf. 29, 7326–7335 (2025).
Chen, J. et al. Source-free cross-modality medical image synthesis with diffusion priors. J. King Saud Univ. Comput. Inf. Sci. 37, 214 (2025).
Wang, Y. et al. Unisyn: A generative foundation model for universal medical image synthesis across mri, ct and pet. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 673–682 (Springer, 2025).
Yazdani, M., Medghalchi, Y., Ashrafian, P., Hacihaliloglu, I. & Shahriari, D. Flow matching for medical image synthesis: Bridging the gap between speed and quality. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 216–226 (Springer, 2025).
Dalmaz, O., Yurt, M. & Çukur, T. Resvit: residual vision transformers for multimodal medical image synthesis. IEEE Trans. Med. Imaging 41, 2598–2614 (2022).
Aburass, S., Dorgham, O., Al Shaqsi, J., Abu Rumman, M. & Al-Kadi, O. Vision transformers in medical imaging: a comprehensive review of advancements and applications across multiple diseases. J. Imaging. Inform. Med. 1–44 (2025).
Kohl, S. et al. A probabilistic u-net for segmentation of ambiguous images. Adv. Neural Inf. Process. Syst. 31, 6965–6975 (2018).
Alblwi, A., Makkawy, S. & Barner, K. E. D-ddpm: Deep denoising diffusion probabilistic models for lesion segmentation and data generation in ultrasound imaging. IEEE Access. 13, 41194–41209 (2025).
Zou, K. et al. A review of uncertainty estimation and its application in medical imaging. Meta-Radiology 1, 100003 (2023).
Faghani, S. et al. Quantifying uncertainty in deep learning of radiologic images. Radiology 308, e222217 (2023).
Ghoshal, B., Tucker, A., Sanghera, B. & Lup Wong, W. Estimating uncertainty in deep learning for reporting confidence to clinicians in medical image segmentation and diseases detection. Comput. Intell. 37, 701–734 (2021).
Lakara, K. & Valdenegro-Toro, M. Disentangled uncertainty and out of distribution detection in medical generative models. arXiv preprint arXiv:2211.06250 (2022).
Mirza, M. & Osindero, S. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014).
Arjovsky, M., Chintala, S. & Bottou, L. Wasserstein generative adversarial networks. In International conference on machine learning, 214–223 (PMLR, 2017).
Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V. & Courville, A. Improved training of wasserstein gans. arXiv preprint arXiv:1704.00028 (2017).
Adler, J. & Öktem, O. Deep bayesian inversion. arXiv preprint arXiv:1811.05910 (2018).
Wolleb, J., Sandkühler, R., Bieder, F., Valmaggia, P. & Cattin, P. C. Diffusion models for implicit image segmentation ensembles. In International Conference on Medical Imaging with Deep Learning, 1336–1348 (PMLR, 2022).
Nichol, A. Q. & Dhariwal, P. Improved denoising diffusion probabilistic models. In International Conference on Machine Learning, 8162–8171 (PMLR, 2021).
Ixi dataset website: https://brain-development.org/ixi-dataset/.
LaMontagne, P. J. et al. Oasis-3: longitudinal neuroimaging, clinical, and cognitive dataset for normal aging and alzheimer disease. MedRxiv 2019–12 (2019).
Zhang, R., Isola, P., Efros, A.A., Shechtman, E. & Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition, 586–595 (2018).
Zhu, S. et al. A multi-resolution hybrid cnn-transformer network with scale-guided attention for medical image segmentation. IEEE J. Biomed. Health Inf. 29, 8385–8394 (2025).
Ou, Y. et al. Patcher: Patch transformers with mixture of experts for precise medical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 475–484 (Springer, 2022).
Kuleshov, V., Fenner, N. & Ermon, S. Accurate uncertainties for deep learning using calibrated regression. In International conference on machine learning, 2796–2804 (PMLR, 2018).
Dumoulin, V., Shlens, J. & Kudlur, M. A learned representation for artistic style. arXiv preprint arXiv:1610.07629 (2016).
Ho, J., Jain, A. & Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 33, 6840–6851 (2020).
Song, Y. et al. Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456 (2020).
Dohmen, M., Klemens, M. A., Baltruschat, I. M., Truong, T. & Lenga, M. Similarity and quality metrics for mr image-to-image translation. Sci. Rep. 15, 3853 (2025).
Woodland, M. et al. Feature extraction for generative medical imaging evaluation: New evidence against an evolving trend. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 87–97 (Springer, 2024).
Kumar, A., Liang, P. & Ma, T. Verified uncertainty calibration. In Adv. Neural Inf. Process. Syst. (NeurIPS) (2019).
Fsl-flirt users’ guide: https://fsl.fmrib.ox.ac.uk/fsl/fslwiki/FLIRT/UserGuide.
Moazami, S., Ray, D., Pelletier, D. & Oberai, A. A. Probabilistic brain extraction in mr images via conditional generative adversarial networks. IEEE Trans. Med. Imaging. 43, 1071–1088 (2023).
Fsl example box: Mri artifacts https://www.fmrib.ox.ac.uk/primers/intro_primer/ExBox1/IntroBox1.html.
Commowick, O. et al. Multiple sclerosis lesions segmentation from multiple experts: The miccai 2016 challenge dataset. Neuroimage 244, 118589 (2021).
Menze, B. H. et al. The multimodal brain tumor image segmentation benchmark (brats). IEEE Trans. Med. Imaging 34, 1993–2024 (2014).
Bakas, S. et al. Advancing the cancer genome atlas glioma mri collections with expert segmentation labels and radiomic features. Sci. data 4, 1–13 (2017).
Bakas, S. et al. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the brats challenge. arXiv preprint arXiv:1811.02629 (2018).
Acknowledgements
A.A.O. would like to acknowledge the support of ARO Grant AWD-00009144. C.J.A. receives grant support from the National MS Society, the National Institutes of Health, and the Keck School of Medicine. She has received consulting honoraria for participation in advisory boards, data safety monitoring committee, and/or MRI Adjudication Committee for Horizon Therapeutics, TG Therapeutics, EMD Serono, Genentech, and Zenas Biopharma; and for participation in educational events for Sanofi Genzyme, MJH Life Sciences, Efficient LLC, and Spire Learning. The authors acknowledge the Center for Advanced Research Computing (CARC) at the University of Southern California for providing computing resources that have contributed to the research results reported within this publication. URL: https://carc.usc.edu.
Funding
This work is supported by ARO Grant AWD-00009144.
Author information
Authors and Affiliations
Contributions
S.R., S.M., and A.A.O. contributed to conceptualizing and designing the study and writing the manuscript. S.R. and S.M. conducted the numerical experiments of the study and implemented the algorithms. A.A.O. oversaw the theoretical aspects of the study. C.J.A. oversaw the neurological interpretations of the study.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Rezvani, S., Moazami, S., Azevedo, C.J. et al. Probabilistic brain MR image transformation using generative models. Sci Rep 16, 2161 (2026). https://doi.org/10.1038/s41598-025-31958-3
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-025-31958-3












