Table 1 GAN architecture, hyperparameters, and training details.

From: Multimodal medical image fusion combining saliency perception and generative adversarial network

Component

Details

Generator architecture

Consists of multiple convolutional layers with ReLU activation, followed by batch normalization. Upsampling layers refine image quality.

Discriminator architecture

A series of convolutional layers with LeakyReLU activation, batch normalization, and a fully connected layer for classification.

Input dimensions

Image patches of size 256 × 256 pixels for training.

Loss function

Binary Cross-Entropy (BCE) loss for adversarial training.

Optimizer

Adam optimizer with a learning rate of 0.0002 for stable convergence.

Batch size

32 images per batch to balance training stability and computational efficiency.

Training epochs

1200 epochs with early stopping based on validation loss.

Regularization

Dropout (0.3) applied to prevent overfitting in both generator and discriminator.

Evaluation metrics

F1-score, Structural Similarity Index (SSIM), Peak Signal-to-Noise Ratio (PSNR) for quality assessment.

Data augmentation

Random flipping, rotation, and contrast adjustments applied to improve generalization.