Abstract
Electromagnetic (EM) simulation is widespread in microwave engineering. EM tools ensure evaluation reliability but incur significant expenses. These can be mitigated by employing surrogate modeling methods, especially to expedite design workflows like local/global optimization or uncertainty quantification. However, building accurate surrogates is a daunting task beyond simple cases (low dimensionality, narrow geometry parameter and frequency ranges). This research suggests a new technique for dependable modeling of microwave circuits. Its main ingredient is a recurrent neural network (RNN) with the main architectural components being bidirectional Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) layers. These are incorporated to accurately represent frequency relationship within circuit characteristics as well as dependencies between its dimensions and outputs considered as vector-valued functions parameterized by frequency. The network’s hyperparameters are adjusted through Bayesian Optimization (BO). Utilization of frequency as a sequential variable handled by RNN is a distinguishing feature of our approach, which leads to the enhancement of dependability and cost efficiency. Another critical factor is dimension- and volume-wise reduction of the model’s domain achieved through global sensitivity analysis. It allows for additional and dramatic accuracy improvements without diminishing the surrogate’s coverage regarding circuit’s operating parameters. Our methodology has been extensively validated using several microstrip structures. The results demonstrate its competitive performance over a range of kernel-based regression techniques and diverse neural networks. The proposed procedure ensures building models of outstanding predictive power while using small training datasets, which is beyond the capabilities of benchmark algorithms.
Similar content being viewed by others
Introduction
Computational tools are indispensable in modern microwave engineering1,2 with the special emphasis on electromagnetic (EM) solvers3,4. EM simulation is versatile and enables quantification of effects that cannot be evaluated using different methods (cross-coupling, dielectric/radiation losses, anisotropy, the impact of environmental components such as connectors or installation fixtures). For many circuits, such as compact structures, substrate-integrated waveguide (SIW)-based circuits, or structures incorporating metamaterials5,6,7,8,9, EM simulation is imperative when it comes to accurate characterization of their electrical properties. However, EM analysis is CPU intensive, which impedes its utilization in procedures requiring multiple system evaluations. Examples include design closure10,11, statistical analysis12,13,14, or global and multi-objective design15,16,17,18. The latter seems to be the most challenging endeavor, typically executed using bio-inspired algorithms with expenses reaching thousands of merit function calls19,20,21,22,23.
Recent years observed significant research focus on expedited EM-driven design methodologies. A range of methods were developed to accelerate gradient-based algorithms (adjoint sensitivity, restricted Jacobian updating, parallelization, mesh deformation24,25,26,27,28). More generic approaches include feature-based techniques29,30, variable-fidelity approaches31,32,33,34, or dimensionality reduction35,36,37,38. Notwithstanding, the main emphasis is currently put on surrogate-assisted procedures39,40,41,42,43, typically arranged as machine learning (ML) algorithms44,45,46. Widely used modelling methods include support vector regression, Gaussian processes, kriging, polynomial chaos expansion, ensemble learning, and diverse types of neural networks47,48,49,50,51,52,53,54,55.
Shifting computational operations to a fast metamodel effectively mitigates the cost-related issues of EM-driven design. At the same time, constructing reliable surrogates is a difficult undertaking except for simple scenarios (small number of design variables, narrow parameter ranges). The main problem is the curse of dimensionality56 and the large volume of the traditional (interval-based) domain. Both factors lead to a fast increase in the size of the training set necessary to build an accurate data-driven model. This is why surrogate-based procedures normally take the form of ML algorithms in which case the surrogate is constructed in a restricted region believed to encapsulate the optimum design. Much of the space is not explored. Clearly, data-driven models constructed during the ML-based search are not general-purpose57. Method for alleviating the mentioned difficulties in a more generic setting include better utilization of available data (ensemble learning58,59), improved handling of large datasets (deep learning60,61), exploring specific system output structure (high-dimensional model representation, HDMR, orthogonal matching pursuit62,63). Alternative techniques include multi-fidelity methods64,65,66, and performance-driven modelling67,68,69,70,71. Therein, the domain is limited to a subspace containing high-quality designs. Approximation of this region requires extra computational effort67; however, the surrogate exhibits predictive power that cannot be matched by conventional methods.
Recently, the growing popularity of artificial neural networks (ANNs) has been observed in modelling of high-frequency structures. ANNs are typically used within ML-based design procedures. The employed architectures are often variations of feedforward networks, specifically multi-layer perceptrons (MLPs)72,73,74,75,76,77,78, convolutional neural networks (CNNs)79,80,81, deep neural networks (DNNs)82,83,84. At times, specialized architectures are applied, e.g., long short-term memory (LSTM) layers85, multi-fidelity86, evolutionary87, autoencoders88). Some studies use inverse ANNs89,90. General-purpose ANN-based modeling is rare due to challenges discussed earlier. Some exemplary works involve DNN53, CNN91, MLP92, graph neural networks93, cascaded networks94, sensitivity networks95, or hybrid networks96. The mentioned works utilize ANNs as regressors without exploring intricate relationships between decision variables and system features encoded in frequency responses. As a result, their operation and reliability are similar to conventional regression techniques.
This paper introduces an innovative strategy for high-performance surrogate modelling of microwave circuits. The proposed method leverages handling S-parameter characteristics as sequential data ensembles parameterized by frequency and exploring their dependence on decision variables (e.g., geometry parameters). The underlying surrogate is a recurrent neural network (RNN). RNNs are well-suited to processing sequential information. The specific RNN architecture incorporates single- and bi-directional Long Short-Term Memory (LSTM) layers, the Gated Recurrent Unit (GRU) layer, and fully connected and (output) regression layers to produce the final predictions of the complete frequency responses. The model’s hyperparameters, including the number of units in each layer, are adjusted through Bayesian Optimization (BO). Treating frequency as a sequential parameter distinguishes our approach from conventional regression and machine learning methods while being advantageous for the dependability and overall efficacy of the modeling process. Dimensionality reduction realized with global sensitivity analysis (GSA) constitutes another mechanism incorporated to improve the surrogate’s accuracy dramatically. GSA aims to yield orthogonal directions responsible for the maximum variability of the system at hand and span the restricted domain along them. The modeling strategy developed in this work is validated with the help of several microstrip circuits and compared to a range of benchmark techniques such as neural networks and diverse regression surrogates. The results underscore the remarkable predictive power of our metamodels, which is superior to all benchmark techniques. Considerable improvement is observed regardless of whether the model is established in the conventional (box-constrained) domain or dimensionality-reduced region. Furthermore, usable surrogates can be constructed using small numbers of training points, which was not the case for most of the comparison methods.
The original contributions of this study include (i) the development of a novel RNN-based metamodel for precise representing of circuit’s characteristics, (ii) handling frequency as a sequential parameter to facilitate the data-driven representation of the scattering parameters and capture the relationships between systems outputs at various frequencies and the design variables, (iii) incorporating LSTM and GRU layers for efficiency frequency-wise dependencies processing, (iv) utilization of Bayesian optimization for boosting the surrogate’s accuracy, (v) development of complete modeling framework that leverages explicit dimensionality reduction through global sensitivity analysis, (vi) demonstrating remarkable reliability of our method and its advantages over several benchmark procedures.
RNN and dimensionality reduction for precise microwave circuit modeling
The fundamental components of the presented modeling approach are elaborated here. We first recall the formulation of the modeling problem in Sect. “Microwave modelling”. The overall structure and the working principles of the suggested RNN are provided in Sect. “RNN-based surrogates with sequential processing of frequency responses”. The domain confinement mechanism is discussed in Sect. “Domain Confinement Using Global Sensitivity Analysis”, whereas Sect. “Modeling Procedure” puts together the operating flow of the complete algorithm.
Microwave modelling
Let Rf(x) represent the primary (EM-simulated) model of the circuit of interest. Here, x = [x1 … xn]T are design parameters. Rf stands for the aggregated system outputs, typically scattering parameters Skj(x,f), where f is frequency, and k and j mark the respective circuit ports. We aim to build a low-cost surrogate Rs(x) that accurately represents Rf(x) within the domain X. Traditionally, X is an interval [l u] defined by the lower and upper bounds for parameters l = [l1 … ln]T and u = [u1 … un]T.
The surrogate’s accuracy is evaluated by means of a suitable error metric (see, e.g.,97,98). Here, the relative root-mean-square error (RRMSE) is used, defined as ||Rs(x) – Rf(x)||/||Rf(x)|| (if system responses contain multiple vectors, the Frobenius norm is employed). To estimate the accuracy over the entire domain, the average error Eaver is computed.

where {xt(k)}k = 1, …, Nt, are independent testing (hold-off) samples. The relative error is convenient as it matches well the visual alignment between the metamodel and EM evaluated outputs. Less than ten percent of RRMSE typically translates into good alignment between Rf and Rs and makes the model suitable for design purposes.
RNN-based surrogates with sequential processing of frequency responses
This section describes the architecture of the proposed recurrent neural network (RNN) surrogate with sequential frequency data processing. The underlying concept is elucidated in Sect. “RNN modeling with sequential frequency processing”. Section “Model Structure” discusses the network structure, essential layers, and the working principles. RNN training and hyperparameter optimization are outlined in Sect. “Hyperparameter Optimization”.
RNN modeling with sequential frequency processing
Predicting the frequency characteristics of microwave signals poses a unique challenge. However, the modeling process may be facilitated by exploring dependencies between the signal state at the given frequency and the lower and higher ones. Thus, instead of traditional methods, Recurrent Neural Networks, which are designed to recognize patterns over sequences, are well-suited for this purpose. The central feature of an RNN is its ability to retain information across time steps through a recurrent structure, allowing it to capture dependencies over extended sequences. When applied to frequency data, RNNs treat each frequency point as a sequential input, which allows the model to learn from the dependencies between these points.
Frequency data often exhibits dependencies that vary across scales—both in the immediate range (e.g., a particular shape of a resonance) and over more extended periods (e.g., relationship between the fundamental operating frequency and the harmonics), necessitating a model that can remember and adjust based on both recent and long-past data. Thus, to account for the same the current model combines several types of RNN layers with optimization strategies such that the predicted value is as close to corresponding frequency characteristic.
Model structure
For an effective learning and accurate frequency response prediction, the proposed model combines a set of LSTM, GRU and Bi-LSTM layers. In this section, we introduce and explore the specific roles of each of the layers. The operating principles of the LSTM layer have been shown in Fig. 1. The purpose of this layer is to capture long-term dependencies in the sequential data. The gated mechanisms regulate the flow of information, selectively remembering or forgetting past inputs to model complex temporal relationships over extended sequences. The second type is a GRU layer (cf. Figure 2), which is a streamlined variant of the LSTM. GRU is designed to reduce computational complexity while retaining the ability to representing sequential data dependencies, which, in our case, are the relationships between the circuit’s response along the considered frequency spectrum. Yet another tool incorporated into the proposed model is the bidirectional LSTM (Bi-LSTM) layer, shown in Fig. 3, which enables the model to capture temporal relationships in a more comprehensive manner by making the system representation at any given frequency dependent on both lower and higher parts of the spectrum. Fully connected and regression layers (Fig. 4) are the final components of the proposed RNN surrogate.
Fully connected layer. It is the final layer of the proposed model. It maps the high-dimensional features from recurrent layers into a prediction space for complex-valued responses. In contrast, the regression layer minimizes prediction error by assessing the discrepancy with actual responses. Together, these layers refine the output for improved prediction accuracy.
Their role of the former is to map the high-dimensional features from recurrent layers into a prediction space for rendering complex-valued responses. The latter minimizes prediction error by assessing the discrepancy with actual responses. Together, these layers refine the output for improved prediction accuracy. The architecture of the overall RNN metamodel can be found in Fig. 5. The input features x and the output signals y are extracted and structured in the pre-processing module. The RNN processes the input features through a sequence of layers, starting with the LSTM layer followed by GRU to capture long-term dependencies, and Bi-LSTM layers for efficient modeling and bidirectional context to build the microwave outputs. These recurrent layers extract temporal features, which are then transformed by fully connected layers into prediction space. Finally, the regression layer evaluates the model by comparing the predicted complex-valued frequency responses to the actual values using the relative error metric.
Initial experimentation with different configurations of LSTM and GRU layers revealed challenges in achieving an optimal balance between underfitting and overfitting. Models using only LSTM or GRU layers struggled to capture the full complexity of the data, but with increased layer depth, it exhibited signs of overfitting. Similarly, a fully Bi-LSTM-based architecture, while providing bidirectional context, resulted in excessive training times without substantial gains in accuracy. Adjusting the number of neurons per layer in these configurations did not yield improvements beyond existing benchmarks. Following multiple rounds of testing, we determined that a hybrid architecture combining LSTM, GRU, and Bi-LSTM layers provided the best trade-off between accuracy, generalization, and training efficiency. Specifically, the LSTM layers capture long-term dependencies in the sequence, GRU layers enable deeper architectures with reduced computational cost compared to LSTM, and Bi-LSTM layers enhance contextual understanding by incorporating bidirectional dependencies. This architecture was finalized after systematically evaluating alternative designs and observing consistent performance gains across different datasets (see Sect. “Modeling Procedure”).
Hyperparameter optimization
The hyperparameter tuning and optimization process is essential for refining the model’s predictive capability, accuracy, and generalization. In this work, it is realized using Bayesian Optimization (BO)99 at the level of the number of RNN units per layer, learning rate, etc. Its purpose is to identify the best possible model architecture. At the same time, the layer-specific parameters, i.e., the weights, are optimized using the ADAM procedure100. Figure 6 summarizes the operations undertaken to identify the model and the roles of the applied algorithmic tools.
Domain confinement using global sensitivity analysis
Appropriate handling of the training data is one of the essential aspects of reliable surrogate modeling. The second is the proper selection of the model domain, which—in this study—is based on dimensionality reduction involving fast global sensitivity analysis (FGSA), initially proposed in69. The idea is to span the domain along the vectors responsible for the most significant changes in the system outputs while leaving out the remaining directions.
This enables the dimensionality-related difficulties to be tackled without compromising the design utility of the model. FGSA is used instead of conventional techniques such as variable screening101,102,103 or traditional global sensitivity analysis (Sobol indices, Jansen’s method104,105,106) capitalizing on its low running cost and flexibility. In particular, the essential directions constructed by FGSA can be arbitrarily oriented (i.e., not aligned with coordinate system axes), which enables accounting for joint effects of pairs or triples of design variables without eliminating individual parameters. The operating flow of FGSA has been outlined in Fig. 7. The random vectors xs(k) are generated using Latin Hypercube Sampling (LHS)107. The eigenvectors and eigenvalues are found using principal component analysis108. The eigenvalues are arranged in descending order λ1 ≥ λ2 ≥ … ≥ λn, i.e., the effect of subsequent eigenvectors on the circuit’s characteristics gradually decreases.
The outline of FGSA69. The vectors ej account for the directions that significantly affect the system outputs; their relevance is assessed by the eigenvalues λj.
The dimensionality-reduced domain of the metamodel is spanned by Nd most critical eigenvectors ej, which collectively account for most of the system’s response variability. Let Cmin be the joint variability threshold. The number Nd is the smallest integer such that99
In this work, Cmin = 0.9 so that the vectors ej defining the domain are associated with 90% of the circuit’s response variability.
The reduced domain Xd is defined as69

where xc = [l + u]/2 is the original domain’s center, whereas aj, j = 1, …, Nd, are real numbers. In other words, Xd is an intersection of the Nd-dimensional subspace determined by ej, j = 1, …, Nd, and the conventional parameter space X.
Modeling procedure
This section puts together the operation of the complete modeling framework utilizing the components elucidated in Sects. “Microwave Modelling” through Sects. “Domain Confinement Using Global Sensitivity Analysis”. The focus is on the training process and model validation.
Data extraction and preprocessing
Data extraction is a critical first step in preparing the input and output variables for RNN training. The dataset consists of four sets of complex-valued frequency responses, and through the pre-processing module in Fig. 5, we structure the data into matrices that represent the antenna’s behavior across frequencies, formatted for sequential processing by the RNN.
Model architecture and design
As mentioned earlier, the proposed architecture integrates multiple types of recurrent layers, each selected for its unique strengths in processing temporal dependencies. The first part of the model is an input layer that feeds into an LSTM (Long Short-Term Memory) layer. LSTM captures long-term dependencies within the frequency sequence by retaining past information across time steps. Following the LSTM layer, a GRU (Gated Recurrent Unit) layer is added to provide efficient processing with fewer parameters, maintaining relevant patterns in the sequence without excessive computational overhead. Subsequently, a Bi-LSTM (Bidirectional LSTM) layer is incorporated to process the sequence in both directions (forward and backward), which enables the model to utilize contextual information from both preceding and succeeding frequency points. This bidirectional setup enhances the model’s understanding of the sequence, as it learns from both past and future data points (here, meaning signal levels corresponding to the lower and higher frequencies) simultaneously.
The outputs from these recurrent layers are processed by a fully connected layer, which performs a linear transformation that maps the high-dimensional feature space into the final prediction space, representing the complex-valued frequency response of the antenna. Finally, a regression layer computes the model’s prediction quality using the Mean Squared Error (MSE) as the loss function. This layer guides the training process by iteratively adjusting the model parameters to minimize prediction error. The combination of these recurrent layers with optimized hyperparameters, forms a robust architecture capable of capturing both short-term and long-term dependencies in the antenna’s frequency response, resulting in highly accurate and generalizable predictions.
Hyperparameter tuning
To optimize the model’s accuracy, Bayesian Optimization (BO) is employed as mentioned in Sect. “Hyperparameter Optimization”. It fine-tunes critical hyperparameters, like RNN layer units and learning rate, by iteratively refining a surrogate model to identify the optimal configuration minimizing the relative error metric. The specific parameters optimized include:
-
Layer Units: The number of units in each RNN layer (LSTM, GRU, Bi-LSTM) to adjust the model’s complexity.
-
Learning Rate: Adaptively tuned to enhance convergence and generalization.
Training process
Model training is conducted using the Adam optimizer (cf. Section “Hyperparameter Optimization”), with an adaptive learning rate, which facilitates stable convergence over the epochs. A key feature of this setup includes LearnRateDropFactor and LearnRateDropPeriod parameters dynamically adjust the learning rate, reducing it to approximately 36% of the initial value by the end of training. This prevents overfitting by allowing larger updates initially, then gradually refining the model.
Model evaluation and prediction
After training, the model undergoes evaluation using a distinct set of test points. The mean relative error (cf. (1)) is calculated to quantify the discrepancies between predicted and true (EM-evaluated) frequency responses. This metric, averaged across the four outputs, serves as an overall measure of model performance. Predictions, encompassing both real and imaginary components in complex format, are saved along with the trained models to design paths for future analysis and validation.
For a supplementary clarification, Fig. 8 shows the diagram explaining the data flow in the proposed model. Furthermore, Fig. 9 illustrates the operating flow of the complete model construction process, which includes the establishment of the confined domain, sampling, acquisition of the EM simulation data, and identification of the RNN metamodel.
Results
The proposed modeling methodology is demonstrated using three circuits. It is also juxtaposed against several benchmark procedures. The material is organized as follows. Section “Benchmark Methods” outlines the benchmark techniques. Section “Setup” covers the setup. The results and discussion are provided in Sects. “Results” and Sect. “Discussion”. Application of the model for circuit optimization and experimental validation of the optimized circuits are included in Sect. “Applications Case Studies and Experimental Validation”.
Benchmark methods
The benchmark methods are outlined in Table 1. These include kriging interpolation, radial basis functions, Gaussian process regression, support vector machines, and artificial neural networks (ANN). The last three models are deep feedforward ANN architectures added to provide a more meaningful comparison with the proposed RNN-based model and to demonstrate that sequential processing of frequency data does provide distinctive benefits, which cannot be achieved by merely increasing the network architecture complexity. These frameworks are widely used in high-frequency engineering in diverse applications (general-purpose modeling, global and multi-objective optimization, and machine learning). Consequently, they can be considered representative in the considered field. The cited references109,110,111,112 provide more extensive information about each method.
Setup
The suggested technique is employed to build surrogate models for the circuits discussed in Sects. “Example I” through Example III. To comprehensively demonstrate the properties of our approach, the surrogates are rendered in both the conventional space X and the reduced domain Xd. This enables observing the advantages of the RNN-based surrogate and the benefits of combining it with dimensionality reduction.
The dimensionality of Xd is determined using FGSA with Cmin = 0.9 (Sect. “Domain Confinement Using Global Sensitivity Analysis”). The models are built using datasets of different sizes, from 50 to 800 samples. The training points are allocated using LHS 107. The model accuracy is evaluated using the RRMSE defined in Sect. “Microwave Modelling” based on 100 independent testing samples (see also (1)). The hyperparameter space has also been summarized in Table 2.
The specific model setup is as follows:
-
Architecture (as described in Sect. “RNN-based Surrogates with Sequential Processing of Frequency Responses”): A deep Recurrent Neural Network comprising LSTM, GRU, and Bi-LSTM layers, followed by a Fully Connected output layer and a Regression layer to predict complex-valued antenna frequency responses.
-
Hyperparameter Tuning: Utilizes Bayesian Optimization to fine-tune the number of units in each layer and the learning rate. The number of BO iterations was set to 25.
-
Layer arrangement: 500 − 750 (LSTM) → 500–1000 (GRU) → 500 − 1000 (Bi-LSTM) → output (real + imaginary).
-
Training Configuration: The model is trained for up to 10,000 epochs using the Adam optimizer, with early stopping and learning rate scheduling to prevent overfitting.
Comparing the training time and computational complexity, in general, we can safely state that the first four benchmark models (kriging, RBF, GPR, SVR) are cheap to construct. These models rely on relatively simple optimization tasks, such as maximum likelihood estimation for kernel-based techniques, which typically complete within seconds to a few minutes, depending on dataset size their training times are negligible compared to the time required to acquire electromagnetic (EM) simulation data, which can take several minutes per simulation. In comparison, the ANN (the fifth benchmark method) training times range from 2–3 min (for 50-point training set) to approximately an hour for the largest datasets. However, it remains computationally efficient due to its fixed architecture and the use of backpropagation for optimization.
In contrast, the proposed model requires significantly longer training times due to Bayesian optimization, which involves multiple network training cycles for hyperparameter tuning. For the smallest dataset (50 samples), training takes approximately 40 min, while for the largest dataset (800 samples), the process extends to tens of hours. Training time also varies based on the complexity of the test case, particularly the input space dimensionality. For instance, training on an 800-sample dataset requires around 12 h for Antenna I (six design variables), approximately 24 h for Antenna II (eleven variables), and 16 h for Antenna III (seven variables). Despite the increased training time, this computational cost remains a fraction of the overall time required for data acquisition. Additionally, this trade-off is justified by the significant improvements in predictive accuracy and generalization achieved by the proposed model.
Results
Here, we report the results obtained for the considered test circuits using the proposed modeling approach and the benchmark procedures. These results are analyzed in depth in Sect. “Discussion”. Design applications and experimental validation of the circuits optimized using our surrogates for selected target operating parameters are covered in Sect. “Applications Case Studies and Experimental Validation”. Note that the test cases are challenging from the modelling perspective primarily due to wide ranges of parameters and frequency, and considerable nonlinearity of circuit’s responses.
Example I
The first verification case is a compact coupler with unequal power division ratio shown in Fig. 10113. The same figure also provides data on design variables, parameter bounds, and the target frequency range . The objective is to construct the model of scattering parameters S11, S21, S31, and S41. The sensitivity analysis has been carried out based on fifty random samples, yielding the normalized eigenvalues are λ1 = 1.00, λ2 = 0.65, λ3 = 0.51, λ4 = 0.46, λ5 = 0.37, λ6 = 0.28, which leads to the dimensionality of the reduced domain equal to Nd = 3, for which we have \({{\sqrt {\sum\nolimits_{j = 1}^{{N_{d} }} {\lambda_{j}^{2} } } } \mathord{\left/ {\vphantom {{\sqrt {\sum\nolimits_{j = 1}^{{N_{d} }} {\lambda_{j}^{2} } } } {\sqrt {\sum\nolimits_{j = 1}^{n} {\lambda_{j}^{2} } } }}} \right. \kern-0pt} {\sqrt {\sum\nolimits_{j = 1}^{n} {\lambda_{j}^{2} } } }} = 0.89\) (almost equal to the acceptance threshold Cmin = 0.9). The results are encapsulated in Table 3. The surrogate-predicted and EM-simulated scattering parameters for the proposed model built in the reduced domain with NB = 800 samples can be found in Fig. 11.
Example II
The second test case is a compact branch-line coupler with compact microstrip resonant cells (CMRCs)114 shown in Fig. 12. As for Example I, we aim to build the surrogate that represents scattering parameters S11, S21, S31, and S41. The substrate’s permittivity is an extra design parameter considered in the range from 2.0 to 5.0 so that the model covers diverse substrate materials.
FGSA run based on fifty random samples, yields the normalized eigenvalues λ1 = 1.00, λ2 = 0.66, λ3 = 0.54, λ4 = 0.48, λ5 = 0.41, λ6 = 0.39, λ7 = 0.30,λ8 = 0.25,λ9 = 0.22,λ10 = 0.16,λ11 = 0.13. The resulting domain dimensionality is Nd = 4 with \({{\sqrt {\sum\nolimits_{j = 1}^{{N_{d} }} {\lambda_{j}^{2} } } } \mathord{\left/ {\vphantom {{\sqrt {\sum\nolimits_{j = 1}^{{N_{d} }} {\lambda_{j}^{2} } } } {\sqrt {\sum\nolimits_{j = 1}^{n} {\lambda_{j}^{2} } } }}} \right. \kern-0pt} {\sqrt {\sum\nolimits_{j = 1}^{n} {\lambda_{j}^{2} } } }} = 0.88\). The results can be found in Table 4, whereas Fig. 13 shows surrogate-predicted and EM-simulated S-parameters for the metamodel rendered using NB = 800 training.
Example III
The last test example is a dual-band power divider illustrated in Fig. 14115. For this circuit, the goal is to build the metamodel representing the S-parameters: S11, S21, S22, and S32. Just as for other examples, FGSA was run using fifty random samples. The normalized eigenvalues are λ1 = 1.00, λ2 = 0.77, λ3 = 0.66, λ4 = 0.64, λ5 = 0.50, λ6 = 0.48, λ7 = 0.45. The reduced domain dimensionality is Nd = 4, which gives \({{\sqrt {\sum\nolimits_{j = 1}^{{N_{d} }} {\lambda_{j}^{2} } } } \mathord{\left/ {\vphantom {{\sqrt {\sum\nolimits_{j = 1}^{{N_{d} }} {\lambda_{j}^{2} } } } {\sqrt {\sum\nolimits_{j = 1}^{n} {\lambda_{j}^{2} } } }}} \right. \kern-0pt} {\sqrt {\sum\nolimits_{j = 1}^{n} {\lambda_{j}^{2} } } }} = 0.89\). The numerical results are provided in Table 5. Note that this is the most challenging example, therefore, an extended training set of 1600 samples was considered as well. The circuit responses at the selected test design are showcased in Fig. 15.
Discussion
The numerical data reported in Tables 3, 4, and 5 unequivocally demonstrates the competitive operation of the presented modeling methodology compared to the benchmark. For all verification structures, the RNN-based surrogate offers improved predictive power, with the advantage over comparison methods being quite significant for certain combinations of circuits and the training dataset sizes. The benefits of dimensionality reduction are also evident and contribute to dramatically improving accuracy, with RRMSE being lower than two percent for Circuit I and slightly above five percent for the most extensive training sets. This level of reliability makes the surrogates suitable for design purposes, which will be discussed in Sect. “Applications Case Studies and Experimental Validation”. The accuracy improvements are noticeable for the lowest-cardinality training sets (50 and 100 samples) but even more pronounced for NB = 400 and 800. This underscores the relevance of sequential processing of frequency characteristics leveraged by our methodology and realized using a dedicated combination of LSTM and GRU layers. It should also be noted that apart from Circuit I (the most straightforward test case), modeling in the conventional parameter space is not feasible, i.e., the values of RRMSE exceed twenty (Circuit II) and thirty percent (Circuit III), even for NB ≥ 800. It can also be observed that deep feedforward ANN models (ANN 1, 2, and 3) do not perform as well as the proposed RNN-based technique, which is indicative of the fact that just increasing the network complexity does not carry over to competitive results, such as those obtained using the proposed methodology. It appears that sequential treatment of the frequency data does have a positive impact on the results.
Another appealing feature of our framework is consistency of results. Our method has been shown superior over all benchmark techniques, for all test circuits, training set sizes, and the selection of the domain (original or reduced). In addition, it exhibits excellent scalability, i.e., fast enhancement of the predictive power as a function of the training dataset cardinality. This is particularly noticeable for a dimensionality-reduced domain. Therein, the relationship between the mean distance between the training sites scales more favorably with NB. Again, the mentioned benefits can be attributed to the specific architecture of the proposed RNN-based surrogate and the advantages of sequential processing of circuit response data (as a function of frequency). Our approach differs from standard modeling approaches (kernel-based and neural-network-based regression techniques), where the circuit characteristics are treated as vector-valued ensembles with frequency often considered a supplementary parameter.
Applications case studies and experimental validation
The surrogate models generated using the proposed methodology were employed to perform parameter tuning of Circuits I, II, and III. The exemplary design scenarios assumed for all circuits are detailed in Table 6. The table showcases the optimized geometry variable vectors obtained by directly optimizing the respective metamodels (with no further correction). The surrogated constructed with NB = 800 training samples were employed in all cases. To provide additional illustration, the optimized designs were fabricated and experimentally validated.
The photographs of the circuit prototypes and their S-parameter responses, as predicted by the surrogate, by EM simulation, and evaluated through measurements, are shown in Fig. 16. Observe a good agreement between the model predictions and EM analysis. The alignment between EM simulations and experimental results is also satisfactory. Minor discrepancies are due to manufacturing inaccuracies and the effects of the SMA connectors.
Application case studies (circuit optimization). The pictures show the circuit prototypes corresponding to metamodel-optimized designs gathered in Table 6 and their frequency characteristics: surrogate prediction (o), EM simulation (grey), and measurement (black): (a) Circuit I, (b) Circuit II, (c) Circuit III.
Conclusion
This study focused on developing an improved methodology for reliable modeling of passive microwave circuits. The approach suggested here leverages the properties of recurrent neural networks (RNNs) employed to process information enclosed in the circuit’s electrical characteristics efficiently. Our technique’s distinctive feature is handling frequency as a sequential parameter, which facilitates building a behavioral (data-driven) system representation. This contrasts with more conventional methods, where frequency responses are treated in parallel (as vector-valued entities), with the frequency often treated as an independent parameter. The proposed RNN architecture incorporates LSTM and GRU layers and the bidirectional LSTM layer to capture frequency-wise dependencies within the system’s outputs. The major hyperparameters of the network are adjusted using Bayesian Optimization (BO). Flexible dimensionality reduction is another tool employed to enhance the model’s quality significantly. It is realized using rapid global sensitivity analysis and implemented to allow arbitrary orientation of the reduced domain (spanned by the vectors associated with the maximum circuit response variations) but without eliminating individual design variables.
Our methodology has been extensively verified using several planar circuits and a range of benchmark methods. To ensure meaningful assessment, the surrogate models were constructed in conventional and dimensionality-reduced domains using training sets of sizes from 50 to 800 samples. The results unanimously demonstrate the advantages of our procedure, which turned out to be superior to all benchmark techniques regarding the surrogate’s predictive power measured using the relative root mean square error (RRMSE). These benefits are consistent for all considered test circuits, domain selection, training dataset cardinality, and the scalability of the model’s accuracy concerning the number of training samples. They also corroborate the adequacy of the assumed RNN architecture and the introduced data handling paradigm. Future work will be oriented toward further improvements of the procedure’s reliability and extending its applicability to higher-dimensional problems.
Data availability
The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.
References
ADS (Advanced Design System) Keysight Technologies Fountaingrove Parkway 1400 Santa Rosa CA 95403–1799 2023.
AWR Microwave Office Cadence Design Systems Inc. San Jose, CA 95134 USA 2023
CST Microwave Studio, https://www.3ds.com/products-services/simulia/products/cst-studio-suite/, Dassault Systems Rue Marcel Dassault 78140 Velizy-Villacoublay, France 2023.
HFSS, ANSYS, http://www.ansoft.com/products/hf/hfss/, 2600 Ansys Dr., Canonsburg PA 15317 USA 2023.
Xu, X., Yuan, T., Wu, J., Liu, J. & Du, Y. Detection of trace dibutyl phthalate based on multiresonant terahertz metamaterial sensor. IEEE Sens. J. 24(2), 1415–1423 (2024).
Herraiz, D., Esteban, H., Herraiz, D., Belenguer, A. & Boria, V. E. A novel in-line ESICL-to-ESIW transition for high-performance empty substrate integrated microwave systems. IEEE Microw. Wirel. Techn. Lett. 34(4), 363–366 (2024).
Tan, X. & Zhang, Y. A Compact rat-race coupler with widely tunable frequency- and power-dividing ratio. IEEE Microw. Wirel. Techn. Lett. 34(1), 17–20 (2024).
Wu, D.-S., Li, Y. C., Xue, Q. & Mou, J. “LTCC bandstop filters with controllable bandwidths using transmission zeros pair”, IEEE Trans. Circuits Syst. II: Express Briefs 67(6), 1034–1038 (2020).
Sen, S. & Moyra, T. “Compact microstrip low-pass filtering power divider with wide harmonic suppression”, IET Microwaves. Ant. Propag. 13(12), 2026–2031 (2019).
Dong, Y., Yang, B., Yu, Z. & Zhou, J. Robust fast electromagnetic optimization of SIW filters using model-based deviation estimation and Jacobian matrix update. IEEE Access 8, 2708–2722 (2020).
Koziel, S., Pietrenko-Dabrowska, A. & Plotka, P. Design specification management with automated decision-making for reliable optimization of miniaturized microwave components. Sc. Rep. 12, 829 (2022).
Ren, Z., He, S., Zhang, D., Zhang, Y. & Koh, C. S. A possibility-based robust optimal design algorithm in preliminary design state of electromagnetic devices. IEEE Trans. Magn. 52(3), 7001504 (2016).
Zhang, Z. et al. A surrogate modeling space definition method for efficient filter yield optimization. IEEE Microw. Wirel. Techn. Lett. 33(6), 631–634 (2023).
Cilici, F. et al. Nonintrusive machine learning-based yield recovery and performance recentering for mm-wave power amplifiers: a two-stage class-A power amplifier case study. IEEE Trans. Microw. Theory Techn. 72(5), 3046–3064 (2024).
Liu, S., Pei, C., Khan, L., Wang, H. & Tao, S. Multiobjective optimization of coding metamaterial for low-profile and broadband microwave absorber. IEEE Ant. Wirel. Propag. Lett. 23(1), 379–383 (2024).
Torun, H. M. & Swaminathan, M. High-dimensional global optimization method for high-frequency electronic design. IEEE Trans. Microw. Theory Techn. 67(6), 2128–2142 (2019).
Jiao, D. & V. C. do Nascimento,. Fast rank-revealing method for solving large global optimization problems and its applications. IEEE Trans. Microw. Theory Techn. 72(8), 4555–4567 (2024).
Koziel, S. & Pietrenko-Dabrowska, A. Rapid surrogate-aided multi-criterial optimization of compact microwave passives employing machine learning and ANNs. IEEE Trans. Microw. Theory Techn. 72(8), 4475–4488 (2024).
Luo, X., Yang, B. & Qian, H. J. Adaptive synthesis for resonator-coupled filters based on particle swarm optimization. IEEE Trans. Microw. Theory Techn. 67(2), 712–725 (2019).
Oyelade, O. N., Ezugwu, A.E.-S., Mohamed, T. I. A. & Abualigah, L. Ebola optimization search algorithm: a new nature-inspired metaheuristic optimization algorithm. IEEE Access 10, 16150–16177 (2022).
Li, X. & Luk, K. M. The grey wolf optimizer and its applications in electromagnetics. IEEE Trans. Ant. Prop. 68(3), 2186–2197 (2020).
Li, Y. & Luo, X. Adaptive synthesis using hybrid genetic algorithm and particle swarm optimization for reflectionless filter with lumped elements. IEEE Trans. Microw. Theory & Techn. 71(12), 5317–5334 (2023).
Xue, L. et al. An unsupervised microwave filter design optimization method based on a hybrid surrogate model-assisted evolutionary algorithm. IEEE Trans. Microw. Theory Techn. 71(3), 1159–1170 (2023).
Wang, J., Yang, X. S. & Wang, B. Z. Efficient gradient-based optimisation of pixel antenna with large-scale connections. IET Microw. Ant. Prop. 12(3), 385–389 (2018).
Koziel, S. & Pietrenko-Dabrowska, A. Variable-fidelity simulation models and sparse gradient updates for cost-efficient optimization of compact antenna input characteristics. Sensors 19, 8 (2019).
Pietrenko-Dabrowska, A. & Koziel, S. Computationally-efficient design optimization of antennas by accelerated gradient search with sensitivity and design change monitoring. IET Microw. Ant. Prop. 14(2), 165–170 (2020).
Zhang, W. et al. EM-centric multiphysics optimization of microwave components using parallel computational approach. IEEE Trans. Microw. Theory Techn. 68(2), 479–489 (2020).
Feng, F. et al. Coarse- and fine-mesh space mapping for EM optimization incorporating mesh deformation. IEEE Microw. Wireless Comp. Lett. 29(8), 510–512 (2019).
Pietrenko-Dabrowska, A. & Koziel, S. Response feature technology for high-frequency electronics Optimization modeling and design automation (Springer, 2023).
Zhang, C., Feng, F., Gongal-Reddy, V., Zhang, Q. J. & Bandler, J. W. Cognition-driven formulation of space mapping for equal-ripple optimization of microwave filters. IEEE Trans. Microw. Theory Techn. 63(7), 2154–2165 (2015).
Bandler, J. W. & Rayas-Sánchez, J. E. An early history of optimization technology for automated design of microwave circuits. IEEE J. Microw. 3(1), 319–337 (2023).
Wang, L. L., Yang, X. S. & Ma, C. J. An efficient gradient-based hybrid parameter-topology optimization for antenna design. IEEE Trans. Ant. Propag. 71(12), 9477–9486 (2023).
Li, J., Yang, A., Tian, C., Ye, L. & Chen, B. Multi-fidelity Bayesian algorithm for antenna optimization. J Syst. Eng. Electr. 33(6), 1119–1126 (2022).
Koziel, S. & Leifsson, L. Simulation-driven design by knowledge-based response correction techniques (Springer, 2016).
Pietrenko-Dabrowska, A. & Koziel, S. Reliable surrogate modeling of antenna input characteristics by means of domain confinement and principal components. Electronics 9(5), 1–16 (2020).
Chellappa, S., Feng, L., de la Rubia, V. & Benner, P. Inf-sup-constant-free state error estimator for model order reduction of parametric systems in electromagnetics. IEEE Trans. Microw. Theory Techn. 71(11), 4762–4777 (2023).
Zhang, J., Feng, F. & Zhang, Q.-J. Rapid yield estimation of microwave passive components using model-order reduction based neuro-transfer function models. IEEE Microw. Wirel. Comp. Lett. 31(4), 333–336 (2021).
Wu, Y., Pan, G., Lu, D. & Yu, M. Artificial neural network for dimensionality reduction and its application to microwave filters inverse modeling. IEEE Trans. Microw. Theory Techn. 70(11), 4683–4693 (2022).
Chen, W. et al. Knowledge-guided and machine-learning-assisted synthesis for series-fed microstrip antenna arrays using base element modeling. IEEE Trans. Ant. Propag. 72(2), 1497–1509 (2024).
Babale, S. A. et al. Machine learning-based optimized 3G/LTE/5G planar wideband antenna with tri-bands filtering notches. IEEE Access 12, 80669–80686 (2024).
Feng, F. et al. Adaptive feature zero assisted surrogate-based EM optimization for microwave filter design. IEEE Microw. Wirel. Comp. Lett. 29(1), 2–4 (2019).
Mahmood, M., Koc, A., Morawski, R. & Le-Ngoc, T. Achieving capacity gains in practical full-duplex massive MIMO systems: a multi-objective optimization approach using hybrid beamforming. IEEE Open J. Comm. Soc. 5, 2268–2286 (2024).
Zhang, Z., Chen, H. C. & Cheng, Q. S. Surrogate-assisted quasi-Newton enhanced global optimization of antennas based on a heuristic hypersphere sampling. IEEE Trans. Ant. Propag. 69(5), 2993–2998 (2021).
Wu, Q., Wang, H. & Hong, W. Multistage collaborative machine learning and its application to antenna modeling and optimization. IEEE Trans. Ant. Propag. 68(5), 3397–3409 (2020).
Bilson, S., Hong Loh, T., Héliot, F. & Thompson, A. Physics-informed machine learning modelling of RF-EMF exposure in massive MIMO systems. IEEE Access 12, 69410–69422 (2024).
Wu, Q., Chen, W., Yu, C., Wang, H. & Hong, W. Multilayer machine learning-assisted optimization-based robust design and its applications to antennas and array. IEEE Trans. Ant. & Propag. 69(9), 6052–6057 (2021).
Pietrenko-Dabrowska, A., Koziel, S. & Golunski, L. Cost-efficient globalized parameter optimization of microwave components through response-feature surrogates and nature-inspired metaheuristics. IEEE Access 12, 79051–79065 (2024).
Manfredi, P. Probabilistic uncertainty quantification of microwave circuits using Gaussian processes. IEEE Trans. Microw. Theory Techn. 71(6), 2360–2372 (2023).
Feng, L., Benner, P., Romano, D. & Antonini, G. Matrix-free transfer function prediction using model reduction and machine learning. IEEE Trans. Microw. Theory Techn. 70(12), 5392–5404 (2022).
Cai, J., King, J., Yu, C., Liu, J. & Sun, L. Support vector regression-based behavioral modeling technique for RF power transistors. IEEE Microw. & Wirel. Comp. Lett. 28(5), 428–430 (2018).
Dong, J., Qin, W. & Wang, M. “Fast multi-objective optimization of multi-parameter antenna structures based on improved BPNN surrogate model. IEEE Access 7, 77692–77701 (2019).
Calik, N., Belen, M., Mahouti, P. & Koziel, S. Accurate modeling of frequency selective surfaces using fully-connected regression model with automated architecture determination and parameter selection based on Bayesian optimization. IEEE Access 9, 38396–38410 (2021).
Koziel, S., Mahouti, P., Calik, N., Belen, M. A. & Szczepanski, S. Improved modeling of miniaturized microwave structures using performance-driven fully-connected regression surrogate. IEEE Access 9, 71470–71481 (2021).
Jia, Q., Li, J., Wei, C. & Liu, J. Microwave photonic reconfigurable high precision instantaneous frequency measurement system assisted by stacking ensemble learning method. J. Lightwave Techn. 41(6), 1696–1703 (2023).
Chen, W., Ji, Y., Qi, Z., Yan, L. & Zhao, X. Automatic determination of set of multivariate basis polynomials based on recursion and application of LSPCR to high-dimensional uncertainty quantification of multi-conductor transmission lines. IEEE Access 12, 18536–18544 (2024).
Pang, Y., Zhou, B. & Nie, F. “Simultaneously learning neighborship and projection matrix for supervised dimensionality reduction”. IEEE Trans Neural Netw. & Learn. Syst. 30(9), 2779–2793 (2019).
Lv, Z., Wang, L., Han, Z., Zhao, J. & Wang, W. Surrogate-assisted particle swarm optimization algorithm with Pareto active learning for expensive multi-objective optimization. IEEE/CAA J. Automatica Sinica 6(3), 838–849 (2019).
Müller, D., Soto-Rey, I. & Kramer, F. An analysis on ensemble learning optimized medical image classification with deep convolutional neural networks. IEEE Access 10, 66467–66480 (2022).
Zhang, L., Ni, Q., Zhai, M., Moreno, J. & Briso, C. An ensemble learning scheme for indoor-outdoor classification based on KPIs of LTE network. IEEE Access 7, 63057–63065 (2019).
Jin, J. et al. A novel deep neural network topology for parametric modeling of passive microwave components. IEEE Access 8, 82273–82285 (2020).
Jin, J. et al. Deep neural network technique for high-dimensional microwave modeling and applications to parameter extraction of microwave filters. IEEE Trans. Microw. Theory Techn. 67(10), 4140–4155 (2019).
Yücel, A. C., Bağcı, H. & Michielssen, E. “An ME-PC enhanced HDMR method for efficient statistical analysis of multiconductor transmission line networks”. IEEE Trans Comp. Packag. & Manuf. Techn. 5(5), 685–696 (2015).
Becerra, J. A. et al. A doubly orthogonal matching pursuit algorithm for sparse predistortion of power amplifiers. IEEE Microw. Wirel. Comp. Lett. 28(8), 726–728 (2018).
Kennedy, M. C. & O’Hagan, A. Predicting the output from complex computer code when fast approximations are available. Biometrika 87, 1–13 (2000).
Wang, F. et al. “Bayesian model fusion: large-scale performance modeling of analog and mixed-signal circuits by reusing early-stage data”. IEEE Trans Comp.-Aided Design Integr. Circuits Syst. 35(8), 1255–1268 (2016).
Jacobs, J. P. & Koziel, S. Two-stage framework for efficient gaussian process modeling of antenna input characteristics. IEEE Trans. Ant. Prop. 62(2), 706–713 (2014).
Koziel, S. & Pietrenko-Dabrowska, A. Performance-driven surrogate modeling of high-frequency structures (Springer, 2020).
Pietrenko-Dabrowska, A., Koziel, S. & Zhang, Q. J. Cost-efficient two-level modeling of microwave passives using feature-based surrogates and domain confinement. Electronics 12(17), 3560 (2023).
Koziel, S., Pietrenko-Dabrowska, A. & Leifsson, L. Improved efficacy behavioral modeling of microwave circuits through dimensionality reduction and fast global sensitivity analysis. Sci. Rep. 14, 19465 (2024).
Koziel, S., Calik, N., Mahouti, P. & Belen, M. A. Accurate modeling of antenna structures by means of domain confinement and pyramidal deep neural networks. IEEE Trans. Ant. Prop. 70(3), 2174–2188 (2022).
Koziel, S., Pietrenko-Dabrowska, A. & Ullah, U. Reduced-cost microwave modeling using constrained domains and dimensionality reduction. Sci. Rep. 13, 18509 (2023).
Fu, J. et al. Feature-assisted neural network surrogate-based multiphysics optimization for microwave filters. IEEE Microwave Wireless Techn. Lett. 34(5), 474–477 (2024).
He, Y. et al. Hybrid method of artificial neural network and simulated annealing algorithm for optimizing wideband patch antennas. IEEE Trans. Antennas Propag. 72(1), 944–949 (2024).
Touhami, A., Collardey, S. & Sharaiha, A. A global optimization method for wideband and small supergain arrays design using artificial neural network. IEEE Open J. Ant. Propag. 4, 1016–1028 (2023).
Liu, Y. et al. An efficient method for antenna design based on a self-adaptive Bayesian neural network-assisted global optimization technique optimization. IEEE Trans. Ant. Propag. 70(12), 11375–11388 (2022).
Sonker, A., Nayak, A. K., Goel, T. & Patnaik, A. Multifunctional antenna design for wireless consumer electronic devices: a soft-computing approach. IEEE Can. J. Electr. Comput. Eng. 46(2), 144–156 (2023).
Y. Liu, P. Chen, J. Tian, J. Xiao, S. Noghanian, and Q. Ye, 2024 “Hybrid ANN-GA optimization method for minimizing the coupling in MIMO antennas,” AEU – Int. J. Electronics Comm 175 155068
Javid-Hosseini, S.-H., Ghazanfarianpoor, P., Nayyeri, V. & Colantonio, P. A unified neural network-based approach to nonlinear modeling and digital predistortion of RF power amplifier. IEEE Trans. Microw. Theory Techn. 72(9), 5031–5038 (2024).
Wu, Q., Chen, W., Yu, C., Wang, H. & Hong, W. Machine-learning-assisted optimization for antenna geometry design. IEEE Trans. Ant. Propag. 72(3), 2083–2095 (2024).
Yu, C., Li, Q., Feng, F. & Zhang, Q. J. Convolutional neural network with adaptive batch-size training technique for high-dimensional inverse modeling of microwave filters. IEEE Microw. Wirels. Techn. Lett. 33(2), 122–125 (2023).
Gupta, A., Karahan, E. A., Bhat, C., Sengupta, K. & Khankhoje, U. K. Tandem neural network based design of multiband antennas. IEEE Trans. Ant. Propag. 71(8), 6308–6317 (2023).
Wang, J., Fan, C., Liao, Y. & Zhou, L. Pattern-to-absorption prediction for multilayered metamaterial absorber based on deep learning. IEEE Microwe. Wirel. Techn. Lett. 34(5), 463–466 (2024).
Stanković, Z. Ž, Olćan, D. I., Dončov, N. S. & Kolundžija, B. M. Consensus deep neural networks for antenna design and optimization. IEEE Trans. Ant. Propag. 70(7), 5015–5023 (2022).
Yasmeen, K., Mishra, K. V., Subramanyam, A. V. & Ram, S. S. Circularly polarized Fabry-Pérot cavity sensing antenna design using generative model. IEEE Sens. Lett. 7(2), 1–4 (2023).
Kouhalvandi, L. & Matekovits, L. Hyperparameter optimization of long short-term memory-based forecasting DNN for antenna modeling through stochastic methods. IEEE Ant. Wirel. Propag. Lett. 21(4), 725–729 (2022).
Tan, J., Shao, Y., Zhang, J. & Zhang, J. Efficient antenna modeling and optimization using multifidelity stacked neural network. IEEE Trans. Ant. Propag. 72(5), 4658–4663 (2024).
Li, J. et al. An adaptive evolutionary neural network-based optimization design method for wideband dual-polarized antennas. IEEE Trans. Ant. Propag. 71(10), 8165–8172 (2023).
G. Qi, S. Bao, Y. Wei, and Y. Zhang, “Accurate antenna design by deep auto-encoder surrogate model assisted particle swarm optimization,” 2022 IEEE 5th Int. Conf. electronic information comm. technology (ICEICT) Hefei China 2022 875-880.
Liu, J. P., Wang, B. Z., Chen, C. S. & Wang, R. Inverse design method for horn antennas based on knowledge-embedded physics-informed neural networks. IEEE Ant. Wirel. Propag. Lett. 23(6), 1665–1669 (2024).
Liu, Y. F., Peng, L. & Shao, W. An efficient knowledge-based artificial neural network for the design of circularly polarized 3-D-printed lens antenna. IEEE Trans. Ant. Propag. 70(7), 5007–5014 (2022).
Khan, M. R., Zekios, C. L., Bhardwaj, S. & Georgakopoulos, S. V. A deep learning convolutional neural network for antenna near-field prediction and surrogate modeling. IEEE Access 12, 39737–39747 (2024).
Su, Y. et al. Time-domain scattering parameters-based neural network inverse model for antenna designs. IEEE Ant. Wirel. Propag. Lett. 23(7), 1976–1980 (2024).
Jin, J., Su, Q., Xu, Y., He, Z. & Lu, Y. Efficient radiation pattern prediction of array antennas based on complex-valued graph neural networks. IEEE Ant. Wirel. Propag. Lett. 21(12), 2467–2471 (2022).
Tang, K., Yu, C. & Liu, Y. Cascaded neural network module for digital predistortion under various operating conditions. IEEE Microw. Wirel. Techn. Lett. 34(1), 96–98 (2024).
Liu, W. et al. Second-order sensitivity neural network modeling approach with applications to microwave devices. IEEE Trans. Microw. Theory Techn. 72(7), 3980–3992 (2024).
Faraji, A. et al. Hybrid batch-normalized deep feedforward neural network incorporating polynomial regression for high-dimensional microwave modeling. IEEE Trans. Circuits Syst. I 71(3), 1245–1258 (2024).
Li, X. R. & Zhao, Z. Evaluation of estimation algorithms part I: incomprehensive measures of performance. IEEE Trans. Aerosp. Electr. Syst. 42(4), 1340–1358 (2006).
Gorissen, D., Crombecq, K., Couckuyt, I., Dhaene, T. & Demeester, P. A surrogate modeling and adaptive sampling toolbox for computer based design. J. Mach. Learning Res.11,2051–2055(2010).
J. Snoek, H. Larochelle, R.P. Adams, 2012 Practical Bayesian optimization of machine learning algorithms. Adv. Neural Inf. Processing Syst https://doi.org/10.48550/arXiv.1206.2944
D.P. Kingma and J. Ba, “Adam: a method for stochastic optimization,”Int. Conf. Learning Representations (ICLR), Banff, AB, Canada 12–14 2014.
Morris, M. D. Factorial sampling plans for preliminary computational experiments. Technometrics 33, 161–174 (1991).
Iooss, B. & Lemaitre, P. A review on global sensitivity analysis methods. In Uncertainty management in simulation-optimization of complex systems (eds Dellino, G. & Meloni, C.) 101–122 (Springer, 2015).
Tian, W. A review of sensitivity analysis methods in building energy analysis. Renew. & Sustain. Energy Rev. 20, 411–419 (2013).
Saltelli, A. Making best use of model evaluations to compute sensitivity indices. Com. Physics. Comm. 145, 280–297 (2002).
Jansen, M. J. W. Analysis of variance designs for model output. Comp. Physics Comm. 117, 25–43 (1999).
Kovacs, I., Topa, M., Buzo, A., Rafaila, M. & Pelz, G. Comparison of sensitivity analysis methods in high-dimensional verification spaces. Acta Tecnica Napocensis. Electr. & Telecommun. 57(3), 16–23 (2016).
B. Beachkofski R. Grandhi, Improved distributed hypercube sampling American Institute of Aeronautics and Astronautics paper AIAA 2002–1274 2002
Jolliffe, I. T. Principal component analysis 2nd edn. (Springer, 2002).
Forrester, A. I. J. & Keane, A. J. Recent advances in surrogate-based optimization. Prog. Aerospace Sci. 45, 50–79 (2009).
C.K.I. Williams C.E. Rasmussen Gaussian processes for regression In D. Touretzky, M.C. Mozer, and M. Hasselmo (Eds) Advances in Neural Information Processing Systems 8 (MIT Press 1995)
M. Awad R. Khanna Support vector regression. In Efficient Learning Machines (Apress, Berkeley, CA, 2015)
Vang-Mata, R. Multilayer perceptrons (Nova Science Pub Inc., 2020).
S. Koziel and A.T. Sigurdsson, “Performance-driven modeling of compact couplers in restricted domains Int. J. RF & Microwave CAE, 28 6 2018
Tseng, C. & Chang, C. A rigorous design methodology for compact planar branch-line and rat-race couplers with asymmetrical T-structures. IEEE Trans. Microw. Theory Techn. 60(7), 2085–2092 (2012).
Lin, Z. & Chu, Q.-X. A novel approach to the design of dual-band power divider with variable power dividing ratio based on coupled-lines. Prog. Electromagn. Res. 103, 271–284 (2010).
Acknowledgements
The authors thank Dassault systemes, France, for making CST Microwave Studio available. This work was partly supported by the Icelandic Research Fund Grant 217771 and by the National Science Centre of Poland Grant 2022/47/B/ST7/00072.
Author information
Authors and Affiliations
Contributions
Conceptualization, K.S., S.K., methodology, K.S., S.K. and A.P.; data generation, S.K.; investigation, K.S. and S.K.; writing—original draft preparation, S.K., K.S., and A.P.; writing—review and editing, S.K. and A.P.; visualization, K.S. and S.K.; supervision, S.K.; project administration, S.K. and A.P.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Sahu, K.C., Koziel, S. & Pietrenko-Dabrowska, A. Surrogate modeling of passive microwave circuits using recurrent neural networks and domain confinement. Sci Rep 15, 13322 (2025). https://doi.org/10.1038/s41598-025-91643-3
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-025-91643-3


















