Introduction

Accurate temperature regulation in continuous stirred-tank heater (CSTH) systems is fundamental in chemical and thermal process industries, where even minor deviations can result in energy inefficiency, off-spec products, or safety risks1. The CSTH’s strong nonlinearities, parameter uncertainties, and susceptibility to external disturbances make it one of the most challenging benchmarks in process control. Among various strategies, the proportional–integral–derivative (PID) controller remains the most widely implemented due to its structural simplicity, reliability, and compatibility with industrial environments2,3. However, the controller’s performance depends critically on appropriate tuning of the proportional, integral, and derivative gains. Improper tuning often leads to oscillatory responses, overshoot, or instability, especially in nonlinear and time-varying systems4,5,6,7.

Traditional analytical approaches (such as Ziegler–Nichols, Cohen–Coon, and other model-based tuning rules) provide acceptable performance for linear systems but are often inadequate for highly nonlinear and parameter-sensitive plants like CSTH8,9. Consequently, research has shifted toward intelligent and metaheuristic optimization-based tuning methods, which can adaptively and efficiently search for optimal controller parameters without relying on analytical gradients10,11.

Over the last decade, a variety of bio-inspired and physics-based metaheuristic algorithms (including differential evolution (DE), particle swarm optimization (PSO), black hole optimization (BHO), Jaya optimization (JOA), and sunflower optimization (SFO)) have been applied for PID and fractional-order PID (FOPID) controller tuning12,13,14,15,16,17. For example, Basil and Marhoon16 demonstrated that FOPID tuning using BHO, JOA, and SFO improved robustness and trajectory tracking in NASA’s X-15 adaptive flight control system. Likewise, hybrid and multi-strategy algorithms such as PSO–ALO, HAOAROA, and FOPID–TID have achieved remarkable results in UAV and multi-rotor flight control, balancing disturbance rejection and dynamic stability better than single-algorithm approaches18,19,20. These advances illustrate the importance of combining diversity-preserving mechanisms with global exploration strategies to enhance convergence reliability in complex nonlinear problems.

Parallel developments in CSTH control have also expanded beyond traditional tuning. Mahmood and Nawaf21 reported that cascade control structures can suppress oscillations and accelerate settling, while Haghani et al.22 and Wang et al.23 proposed probabilistic and manifold-based optimization frameworks for robust fault isolation and performance monitoring. Predictive and adaptive control methods, such as broad learning-based predictive control24 and adaptive subspace predictive control (ASPC)25, have improved convergence and adaptability under varying process conditions. Neural network26 and reinforcement learning27 controllers further enhanced flexibility and disturbance tolerance. Meanwhile, the theoretical work of Do28 on systems driven by Lévy processes provided a strong mathematical foundation for employing Lévy-flight perturbations in control-oriented optimization.

Despite these advancements, achieving a simultaneous balance between global exploration and local exploitation remains a persistent challenge. Algorithms such as PSO, DE, SMA, GEO, HO, and NRBO have yielded performance gains but still encounter premature convergence, population collapse, or stagnation during optimization in complex nonlinear search landscapes. Moreover, deterministic opposition-based strategies and conventional trap-avoidance mechanisms do not always ensure sufficient diversity preservation or convergence reliability in real-world control problems.

To overcome these challenges, this study introduces a modified Newton–Raphson-based optimization (mNRBO) algorithm designed to enhance both optimization robustness and control performance. The original Newton-Raphson-based optimizer (NRBO)29 integrates Newton–Raphson search logic with a trap-avoidance mechanism to accelerate convergence and improve solution stability The proposed algorithm extends the standard NRBO29 by incorporating two synergistic strategies:

  • Random opposition learning (ROL)30 to periodically introduce stochastic diversity and prevent population stagnation, and.

  • Lévy-flight-based guided learning (LGL)28 to provide adaptive long-jump perturbations for improved global exploration and efficient escape from local optima.

By embedding these mechanisms into the Newton–Raphson search logic, mNRBO achieves a well-balanced exploration–exploitation trade-off, enabling faster convergence and superior global search ability.

Before being applied to the CSTH control problem, the proposed mNRBO was first validated on a comprehensive set of classical benchmark functions, encompassing unimodal, multimodal, and composite test functions. These benchmarks were used to evaluate convergence speed, stability, precision, and global optimization capability against several state-of-the-art algorithms such as NRBO29, slime mould algorithm (SMA)31, golden eagle optimizer (GEO)32, and hippopotamus optimization algorithm (HO)33. The results confirmed that mNRBO consistently achieved the lowest mean error and standard deviation values, establishing its robustness and versatility as a high-performance optimizer suitable for complex nonlinear systems.

Subsequently, the algorithm was employed to optimize the PID controller parameters for a nonlinear CSTH process modeled through mass–energy balance equations. A multi-objective cost function integrating rise time, settling time, overshoot, and steady-state error was used for evaluation under realistic constraints. Comparative results with competing algorithms demonstrated that mNRBO achieved superior dynamic performance, with lower cost function values, faster transient response, and improved disturbance rejection. Quantitative metrics including recovery time, peak deviation, and integral absolute error further confirmed its robustness against parameter variations and external perturbations.

The major novelties and contributions of this study can be summarized as follows:

  1. 1.

    Proposal of a novel modified Newton–Raphson-based optimization (mNRBO) algorithm integrating ROL and LGL to enhance exploration–exploitation balance.

  2. 2.

    Comprehensive benchmarking of mNRBO against leading metaheuristic algorithms across standard test functions, demonstrating improved global search precision and convergence stability.

  3. 3.

    Application of mNRBO for PID tuning in nonlinear CSTH temperature control, achieving significant performance improvements in rise time, settling time, overshoot, and steady-state error.

  4. 4.

    Ablation and statistical significance analyses confirming the individual and synergistic effects of ROL and LGL via Wilcoxon’s non-parametric test.

  5. 5.

    Quantitative disturbance evaluation using recovery time, peak deviation, and integral of absolute error metrics, evidencing superior disturbance rejection and steady-state accuracy under uncertainties.

  6. 6.

    Industrial relevance and practicality, as the method maintains a standard PID structure compatible with existing control hardware, ensuring real-world implementability.

The remainder of this paper is organized as follows. Section “Overview of NRBO and proposed mNRBO algorithms” details the NRBO framework and the proposed modifications through ROL and LGL. Section “Dynamic modeling of the CSTH-based temperature control system” presents the dynamic modeling of the CSTH system. Section “Proposed control methodology” describes the application of mNRBO for PID tuning. Section “Results and discussion” discusses the comparative and statistical results, including benchmark analyses and disturbance rejection performance, and Section “Conclusion and future directions” concludes with key findings and directions for future work.

Overview of NRBO and proposed mNRBO algorithms

The basics of the NRBO

The NRBO represents a novel approach to solving complex optimization problems by harmonizing classical Newton-Raphson concepts with population-based stochastic search techniques29. Unlike conventional methods that rely purely on derivative-driven refinements, the NRBO introduces a hybrid strategy that enhances search efficiency, balances exploration and exploitation, and mitigates premature convergence. At the core of NRBO lies an iterative search mechanism that begins by estimating optimal solutions using Newton-Raphson’s gradient-based logic. This mechanism is enriched with probabilistic perturbations and diversity-preserving dynamics to effectively traverse the solution space. The algorithm operates on three key pillars: a Newton-inspired refinement strategy that leverages directional gradients, a stochastic variation mechanism that ensures widespread search diversity, and a trap avoidance strategy (TAS) designed to steer agents away from local optima.

The optimization challenge is framed as the minimization of a target objective function subject to bounded constraints. Each decision variable is confined within a defined interval [\({LB}^{j}\), \({UB}^{j}\)], ensuring feasibility across all dimensions \(j=\text{1,2},\ldots,Dim\). This constraint structure reflects real-world limitations, where variable ranges are typically dictated by practical or physical considerations.

A population of candidate solutions, denoted by \(Np\)​, is initialized randomly within the allowable bounds. Each agent in the population is described by a decision vector \({Y}_{j}^{n}\)​, where nnn indexes the individual and jjj the variable dimension. The initialization is given by:

$${Y}_{j}^{n}\left(Iter=0\right)={LB}^{j}+\left(\left({UB}^{j}-{LB}^{j}\right)\times rand\right), \quad n=1:Np,\quad j=1:Dim$$
(1)

During each iteration, the objective function is evaluated for all agents, yielding fitness values that are used to identify the best-performing solution \({Y}^{BEST}\) and the poorest one \({Y}^{WORST}\). These agents guide the adaptation process for the rest of the population.

The core update mechanism, termed the Newton-Raphson Search Rule (\(NRSR\)), revises each solution vector by computing a directionally informed shift based on differences among selected individuals. For two reference agents \({Y}_{1}\)​ and \({Y}_{2}\)​, the \(NRSR\) produces a step vector as:

$$NRSR({Y}_{1},{Y}_{2})=randn\times \left(\frac{\left({Y}_{1}-{Y}_{2}\right)\times \varDelta{Y}_{12}}{2\times\left({Y}_{1}+{Y}_{2}-2\times{Y}_{n}\right)}\right)$$
(2)

where the differential component \(\varDelta{Y}_{12}\)​ is defined by:

$$\varDelta{Y}_{12}=rand\times|{Y}_{1}-{Y}_{2}|$$
(3)

The updated positions of agents are then computed through a combination of three dynamically generated vectors \({Y}_{1}\)​, \({Y}_{2}\)​, and \({Y}_{3}\)​, blended using a probabilistic coefficient \(R{d}_{1}\in\left[\text{0,1}\right]\):

$${Y}_{n}^{j}(Iter+1)=R{d}_{1}\times\left(R{d}_{1}\times{{Y}_{1}}_{n}^{j}\left(Iter\right)+\left(1-R{d}_{1}\right)\times{{Y}_{2}}_{n}^{j}\left(Iter\right)\right)+\left(\right(1-R{d}_{1})\times{{Y}_{3}}_{n}^{j}(Iter\left)\right),\quad n=1:Np,\quad j=1:Dim$$
(4)

Each component vector is constructed by incorporating gradient-derived shifts, elite guidance, and inter-agent differentials. The vectors are defined as:

$${{Y}_{1}}_{n}\left(Iter\right)={Y}_{n}\left(Iter\right)-NRSR\left({Z}_{w},{Z}_{v}\right)+a\text{*}\left({Y}^{BEST}-{Y}_{n}\left(Iter\right)\right)+b\text{*}({Y}_{R1}\left(Iter\right)-{Y}_{R2}\left(Iter\right))$$
(5)
$${{Y}_{2}}_{n}\left(Iter\right)={Y}^{BEST}-NRSR\left({Z}_{w},{Z}_{v}\right)+a\text{*}\left({Y}^{BEST}-{Y}_{n}\left(Iter\right)\right)+b\text{*}({Y}_{R1}\left(Iter\right)-{Y}_{R2}\left(Iter\right))$$
(6)

Here, \({Y}_{R1}\)​ and \({Y}_{R2}\)​ are randomly selected individuals, while aaa and bbb are tuning parameters controlling exploration and exploitation. The intermediate reference points \({Z}_{w}\)​ and \({Z}_{v}\)​ are calculated as:

$${Z}_{w}=R{d}_{2}\text{*}\left(\left(Mean\left({Y}_{n}-randn\text{*}NRSR\left({Y}^{WORST},{Y}^{BEST}\right),{Y}_{n}\right)\right)+R{d}_{2}\text{*}\varDelta Y\right)$$
(7)
$${Z}_{v}=R{d}_{2}\text{*}\left(\left(Mean\left({Y}_{n}-randn\text{*}NRSR\left({Y}^{WORST},{Y}^{BEST}\right),{Y}_{n}\right)\right)-R{d}_{2}\text{*}\varDelta Y\right)$$
(8)

The third update component \({{Y}_{3}}_{n}\)​ introduces iteration-based scaling to support adaptive transition from exploration to exploitation:

$${{Y}_{3}}_{n}\left(Iter\right)={Y}_{n}\left(Iter\right)+\left(2\text{*}\frac{Iter}{Ma{x}_{iter}}\right)-1+a\text{*}\left({{Y}_{2}}_{n}\left(Iter\right)-{{Y}_{1}}_{n}\left(Iter\right)\right)$$
(9)

To safeguard the search process against stagnation in local optima, NRBO introduces the Trap Avoidance Strategy (TAS). This mechanism generates alternative solutions depending on a selection condition involving a binary variable \(\beta\), determined by:

$$\beta=\left\{\begin{array}{ll}1& if\quad R{d}_{3}>0.5\\ 0& otherwise\end{array}\right.$$
(10)

The adaptive coefficients \({\mu}_{a}\) and \({\mu}_{b}\)​ depend on \(\beta\) and a random variable \(R{d}_{3}\)​, providing additional variability:

$${\mu}_{a}=\left(1-\beta\right)+3\times\beta\times R{d}_{3}$$
(11)
$${\mu}_{b}=\left(1-\beta\right)+\beta\times R{d}_{3}$$
(12)

Two offset components are then computed as:

$$XA\left(Iter\right)={\theta}_{1}\times\left(\left({\mu}_{a}\times{Y}^{BEST}\right)-\left({\mu}_{b}\times{Y}_{n}\left(Iter\right)\right)\right)$$
(13)
$$XB\left(Iter\right)=\left(\frac{{\theta}_{2}\times2\times Iter}{Ma{x}_{Iter}}-{\theta}_{2}\right)\times \left(\left({\mu}_{a}\times Mean\left({Y}_{n}\right)\right)-\left({\mu}_{b}\times{Y}_{n}\left(Iter\right)\right)\right)$$
(14)

The TAS-based update rule is:

$${Y}_{TAS}\left(Iter\right)=\left\{\begin{array}{ll}{Y}_{n}\left(Iter+1\right)+XA\left(Iter\right)+XB\left(Iter\right),& if\quad {\mu}_{a}<0.5\\ {Y}^{BEST}+XA\left(Iter\right)+XB\left(IteR\right),& otherwise\end{array}\right.$$
(15)

The complete NRBO procedure can be outlined in the following sequence:

  1. 1.

    Initialization: Define all algorithm parameters including population size, boundaries, dimensionality, and maximum iteration count. Set the weighting constants​, and randomness coefficients.

  2. 2.

    Population generation: Randomly initialize each search agent within the prescribed bounds.

  3. 3.

    Fitness evaluation: Assess the fitness of all agents using the target objective function.

  4. 4.

    Ranking: Identify the best and worst agents based on their fitness values.

  5. 5.

    Search phase: Use a random number to switch between the NRSR-based update and TAS-enhanced update.

  6. 6.

    Position update: Apply either the Newton-inspired update or the trap avoidance strategy to generate new candidate solutions.

  7. 7.

    Iteration: Repeat the ranking, updating, and evaluation steps until the maximum number of iterations is completed.

  8. 8.

    Termination: Output the best solution found as the final outcome of the optimization process.

This adaptive structure allows NRBO to effectively navigate complex search landscapes, striking a balance between convergence speed and global search robustness.

The proposed mNRBO algorithm

The original NRBO employs two search mechanisms: the Newton-Raphson Search Rule (NRSR) and the trap avoidance strategy (TAS). Both rely heavily on the best search agent; however, if this agent converges prematurely to a local optimum, the population quickly loses diversity and stagnates. To overcome these drawbacks, the modified NRBO (mNRBO) incorporates two additional strategies: random opposition learning (ROL) and Lévy-flight-based guided learning (LGL). ROL extends the concept of opposition-based learning30 by introducing a random factor when generating opposite solutions. For a candidate \({x}_{j}\) bounded within \([{L}_{j},{U}_{j}]\), its random opposite is defined as:

$${x}_{j}^{ROL}={L}_{j}+{U}_{j}-{x}_{j}+r \,({U}_{j}-{L}_{j})$$
(16)

where \(r\in\left[\text{0,1}\right]\) is a random number. This ensures that the population explores new regions rather than deterministically revisiting mirrored solutions. In the mNRBO cycle, ROL is invoked periodically to reintroduce diversity and prevent premature convergence.

LGL combines guided exploitation with Lévy-flight exploration28. The new position of an agent is updated as:

$${x}^{LGL}=x+\alpha\cdot\text{Levy}\left(\beta\right)\, ({x}_{best}-x)$$
(17)

where \({x}_{best}\)is the current best solution, \(\alpha\)is a scaling parameter, and \(\text{Levy}\left(\beta\right)\)denotes a Lévy-distributed step length (\(1<\beta \le 3\)). This formulation allows frequent small adjustments around promising regions while occasionally performing long jumps, helping the optimizer escape local minima. Within the mNRBO framework (Fig. 1), ROL enhances the exploration stage by generating diverse candidate positions, while LGL strengthens the exploitation stage through adaptive jumps around the best solution. Together, these strategies maintain a robust balance between exploration and exploitation, enabling the optimizer to converge faster and with greater reliability compared to the standard NRBO‎.

Fig. 1
Fig. 1
Full size image

Flowchart of the proposed mNRBO.

The effects of the improvement strategies

To better understand the individual and combined contributions of the proposed strategies, an ablation analysis was conducted. Figure 2 presents the convergence behavior of four variants: the baseline NRBO, NRBO enhanced with ROL (NRBO + ROL), NRBO combined with LGL (NRBO + LGL), and the complete mNRBO integrating both mechanisms. The curve corresponding to NRBO shows the expected convergence trend but with signs of stagnation, as the search agents tend to cluster around local optima. This behavior highlights the limited ability of the baseline algorithm to maintain diversity once an early promising region is identified.

Fig. 2
Fig. 2
Full size image

Effects of different improvement strategies.

When ROL is incorporated, the resulting NRBO + ROL variant achieves noticeably broader exploration. The random generation of opposite solutions periodically disperses agents into new regions of the search space, reducing the risk of premature convergence.

However, while this diversity is maintained longer than in standard NRBO, the absence of a strong exploitation mechanism results in slower refinement near the optimal region. In contrast, NRBO + LGL emphasizes exploitation.

The Lévy-flight perturbations, guided toward the best solution, enable more intensive local search and facilitate occasional jumps that help avoid deceptive minima. Although this variant converges faster than NRBO, the lack of continuous diversity restoration sometimes causes the population to collapse prematurely, particularly in high-dimensional or irregular landscapes.

The full mNRBO, which integrates both ROL and LGL, demonstrates the most balanced performance. As shown in Fig. 2, the algorithm maintains a steady decrease in objective function values with fewer fluctuations, indicating an effective interplay between exploration and exploitation. ROL ensures that diversity is preserved throughout the process, while LGL accelerates convergence around promising areas without risking stagnation. This complementary interaction explains why the full mNRBO achieves lower final cost values and greater stability compared with all intermediate variants. In summary, the ablation results confirm that each enhancement contributes uniquely: ROL strengthens exploration, LGL reinforces exploitation, and their integration yields the robust search dynamics necessary for tackling complex nonlinear control problems. The findings underscore that the superior performance of mNRBO is not due to either mechanism in isolation but rather to their synergistic effect within the NRBO framework.

Exploration-exploitation analysis

The effectiveness of any metaheuristic algorithm depends on how well it balances exploration of the search space with exploitation of promising regions. Figure 3 illustrates this trade-off for the proposed mNRBO in comparison with the original NRBO and other benchmark algorithms. As shown in the figure, the exploration stage of mNRBO is strengthened by the introduction of ROL. This mechanism periodically generates randomized opposite solutions, which broadens the search space and prevents the population from clustering too early around local optima. Consequently, mNRBO sustains higher population diversity during the early and mid-iterations, enabling the algorithm to sample from wider regions of the solution landscape. On the other hand, the LGL mechanism enhances exploitation in later iterations. By allowing small local refinements interspersed with occasional long jumps, LGL ensures that the algorithm not only intensifies the search around the best solution but also maintains the ability to escape deceptive minima. This dual behavior explains the smoother convergence trend observed in Fig. 3, where mNRBO consistently avoids stagnation and progresses toward the global optimum more reliably.

Fig. 3
Fig. 3
Full size image

Exploration and exploitation analysis of mNRBO.

The combined impact of ROL and LGL is reflected in the convergence trajectories. While other algorithms exhibit fluctuations or premature plateaus, mNRBO demonstrates a steady decline in objective function values, indicating both effective exploration and efficient exploitation. This explains why, in later sections, the algorithm achieves faster convergence, lower cost values, and superior closed-loop control performance. Ultimately, Fig. 3 confirms that the strength of mNRBO lies not merely in achieving improved numerical results but in the mechanistic balance it maintains between global search and local refinement, a balance that is critical for solving complex nonlinear control problems.

Dynamic modeling of the CSTH-based temperature control system

The CSTH system is widely employed in industrial processes requiring precise thermal regulation. This system ensures consistent heating by utilizing a jacketed structure to manage the temperature of the fluid contained within the tank. The CSTH’s operation encompasses complex interactions among fluid flow, heat exchange, and control mechanisms. These interactions are often described by nonlinear dynamic models, making accurate system modeling and parameter tuning essential for maintaining energy efficiency and reliable performance. A general overview of the CSTH configuration is depicted in Fig. 434.

Fig. 4
Fig. 4
Full size image

Simplified diagram of CSTH temperature control system.

Governing physical principles

The CSTH system operates based on fundamental conservation laws involving mass and energy, alongside heat transfer mechanisms.

Mass balance section

The conservation of mass across the system can be described through a mass balance equation:

$$\frac{dm}{dt}={\rho}_{f}{Q}_{in}-{\rho}_{f}{Q}_{out}$$
(18)

In the mass balance equation, \(m\) denotes the mass of the fluid within the tank, while \({\rho}_{f}\) epresents the fluid density. The terms \({Q}_{in}\)​ and \({Q}_{out}\) correspond to the inlet and outlet flow rates, respectively.

Energy balance section

To describe thermal dynamics within the tank, an energy balance equation is employed. It accounts for heat exchange due to both conduction and fluid inflow/outflow:

$$\frac{d\left(T\right)}{dt}=\frac{1}{\rho{C}_{P}V}[\rho{C}_{p}{Q}_{in}\left({T}_{in}-T\right)+UA\left({T}_{j}-T\right)]$$
(19)

where, \(T\), \({T}_{in}\), \({T}_{j}\), \({C}_{p}\), \(U\), \(A\) and \(V\) are tank fluid temperature, inlet fluid temperature, temperature of the jacket fluid, specific heat capacity, overall heat transfer coefficient, heat transfer area and volume of the tank respectively.

Heat transfer equation

The heat transfer rate from the jacket to the tank fluid is quantified using:

$${Q}_{h}=UA\left({T}_{j}-T\right)$$
(20)

This equation highlights the direct relationship between heat transfer efficiency and the temperature difference across the heat exchange surface.

Transfer function derivation for CSTH system

To analyze the system’s dynamic response, the Laplace transform is applied to the energy balance equation. Rearranging the equation yields the transfer function between heater input and tank temperature:

$$\frac{T\left(s\right)}{Qh\left(s\right)}=\frac{K}{\tau s+1}$$
(21)

where, \(\tau\) and \(K\) are system time constant and process gain respectivly. Additionally, the influence of inlet temperature on tank temperature can also be modeled:

$$\frac{T\left(s\right)}{{T}_{in}\left(s\right)}=\frac{\acute{K}}{\acute{\tau}s+1}$$
(22)

These transfer functions serve as linear approximations to describe the CSTH’s temperature dynamics under small perturbations.

Parameter configuration of the CSTH system

Accurate modeling of the CSTH system requires a set of physical, thermal, and operating parameters that define its dynamic behavior. To maintain clarity and readability, only the key parameters relevant for mass and energy balance modeling and controller design are reported in the main text. Table 1 summarizes these essential parameters, including fluid properties, main geometrical attributes, heat transfer characteristics, and control setpoints that directly affect temperature regulation. These parameters capture the dominant thermal dynamics and provide the foundation for simulation and controller design. For reproducibility, a comprehensive list of all structural dimensions, transmitter gains, valve coefficients, hydraulic resistances, pump pressures, and auxiliary constants used for high-fidelity modeling is provided in Appendix A (Supplementary Material).

Table 1 Definition of parameters in CSTH system34.

Proposed control methodology

Standard PID controller

The PID controller is one of the most established and widely applied control strategies in industrial automation and process control due to its straightforward structure and reliable performance35. The standard form of the PID controller in the Laplace domain is given as:

$$\frac{U\left(s\right)}{E\left(s\right)}={K}_{P}\left(1+\frac{1}{{T}_{I}s}+\frac{{T}_{D}s}{\alpha{T}_{D}s+1}\right)$$
(23)

where,\(U\left(s\right)\), \(E\left(s\right)\), \({K}_{P}\), \({T}_{I}\) and \({T}_{D}\) are the controller output in the Laplace domain, the error signal (difference between the reference input and system output), the proportional gain, the integral time constant and the derivative time constant respectivly. In addition, \(\alpha\) is a small positive constant \((0<\alpha<1)\) used to implement a low-pass filter on the derivative term. This formulation enhances the classical PID controller by incorporating a first-order filter on the derivative action, which is crucial for attenuating high-frequency measurement noise and improving the robustness of the controller in practical applications. The proportional term \({K}_{P}\) provides an immediate response to the error, the integral term ​\(\frac{1}{{T}_{I}s}\) accumulates the past error to eliminate steady state offset, and the derivative term \(\frac{{T}_{D}s}{\alpha{T}_{D}s+1}\)​ predicts future error trends while filtering out high-frequency noise. The effectiveness of this controller lies in its ability to balance responsiveness, stability, and steady-state accuracy. The block diagram of the standard PID controller is depicted in Fig. 5, illustrating how each component contributes to the overall control loop.

Fig. 5
Fig. 5
Full size image

Block diagram of standard PID controller.

Objective function and application of proposed mNRBO

To evaluate the performance of the proposed mNRBO algorithm in controlling the highly nonlinear CSTH system, a step change in the setpoint (\({Y}_{T,sp}\left(t\right)\)) was introduced. Specifically, the reference value was increased from 76 to 81% at \(t=100\,\hbox{s}\), simulating a realistic operating scenario with abrupt variations in desired output.

A modified version of the cost function from14 was used to assess the transient and steady-state response of the system. The objective function is defined as:

$${F}_{cost}=\left(1-\phi\right)\left(\frac{{N}_{po}+{N}_{pse}}{100}\right)+\phi({N}_{st}-{N}_{rt})$$
(24)

where, \({N}_{rt}\),\({N}_{st}\), \({N}_{po}\) and \({N}_{pse}\) are normalized rise time (10–90%), normalized settling time (\(2\%\) tolerance region), normalized percent overshoot and normalized percent steady-state error at the final time \({t}_{f}=1000\) s. In addition, the scaling factor is \(\phi={e}^{-1}=0.3679\). This cost function ensures a trade-off between dynamic response (rise and settling time) and output accuracy (overshoot and steady-state error). The value of \(\phi\) can be adjusted to emphasize either time-domain dynamics or final value accuracy, depending on the specific application requirements. In this study, a value of \(\phi=0.3679\) was chosen to provide balanced performance. so that no single metric dominates the optimization, thereby achieving a balanced trade-off between fast response, low overshoot, and minimal steady-state error. To perform optimization using mNRBO, the search was bounded by the following constraints on the PID parameters and the filter coefficient:

$$\left\{ {\begin{array}{l} {K_{P}^{{min}} \le K_{P} \le K_{P}^{{max}} } \\ {T_{I}^{{min}} \le T_{I} \le T_{I}^{{max}} } \\ {T_{D}^{{min}} \le T_{D} \le T_{D}^{{max}} } \\ {\alpha ^{{min}} \le \alpha \le \alpha ^{{max}} } \\ \end{array} } \right.$$
(25)

These boundaries ensure that the optimized parameters remain within practical and physically meaningful limits, preventing instability or excessive actuator effort. The mNRBO algorithm was executed iteratively to minimize \({F}_{cost}\)​, yielding the optimal tuning values for the PID controller under the defined operational constraints. Figure 6 illustrates the proposed control methodology applied to the CSTH system, showcasing how the mNRBO algorithm integrates with the feedback control loop to enhance system performance.

Fig. 6
Fig. 6
Full size image

The proposed control methodology for highly nonlinear CSTH system.

Results and discussion

Comparisons with recently reported metaheuristic algorithms

In order to evaluate the effectiveness and robustness of the proposed mNRBO algorithm, a comparative study was conducted against four state-of-the-art metaheuristic optimization techniques: NRBO29, HO33, GEO32, and SMA31. All algorithms were implemented using default parameter settings, with a population size of 20, over 30 independent runs, and each run comprising 50 iterations.

Table 2 presents the statistical metrics of the cost function obtained through the five algorithms. The proposed mNRBO algorithm outperformed the other methods in terms of achieving the minimum, average, and standard deviation values of the cost function. Specifically, mNRBO attained the lowest minimum cost value of 53.2881, compared to 55.7018 by NRBO, 56.1413 by HO, 58.4120 by GEO, and 57.4850 by SMA. Moreover, mNRBO exhibited the lowest standard deviation (0.9038), indicating superior stability and convergence consistency across multiple runs. The statistical significance of mNRBO’s improvement is confirmed by the p-values, all below 0.05 when compared to the other algorithms, highlighting a statistically significant enhancement in performance. It is worth noting that all p-values were obtained using the Wilcoxon signed-rank non-parametric test, which does not assume normality and is widely recommended for the statistical comparison of stochastic optimization algorithms36,37. This test ensures that the reported significance values reliably capture the performance differences among algorithms across multiple independent runs.

Table 2 Statistical metrics of cost function for mNRBO, NRBO, HO, GEO and SMA.

Figure 7 illustrates the evolution of PID parameters \(({K}_{P},{T}_{I},{T}_{D}\,and\,\alpha)\) across iterations for all algorithms. The results show that mNRBO maintains smoother and more stable parameter trajectories, converging earlier and with less fluctuation. This indicates the algorithm’s strong capability in effectively exploring and exploiting the search space. Correspondingly, Fig. 8 shows the variation of the cost function with respect to the number of iterations. Again, mNRBO demonstrates faster convergence to a lower cost value, reinforcing its robustness in achieving optimal solutions with fewer iterations.

Fig. 7
Fig. 7
Full size image

Variation of standard PID controller parameters (\({K}_{P}\), \({T}_{I}\), \({T}_{D}\) and \(\alpha\)) with respect to iterations via mNRBO, NRBO, HO, GEO and SMA.

Fig. 8
Fig. 8
Full size image

Variation of cost function (\({F}_{cost}\)) with respect to iterations via mNRBO, NRBO, HO, GEO and SMA.

Table 3 summarizes the final optimal PID parameters obtained through each algorithm along with their respective search ranges. The mNRBO algorithm delivered optimized gains of \({K}_{P}=6.6595\), \({T}_{I}=138.6948\), \({T}_{D}=27.7611\)while also resulting in the smallest cost value. These parameters lie well within the defined bounds and reflect a precise tuning capability of the mNRBO method. Furthermore, the cost function value achieved by mNRBO is notably lower than those derived from NRBO, HO, GEO, and SMA, emphasizing the enhanced control accuracy enabled by the proposed method.

Table 3 Ranges and optimal parameters of standard PID controller.

The impact of these optimally tuned parameters is further validated through time-domain simulations of the system response. Figure 9 displays the temperature response of the tank under PID control for each algorithm, with mNRBO providing the fastest and most stable response. A magnified view in Fig. 10 confirms the improved settling and tracking behavior achieved by the proposed approach. Table 4 quantitatively supports these observations by reporting key stability metrics, including normalized rise time \(\left({N}_{rt}\right)\), normalized settling time \(\left({N}_{st}\right)\), normalized percent overshoot \(\left({N}_{po}\right)\), and normalized percent steady-state error \(\left({N}_{pse}\right)\). The mNRBO method yields the best performance across all metrics, with the shortest rise and settling times (62.0506 s and 206.8783 s respectively), and the lowest overshoot (1.4098%) and steady-state error (0.0060%).

Fig. 9
Fig. 9
Full size image

mNRBO, NRBO, HO, GEO and SMA based time responses for tank temperature sensor.

Fig. 10
Fig. 10
Full size image

Zoom view of Fig. 9.

Table 4 Numerical values for time responses obtained via mNRBO, NRBO, HO, GEO and SMA.

These results demonstrate that the mNRBO algorithm not only improves convergence efficiency and solution accuracy in the PID tuning process but also enhances overall system performance. Compared to the other benchmark algorithms, mNRBO consistently provides superior results in both statistical metrics and control response, validating its potential as a reliable and powerful optimization strategy for complex control systems.

Performance evaluation of mNRBO-tuned PID controller in terms of setpoint tracking and disturbance rejection

To assess the effectiveness of the proposed mNRBO algorithm in tuning PID controller parameters, its performance was evaluated under scenarios involving both variable setpoints and disturbances. The tracking ability of the controller in response to a dynamically changing reference signal was first examined. As illustrated in Fig. 11, the mNRBO-tuned controller successfully maintained precise tracking across a range of setpoints, demonstrating minimal overshoot, fast rise time, and negligible steady-state error. This confirms the robustness and adaptability of the proposed method in maintaining optimal performance under time-varying reference conditions.

Fig. 11
Fig. 11
Full size image

Performance of proposed mNRBO for variable setpoint.

The system’s disturbance rejection performance was evaluated through testing different disturbance types against the controller. As shown in Fig. 12, two categories of disturbances were considered. Figure 13 demonstrates how the mNRBO-based tuning method improves controller resilience through rapid recovery and stable operation free of oscillatory behavior when facing disturbances. The controller maintained system stability and kept the output near the desired trajectory despite the presence of parameter uncertainties. A detailed assessment confirms that the mNRBO-tuned PID controller demonstrates exceptional performance in handling setpoint adjustments and unexpected disruptions which makes it suitable for application in unpredictable settings.

Fig. 12
Fig. 12
Full size image

Type of disturbances (a) change of \({Q}_{i}\), (b) change of \({T}_{ci}\).

Fig. 13
Fig. 13
Full size image

Performance of the proposed mNRBO with respect to disturbance and parameter changes.

In addition to the qualitative assessment illustrated in Figs. 9, 10, 11, 12 and 13, quantitative stability metrics were computed to further evaluate the disturbance rejection characteristics of the controllers. The selected indicators include the normalized settling time (\({N}_{st}\)), defined as the time required for the output to return within ± 2% of the setpoint after a disturbance; the normalized percent overshoot (\({N}_{po}\)), representing the maximum transient deviation from the reference; and the integral absolute error (\(IAE\)), which quantifies the accumulated deviation over the entire simulation horizon. The IAE was evaluated using Eq. (26):

$$IAE=\underset{0}{\overset{{t}_{final}}{\int}}\left|e\left(t\right)\right|dt$$
(26)

where \({t}_{final}=2000\) s. The results are summarized in Table 5. The proposed mNRBO-based controller achieves the lowest IAE (497.9906), the shortest settling time (1.7199E+03), and competitive overshoot (7.9070) compared with the other optimization algorithms. These findings confirm that mNRBO provides faster disturbance recovery, improved stability, and enhanced resilience against parameter perturbations, establishing its advantage over NRBO, HO, GEO, and SMA during dynamic operating conditions. The overall analysis verifies that the mNRBO-tuned PID controller combines rapid disturbance rejection with minimal error accumulation, ensuring both transient stability and steady-state accuracy under uncertain and time-varying conditions.

Table 5 Stability metrics under disturbance conditions.

Conclusion and future directions

This study introduced a new modified algorithm called mNRBO for optimal PID tuning in nonlinear process control and demonstrated its effectiveness on a CSTH system. By incorporating ROL and LGL, the algorithm achieved a balanced exploration–exploitation trade-off, resulting in faster convergence, improved tracking performance, and enhanced disturbance rejection compared with NRBO, HO, GEO, and SMA. The use of a standard PID structure ensures compatibility with existing control infrastructures, making the approach practical for industrial deployment. Future work will focus on hardware-in-the-loop validation and real-time implementation to assess performance under measurement noise, actuator constraints, and latency. Extending mNRBO to multi-input–multi-output (MIMO) systems and exploring hardware-aware, computationally lightweight variants will address scalability challenges. Moreover, integration with adaptive or model predictive control frameworks could further enhance responsiveness to process variability, while hybridization with machine learning surrogates may enable scalable, data-driven optimization in resource-constrained settings. Overall, mNRBO provides a computationally efficient and industrially relevant solution for nonlinear PID tuning, with clear potential for application in large-scale thermal process industries where reliability, energy efficiency, and precision are critical.