Introduction

With the emergence of FRP bars and their adoption in the construction industry, numerous studies have investigated their performance in various structural members. Among these, columns have received particular attention as one of the most critical structural elements.

The use of longitudinal and transverse reinforcement made of GFRP in columns, subjected to concentric and eccentric loading, has been investigated by numerous researchers 1,2,3,4. Hosseini and Sadeghian 5 proposed a deformability index by studying different configurations of transverse reinforcement, such as bar diameter and overlap length. Hadi et al. 6 examined various loading conditions in these columns and reported that neglecting the contribution of FRP bars to compressive strength leads to a significant discrepancy between analytical and experimental results. Due to the limited number of studies on the behavior of columns reinforced with BFRP bars, various parameters—including the percentage of longitudinal reinforcement, eccentricity, stirrup spacing, and stirrup diameter—have been investigated. These were compared with the corresponding parameters in similar columns reinforced with GFRP and steel bars 7,8,9,10. Researchers have also evaluated the incorporation of fibers in concrete for columns reinforced with GFRP bars 11,12. Alanazi et al. 13 further studied the effect of using steel tubes in GFRP-reinforced columns.

With numerous experimental studies conducted on columns reinforced with FRP, some researchers have proposed design equations to reduce the need for costly and repetitive experimental work 4,14,15. With the rise of neural networks and the advancement of machine learning, various models and equations have since been developed to predict the behavior of structural elements. Machine learning employs different techniques for data classification and prediction and is capable of learning relatively complex relationships from large volumes of information and data 16,17. To date, machine learning methods have been used to predict various characteristics of columns, including the compressive strength of concrete-filled steel tube (CFST) columns 18,19,20, columns confined with FRP wraps 21,22,23,24,25,26,27,28,29,30, the bond strength between FRP bars and concrete 31,32,33, and the strength of columns reinforced with FRP bars 34,35,36,37,38,39.

In the context of strength prediction for FRP-reinforced columns using machine learning, Arora et al. 40 applied an artificial neural network (ANN) to estimate the axial capacity of FRP-reinforced concrete columns and compared the results with 14 existing analytical models. All 242 data points used in their study involved columns under concentric loading; eccentric loading was not considered. Performance indices indicated that the proposed ANN model achieved the highest accuracy among the models, with a correlation coefficient (R) of 0.9758. Zhang et al. 41 investigated the behavior of hollow concrete columns (HCCs) using machine learning techniques such as light gradient boosting (LGB), extreme gradient boosting (XGB), and categorical gradient boosting (CGB). Among these, the CGB model demonstrated the highest accuracy, with all models showing less than 10% error compared to experimental results. Cakiroglu et al. 37 compiled 117 experimental data points to develop machine learning models including Random Forest, Kernel Ridge Regression, Categorical Gradient Boosting, Support Vector Machine, Gradient Boosting Machine, Lasso Regression, Adaptive Boosting, and Extreme Gradient Boosting. All columns in their dataset were subjected to concentric loading. Among the models tested, Gradient Boosting Machine, Random Forest, and XGBoost delivered the highest predictive accuracy. Tarawneh et al. 42 used an artificial neural network to design slender FRP-reinforced columns. Their model incorporated a factor to account for slenderness and introduced slenderness curves to support the design process. They also proposed a genetic expression programming (GEP)-based equation to predict the slenderness ratio, concluding that incorporating this ratio reduced the correlation between axial strength and slenderness. Balili et al. 36 proposed neural network and theoretical models to estimate the axial capacity of GFRP-reinforced columns incorporating glass fibers, using a dataset of 275 samples. The maximum errors reported for the neural network and theoretical models were 1.9% and 3.2%, respectively. Huang et al. 38 used 151 experimental data points on FRP-reinforced concrete columns and concluded that the backpropagation artificial neural network (BP-ANN) outperformed other mathematical models in terms of accuracy and minimized computational error.

Several equations have been proposed to predict the axial strength of FRP-reinforced concrete columns. Some of these are recommended by established design guidelines, while others have been developed by individual researchers. Certain equations account for the contribution of FRP bars to the axial strength of columns; however, several codes conservatively neglect this contribution. This conservative approach is based on experimental findings showing that FRP bars typically sustain only about 5% to 10% of the column’s maximum load 3. Table 1 presents a selection of well-known equations reported in the literature.

Table 1 Equations for axial strength of FRP reinforced concrete columns.

As can be seen, CSA-S806-12 does not account for the contribution of FRP bars to the axial strength of columns, whereas other equations do consider this component. In these equations, \(P_{n}\) is the nominal axial strength of column, \(A_{g}\) is the gross sectional area of columns, \(A_{FRP}\) is sectional area of FRP bars, \(E_{FRP}\) is the modulus of elasticity of FRP bars, \(\varepsilon_{fg}\) and \(\varepsilon_{p}\) are both the ultimate strain of FRP bars and \(f^{\prime}_{c}\) is the compressive strength of concrete. The structure of these equations is generally similar; the main differences lie in the values of the coefficients used.

A review of existing models and equations for estimating the compressive strength of FRP-reinforced concrete columns reveals that none directly account for the effects of bending (i.e., load eccentricity), column height, or the type and characteristics of the transverse reinforcement. Therefore, this study aims to develop more comprehensive equations that incorporate all of these parameters.

Objectives and outline of the research

An evaluation of the existing studies on FRP-reinforced columns reveals certain ambiguities and important limitations. One key limitation is the lack of direct consideration of the effect of eccentricity on the compressive strength of FRP-reinforced columns. Furthermore, in the reported studies, FRP bars have typically been used simultaneously for both longitudinal and transverse reinforcement. As a result, combinations involving FRP bars as transverse reinforcement and steel bars as longitudinal reinforcement—and vice versa—have not been thoroughly examined.

This study aims to address these gaps by proposing comprehensive formulas to predict the axial capacity and flexural strength of columns reinforced with FRP bars. The proposed formulas account for eccentricity, the shape of the column cross-section, and the use of FRP bars in either longitudinal and/or transverse reinforcement.

Figure 1 illustrates the fundamental steps undertaken in this study. First, a dataset of 525 samples was compiled, consisting of 12 input parameters and one output. In the next step, Pearson correlation coefficients were used to evaluate the relationships between input variables and their correlation with the output parameter. Additionally, cumulative frequency plots were generated.

Fig. 1
figure 1

Outline of the research.

Four modeling approaches were employed: artificial neural networks (ANN), gene expression programming (GEP), the group method of data handling (GMDH), and multiple linear regression (MLR). Among these, three methods yielded predictive equations. The performance of the proposed models was then evaluated and compared with existing equations using standard efficiency metrics to identify the most accurate approach for predicting the axial capacity of FRP-reinforced columns.

Datasets of the study

Collection of experimental data

A total of 525 datasets were collected from 66 reported experimental studies 1,2,4,6,14,15,36,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98,99,100,101,102,103. The data include columns with various cross-sectional shapes, such as circular, square, and rectangular sections. These columns are reinforced with longitudinal FRP bars, transverse FRP bars, or both. Both concentric and eccentric loading conditions are considered, and the transverse reinforcement consists of either ties or spirals. The experimental data cover a wide range of parameters, with the maximum axial load (Pmax) serving as the output variable. Table 11 summarizes the test references, number of specimens, column shapes, materials of longitudinal and transverse reinforcements, and load eccentricity.

A schematic view of the experiment and the parameters considered as inputs are shown in Fig. 2. Twelve input parameters are taken into account, which are listed alongside the output variable in Table 2. The predictive models in this study were validated using a conventional train/test split approach: 70% of the dataset (367 data points) was used for training, while the remaining 30% (158 data points) was reserved for testing.

Fig. 2
figure 2

Schematic view of input parameters in this study.

Table 2 Introduction to input and output parameters.

Table 3 presents the statistical characteristics of the dataset, including the maximum, minimum, mean, standard deviation, and coefficient of variation for each parameter, based on 525 selected data points.

Table 3 Statistical values of selected data.

Figure 3 presents the correlation matrix of the collected datasets, based on the Pearson correlation coefficient 104. This matrix illustrates the relationships between the various parameters. For the selected dataset, the correlation between the input variables and Pmax ranges from –0.48 to 0.65.

Fig. 3
figure 3

Correlation matrix of data used in this study.

The highest correlation is observed between \((A_{g} ,P_{max} )\), \((A_{g} ,A_{t} )\) , and \((A_{g} ,A_{f} )\) , with a coefficient of 0.65. This indicates that an increase in will have the most significant positive impact on Pmax compared to the other parameters. The second most influential parameter is (the cross-sectional area of the stirrup), with a correlation coefficient of 0.59. According to the figure, the e/D ratio exhibits the strongest negative correlation with Pmax, at –0.48. This means that as the e/D ratio increases, Pmax decreases, and among all parameters, this ratio has the greatest negative influence on the compressive strength of the column.

Figure 4 presents the distribution of data points across various ranges for each input parameter, along with their corresponding cumulative percentages. As shown in Fig. 4a, the majority of columns in the dataset have cross-sectional areas smaller than 5000 mm2, and approximately 90% of the columns have areas less than 10,000 mm2. Beyond this threshold, the number of data points declines sharply.

Fig. 4
figure 4figure 4

Frequency distribution of input data. (a) Ag, (b) H, (c) fc, (d) Af, (e) fyl, (f) El, (g) fyt, (h) Et, (i) e/D, (j) S, (k) At, (l) type

Figure 4b illustrates that most of the columns have heights ranging between 1000 and 2000 mm, followed by a smaller but notable portion with heights under 1000 mm. Columns exceeding 2000 mm in height comprise only about 5% of the total dataset. Regarding concrete compressive strength, no column was constructed with strength below 20 MPa. Approximately 60% of the columns fall within the 20–40 MPa range. A total of 131 columns have compressive strengths between 40 and 80 MPa, and 62 of these fall within the 60–80 MPa subset. Columns with compressive strengths above 80 MPa account for only about 3% of the dataset. In terms of longitudinal reinforcement, nearly 90% of the columns have reinforcement areas less than 2000 mm2, with only 10% exceeding this value. The yield strength of both longitudinal and transverse reinforcement bars shows a similar distribution. The majority of samples have reinforcement yield strengths between 500 and 1500 MPa, followed by a significant portion in the 1500–2000 MPa range. The elastic modulus of the longitudinal reinforcement is less than 50 GPa in approximately 36% of the dataset, whereas most samples fall within the 50–100 GPa range. A similar trend is observed for the transverse reinforcement bars. The distribution of load eccentricity, shown in Fig. 4i, indicates that about 70% of the samples either have no eccentricity or exhibit eccentricity less than 0.2 times the section height (a total of 369 samples). The next most frequent group includes samples with an e/D ratio between 0.2 and 0.4. Overall, 54% of the samples are subjected to concentric loading, while 46% are eccentrically loaded. As shown in Fig. 4j, approximately 70% of the columns have stirrup spacing less than 100 mm. For 25% of the samples, spacing falls between 100 and 200 mm, while only about 5% have stirrup spacing greater than 200 mm. The cross-sectional area of transverse reinforcement bars is less than or equal to 200 mm2 in 84% of the dataset. Finally, regarding stirrup type (Fig. 4l), only four columns were constructed without stirrups. Conventional stirrups were used in 278 samples, while spiral stirrups were used in 243 samples.

Methodology

Artificial neural network

Neural networks are composed of simple operational elements that work together in parallel. These elements are inspired by biological nervous systems. In nature, the functionality of neural networks is determined by the way components are connected.

After configuring or training the neural network, applying a specific input to it results in a specific output. As shown in the Fig. 5, the network adjusts based on the correspondence and alignment between the input and the target until the network’s output matches our desired output (target). Generally, a large number of these input–output pairs are used in this process, referred to as supervised learning, to train the network.

Fig. 5
figure 5

How neural network works redrawn from 105.

In the Fig. 6, a model of neuron with a single input is shown. The input p is applied to the neuron and is weighted by multiplying it by the weight “w”, resulting in a weighted input that is then fed into the transfer function f to produce the final output. By adding a bias to the neuron structure, a biased neuron is created. The bias input is a constant value (l), and the bias value is added to the product “w.p”, effectively shifting the function to the left.

Fig. 6
figure 6

The model of a neuron with one input redrawn from 105.

At least one neuron and sometimes more, make up a network layer. A neural network has at least the input layer, the hidden layer and the output layer. All of the layers in the network consist of nodes or neurons that work like the ones in the brain. The number of nodes in each layer varies with both how many features there are in the input and what the model is expected to do. The input layer takes external data, then sends it to the following layer so that various computations can be performed. This layer mainly receives data, moves it forward and processes nothing itself. Such a layer is able to receive text, sound and visual data. The hidden layer which lies between the input and output layers, is vital for the model. A neural network is built with at least one hidden layer and extra ones are included if the problem is more difficult. Putting in extra layers makes the network bigger and more complex to calculate. They take care of processing the data that was inputted from the layer before. The last layer, known as the output layer, computes the result that the network will give. Basically, it gets input from the hidden layers, works on it and produces the output that will be used 106.

In this study, two layers were considered for the network, including the hidden layer and the output layer. In order to optimize and select the best network, different neurons from 1 to 20 were selected and trained for the hidden layer. The results of mean square error and values of R for networks with various neurons are depicted in Figs. 7 and 8. The network type was selected as feed-forward backprop. Training and adaption learning functions were selected as TRAINLM and LEARNGDM, respectively. The performance function is also considered MSE. Transfer functions for hidden and output layers were considered tan-sigmoid and pure linear, respectively. For the selection of the best network, the amounts of MSE and R were considered (the lowest amount of MSE and the closest amounts of R to one). As can be seen in Figs. 7 and 8, the network with the number of 12 neurons in the hidden layer was considered the optimum network. The schematic view of the best network is shown in Fig. 9.

Fig. 7
figure 7

MSE values for networks with various neurons.

Fig. 8
figure 8

R values for networks with various neurons.

Fig. 9
figure 9

Schematic view of ANN network with optimum neurons.

The training process stops if the validation set error increases for six consecutive iterations. In Fig. 10, related to the optimized network, it is shown that this stopping occurred at iteration 22. Also, no overfitting occurred until iteration 16, where the best performance on the validation set was achieved.

Fig. 10
figure 10

Training process in the selected network.

Multiple linear regression

A statistical model known as “linear regression” calculates the linear relationship between a scalar response (dependent variable), and one or more explanatory variables (regressors or independent variables). Multiple linear regression refers to the procedure where there are more than one explanatory variable while simple linear regression deals with just one 107. In multiple linear regression, the parameters of a linear model are estimated with the help of an objective function and variable values. In linear regression, the considered model is a linear relationship in terms of model parameters. For the calculation of multiple linear regression, the following formula is used:

$$y_{i} = \beta_{0} + \beta_{1} x_{i1} + \beta_{2} x_{i2} + .... + \beta_{p} x_{ip} + \varepsilon_{i}$$
(1)

For every observation \(i = 1,2,.....,n\)

The number of observations of dependent and independent variables are considered n and p, respectively.

Consequently, \(y_{i}\) is the ith observation of dependent variable. The jth independent variable’s ith observation is denoted as \(x_{ij}\) (\(j = 1,2,.....,p\)). \(\beta_{0}\) is the constant term (y-intercept), \(\beta_{p}\) is the slope coefficients for each explanatory variable, whereas the ith independent identically distributed normal error is denoted by \(\varepsilon_{i}\). As is evident, this is more than just a straightforward line equation. In actuality, it’s a hyperplane Eq. 108.

In this study, the training and testing data consist of 70 and 30 percent of the whole data, respectively. After training the model, the coefficient of each input parameter and value of intercept were determined and the final equation for estimation of columns strength based on the MLR method is as follow:

$$\begin{gathered} P_{\max } = 204.5 + 0.015A_{g} - 0.085H + 14.37f^{\prime}_{c} + 0.23A_{f} + 0.24f_{yl} + 4.18E_{l} - 0.14f_{yt} + 0.59E_{t} \hfill \\ - 2942\left( {{e \mathord{\left/ {\vphantom {e D}} \right. \kern-0pt} D}} \right) - 3.32S + 3.44A_{t} - 244(type) \hfill \\ \end{gathered}$$

Evaluation metrics for MLR for training and testing data are computed and presented in Table 4.

Table 4 Evaluation metrics for MLR model.

Group method of data handling

Group method of data handling (GMDH) was invented by Professor Alexey G. Ivakhnenko in 1968 109. An external test criterion is used by these GMDH algorithms to search for and choose the fittest polynomial equation. With a second-order polynomial, the ANN-GMDH method builds a group of neurons. The approximate function \(\mathop f\limits^{ \wedge }\) with the output \(\mathop y\limits^{ \wedge }\) produced by combining the second-order polynominals of all neurons, is proposed for a set of inputs \(x = \left\{ {x_{1} ,x_{2} ,x_{3} ,....} \right\}\) in order to make the least error with the real output (y). So y can be described as follows for M experimental data with n output and one target:

$$y_{i} = f\left( {x_{i1} {,}x_{i2} {,}x_{i3} {,} \ldots {,}x_{in} } \right)\left( {i = 1{,}2{,} \ldots {,}M} \right)$$
(2)

The output variables which are predicted by GMDH can be described for any input vector of \(X = (x_{i1} ,x_{i2} ,x_{i3} ,....,x_{in} )\), as follows:

$$\mathop y\limits^{ \wedge }_{i} = \mathop f\limits^{ \wedge } \left( {x_{i1} {,}x_{i2} {,}x_{i3} {,} \ldots {,}x_{in} } \right)\left( {i = 1{,}2{,} \ldots {,}M} \right)$$
(3)

Minimizing the square of error between y and \(\mathop y\limits^{ \wedge }\) should be done by the ANN-GMDH method, Put differently:

$$\mathop \sum \limits_{i = 1}^{M} \left[ {\mathop f\limits^{ \wedge } \left( {x_{i1} ,x_{i2} ,x_{i3} ,....,x_{in} } \right) - y_{i} } \right]^{2} \to {\text{min}}$$
(4)

The connection between the variables that are input and output which is known as the polynomial Ivakhnenko, can be represented in the following way using the polynomial function:

$$y = a_{0} + \mathop {\mathop \sum \nolimits }\limits_{n}^{1} a_{i} x_{i} + \mathop {\mathop \sum \nolimits }\limits_{n}^{1} \mathop {\mathop \sum \nolimits }\limits_{n}^{1} a_{ij} x_{i} x_{j} + \mathop {\mathop \sum \nolimits }\limits_{n}^{1} \mathop {\mathop \sum \nolimits }\limits_{n}^{1} \mathop {\mathop \sum \nolimits }\limits_{n}^{1} a_{ijk} x_{i} x_{j} x_{k} + {\mkern 1mu} \ldots$$
(5)

This polynomial with two variables can be presented as follows:

$$\mathop y\limits^{ \wedge } = G\left( {x_{i} {,}x_{j} } \right) = a_{0} + a_{1} x_{i} + a_{2} x_{j} + a_{3} x_{i} x_{j} + a_{4} x_{i}^{2} + a_{5} x_{j}^{2}$$
(6)

Regression techniques are used to identify the unknown coefficients \(a_{i}\) in order to minimize the discrepancy between the computed values and the actual output for each input variable \(x_{i}\) and \(x_{j}\). It is possible to adapt the inputs to all pairings of input–output sets optimally by obtaining coefficients for each function G (i.e., each created neuron) that minimize the total neuron error.

$$E = \frac{{\mathop \sum \nolimits_{i = 1}^{M} (y_{i} - G_{i} )^{2} }}{M} \to {\text{min}}$$
(7)

In this study, with the aim of using the GMDH method to predict data, GMDH Shell software (GMDH Shell DS 3.8.9) 110 was used. Furthermore, this software drastically cuts down on computation time by not requiring any prior data standardization. Using neural networks, it can create a candidate model with high prediction strength by identifying and considering non-linear interactions and connections between data. Additionally, it can provide a useful equation to predict desirable outputs. After running the software with the input and output data, the computed parameters of the software are presented in Table 5. In Fig. 11, the predicted values by the GMDH shell and their distance from the actual values are shown. The proposed formula by GMDH Shell is presented in Eq. 8.

Table 5 The measured parameters by GMDH Shell.
Fig. 11
figure 11

Difference between actual and predicted data by GMDH.

The parameters defined as model inputs, which are listed in Table 3, are also considered as inputs for the formula. To use Eq. 8, the input parameters are substituted into the formula, and the mathematical operations are performed in the following order: exponentiation, square root, multiplication, division, addition, and subtraction. To simplify the formula, Eq. 8 is divided into four parts, which are then summed together. Accordingly, the parameters \(f_{1} ,f_{2} ,f_{3} \,\,\,and\,\,f_{4}\) are first calculated in sequence and finally added together to obtain the desired compressive strength. \(f_{1} ,f_{2} ,f_{3} \,\,\,and\,\,f_{4}\) are defined in the appendix.

In Fig. 12, the effect of each input parameter on output can be seen. Gross area of cross section, area of transverse reinforcement, and area of longitudinal bars are the most effective parameters with direct effect, so increasing these parameters will increase the compressive strength of the column, while the “e/D” parameter has the most negative effect on the output, so by increasing this parameter, the compressive strength of the column will decrease. This figure also shows that fyl, Et, and S are other parameters with adverse impacts on the axial capacity of the column.

$${\text{P}}_{{{\text{max}}}} \, = \,f(f_{1} ,f_{2} ,f_{3} ,f_{4} ) = f_{1} + f_{2} + f_{3} + f_{4}$$
(8)
Fig. 12
figure 12

Impact of each input parameter on output.

Gene expression programming

Gene Expression Programming is an algorithm for creating mathematical models and computer programs that are based on evolutionary computations and drawn from the principles of natural evolution. Ferreira created this technique in 1999, and it was formally unveiled in 2001 111. To address the shortcomings of the first two genetic algorithms, the GEP algorithm really combines the dominant viewpoints of the two. In this method, the chromosomes’ phenotype—a tree structure with varying length and size—is similar to the genetic programming algorithm, while their genotype has a linear structure similar to the genetic algorithm. Therefore, the GEP algorithm, by overcoming the limitation of the dual role of chromosomes in the previous algorithms, provides the possibility of applying multiple genetic operators with the guarantee of the permanent health of the child chromosomes, and with a speed faster than genetic programming (GP) due to the higher structural diversity than genetic algorithm (GA), it expands the space of possible answers to search more thoroughly. From this perspective, GEP has really been successful in surpassing the first and second thresholds—the replicator threshold and the phenotype threshold—assumed in natural evolution processes.

With this approach, a collection of functions and terminals are used to model different phenomena. The primary arithmetic functions (division, multiplication, addition, and subtraction), trigonometric functions, and other mathematical functions (sin, cos, x2, exp, etc.) or functions defined by user that may be appropriate for the model’s elucidation typically make up a set of functions. Constant values and the independent variables in the problem make up the set of terminals. This approach typically relies on a genetic algorithm to select data points based on the merit function, using the population of data as a source. Certain operators (genes) are also used to perform genetic variations. Unlike GA and GP, GEP uses multiple genetic operators to replicate data simultaneously. The renowned Roulette wheel is used to choose the data. The purpose of duplication is to store multiple relevant pieces of data from one generation to the next. The mutation operator aims to optimize the targeted chromosomes internally through random optimization.

A chromosome, which may include one or more genes, represents an individual in GEP. A single-gene chromosome is referred to as a monogenic chromosome, while a chromosome made up of several genes is called a multigenic chromosome. A multigenic chromosome is particularly helpful for breaking down difficult problems into smaller, more manageable components, each of which is encoded by a different gene. This allows for the modular creation of intricate, hierarchical structures. Every gene in a multigenic chromosome is translated into a sub-ET (expressed tree), and the sub-ETs are usually combined into one by a linking function, also known as a linker 112,113. For modeling, some parameters such as the number of chromosomes, head size, number of genes, and linking function that are considered in this study are presented in Table 6. In this table, all the parameters are considered default except head size, chromosomes, and genes. Head size and genes should be determined before modeling, but the number of chromosomes can be changed during training to improve the accuracy of the model. After modeling with different head sizes and chromosomes, these parameters are selected as 14 and 4, respectively. The number of chromosomes is also changed between 50 and 500 to improve the accuracy of the model. Root mean square error (RMSE) is also considered for the type of fitness function of error.

Table 6 Modeling parameters in GEP.

After modeling, the values predicted by GEP along with their experimental values are presented in Figs. 13a and b for training and testing data, respectively. In these figures, the difference between the actual and estimated values of each data is relatively small. Figures 13a and b confirm the accuracy and capability of this model in predicting the results. So based on this conclusion, the equation based on GEP, which is presented in Eq. (9), can be used reliably to predict the axial capacity of FRP-reinforced RC columns in the coming applications.

Fig. 13
figure 13

Experimental versus predicted data in GEP for (a) training data, (b) testing data.

A tree diagram of the model proposed by GEP is illustrated in Fig. 14. As can be seen, this model consists of four sub-ETs, which represent the number of chromosomes. For the equation, these sub-ETs should be appended together by a linking function, which is considered multiplication in this study.

Fig. 14
figure 14

Four sub-ETs from GEP.

By multiplying sub-ETs, the final equation for predicting the axial capacity of FRP-reinforced columns is presented in Eq. (9). Regarding Eq. 9, first, the input parameters are inserted, and the values of \(g_{1} ,g_{2} ,g_{3} \,\,\,and\,\,g_{4}\) are calculated. Finally, the compressive strength is obtained by multiplying these four parameters. \(g_{1} ,g_{2} ,g_{3} \,\,\,and\,\,g_{4}\) are defined in the appendix.

$$P_{\max } = g(g_{1} ,g_{2} ,g_{3} ,g_{4} ) = g_{1} \times g_{2} \times g_{3} \times g_{4}$$
(9)

The sensitivity analysis done by the software is presented in Fig. 15. As the figure shows, the most important parameter that can affect the compressive strength of the column is e/D, with a percentage effect of 34.54. The second and third positions of effect are for Ag and fc, with a percentage of 26.54 and 17.37, respectively. The figure also shows that fyl and El had no effect on the axial capacity of the column.

Fig. 15
figure 15

Importance of different variables on output in equation proposed by GEP.

Comparison of the proposed formulas

Following the modeling process, the performance of each algorithm was evaluated using standard error metrics and efficiency criteria, including mean absolute percentage error (MAPE), coefficient of determination (R2), root mean square error (RMSE), and mean absolute error (MAE), as recommended in 23 reference. These metrics are defined by the following expressions:

$$MAE = \frac{1}{n}\sum\limits_{i = 1}^{n} {\left| {P_{(model)} - P_{(actual)} } \right|}$$
(10)
$$RMSE = \sqrt {\frac{1}{n}\sum\limits_{i = 1}^{N} {(P_{(model)} - P_{(actual)} )^{2} } }$$
(11)
$$MAPE = \frac{1}{n}\left[ {\frac{{\sum\limits_{i = 1}^{M} {\left| {P_{(\bmod el)} - P_{(actual)} } \right|} }}{{\sum\limits_{i = 1}^{M} {P_{(actual)} } }} \times 100} \right]$$
(12)
$$R^{2} = 1 - \frac{{\sum\limits_{i = 1}^{N} {\left( {P_{(actual)} - P_{(model)} } \right)^{2} } }}{{\sum\limits_{i = 1}^{N} {\left( {P_{(actual)} } \right)^{2} } }}$$
(13)

In these equations, n denotes the number of data points—in this study, 525—and Pmodel and Pactual represent the axial capacity predicted by the machine learning model and the experimentally measured capacity, respectively. Lower values of MAE, MAPE, and RMSE, along with R2 values approaching 1, indicate higher predictive accuracy. The performance metrics for the different models are summarized in Table 7. Among the four evaluated machine learning algorithms—artificial neural networks (ANN), multiple linear regression (MLR), the group method of data handling (GMDH), and gene expression programming (GEP)—the ANN model demonstrated the highest predictive performance. On the test dataset, it achieved the lowest MAE (130.13), RMSE (211.45), and MAPE (11.05%), along with a high R2 value of 0.96.

Table 7 Efficiency criteria for various models.

In contrast, the MLR model exhibited the weakest performance, suggesting a highly nonlinear relationship between the input parameters and the target variable. Both the GMDH and GEP models also showed strong predictive capabilities, as reflected in their error metrics.

Furthermore, the mean and standard deviation of prediction errors, also presented in Table 7, were lowest for the ANN, GMDH, and GEP models. This further confirms their robustness and reliability in predicting the axial capacity of FRP-reinforced columns. In comparison, the average and standard deviation of errors for the MLR model were approximately six and sixteen times higher than those of the ANN model, respectively, reinforcing its relative inadequacy for modeling complex nonlinear behavior in this context.

In Figs. 16 and 17, the performance of the models in predicting different data points is shown for the training and testing datasets, respectively. Each point represents one data sample, and based on its location, the accuracy of the models can be assessed. In general, the closer the points are to the 45-degree line (green line), the more accurate the model’s predictions. If all points lie exactly on this line, it indicates that the predicted values are identical to the experimental values.

Fig. 16
figure 16

Performance of models in prediction of each training data. (a) ANN, (b) MLR, (c) GMDH, (d) GEP

Fig. 17
figure 17

Performance of models in prediction of each testing data. (a) ANN, (b) MLR, (c) GMDH, (d) GEP

Figures 16 and 17 demonstrate that the ANN model has the best performance, as most of its points are clustered near the 45-degree line. Overall, the ANN, GMDH, and GEP algorithms show the best predictive performance for both the training and testing datasets. In contrast, MLR delivers the weakest results among the evaluated models, with many of its outputs falling beyond the 40% error line.

As shown in Table 8, only four data points have errors greater than 60% for the ANN model, and the majority of its predictions (89.7% of the data) have errors below 20%. For GMDH and GEP, 76% and 73% of the data, respectively, fall within the same error range. The number of data points across different error intervals is nearly identical for GMDH and GEP, but GMDH performs slightly better, as it has more data points with errors under 40% (481 compared to 479).

Table 8 Number of data in different ranges of errors for all models.

MLR, on the other hand, has the highest number of predictions with errors greater than 60% among all models, confirming its comparatively lower accuracy.

In order to better understand the formulas presented in this study, a column with various eccentricities from the study by Hadhood et al. 2 is selected in this section, and the compressive strengths and moments are estimated using the proposed formulas. The results obtained from each equation, alongside their corresponding experimental values, are presented in Table 9. The axial load–moment diagram from the experiment, as well as the results of each method, are illustrated in Fig. 18. As shown, the results from the GMDH and GEP formulas are very close to the experimental data, while the MLR results show a relatively significant discrepancy. This example demonstrates the acceptable accuracy of the formulas proposed in this study.

Table 9 Axial load and moment of columns predicted by equations.
Fig. 18
figure 18

Comparison of the interaction diagrams of a tested column with the proposed equations.

Comparison of different formulas

As mentioned earlier, all existing equations (some of which are presented in Table 1) predict only the sectional capacity of columns under axial load without directly considering eccentricity. The performance of these equations is compared using various metrics, including MAE, MAPE, RMSE, and R2, with the results shown in Table 10. It can be seen that the equations developed by Mohammad et al. (B), Tobbi et al., and AS-3600 are more accurate than the others, with R2 values of 0.9, 0.883, and 0.883, respectively. Among these, the equation by Mohammad et al. (B) performs best due to its lower MAE, RMSE, and MAPE values. Table 10 also shows that the equations proposed in this study not only account for eccentricity but also exhibit higher accuracy compared to the existing equations in predicting the axial capacity of columns.

Table 10 Performance checking variables for different equations and models.

The performance of these equations is also compared to a machine learning model developed by Arora et al. 40 using ANN. As can be seen, the ANN model done by Arora et al. 40 demonstrates the best performance among the existing equations, owing to its higher R2 value. When compared with the models proposed in this study 40 , the ANN and GMDH models show greater accuracy in prediction, while the GEP model performs comparably. Overall, given that all models have R2 values above 0.9, it can be concluded that they all demonstrate strong performance in predicting the compressive strength of columns reinforced with FRP bars under concentric loading.

In Fig. 19, the dispersion of predicted data from both the existing equations and the proposed equations (GMDH and GEP) is compared. In this figure, each point represents the prediction error for a given data sample. As mentioned earlier, the error range for the proposed equations is within ± 20%, indicating relatively high accuracy in estimating the axial capacity of the columns. In contrast, for the existing equations, most points lie far from the 45-degree line. Even for the most accurate among them—Mohammad et al. (B)—the error range is generally between ± 40% and ± 60%.

Fig. 19
figure 19

Comparison of the proposed model with other studies.

The Taylor diagram for the models employed in this study is presented in Fig. 20. This diagram provides a comprehensive visual summary of model performance by simultaneously illustrating the correlation coefficient, standard deviation, and root-mean-square error (RMSE). The horizontal and vertical axes represent the standard deviation of the predicted values, while the black dashed lines denote the Pearson correlation coefficient. Additionally, the blue contours correspond to RMSE values, as outlined by references 114,115.

Fig. 20
figure 20

Taylor diagrams for proposed models.

The accuracy of each model is indicated by its proximity to the reference point on the diagram. Models that lie closer to the origin and along higher correlation coefficient lines (i.e., smaller angular deviation from the x-axis) demonstrate superior predictive performance.

Among the models analyzed, the artificial neural network (ANN) exhibited the highest overall accuracy, with a correlation coefficient exceeding 0.97, a well-aligned standard deviation, and minimal RMSE—indicating its strong reliability in predicting the axial load capacity of FRP-reinforced columns.

The GMDH and GEP models also showed commendable performance, with correlation coefficients of approximately 0.96 and 0.94, respectively. These models strike a favorable balance between predictive accuracy and model interpretability, making them viable for practical engineering applications.

In contrast, the multiple linear regression (MLR) model showed the weakest performance, with a correlation coefficient of roughly 0.6 and a comparatively larger standard deviation. This suggests a limited ability to capture the complex behavior of FRP-reinforced columns under axial loading.

Overall, the Taylor diagram corroborates the results obtained from other statistical metrics and visually reinforces the superior performance of the ANN model, along with the strong predictive capability of the GMDH and GEP models.

Summary and conclusions

This study employs machine learning techniques to develop predictive models for estimating the compressive strength of concrete columns reinforced with fiber-reinforced polymer (FRP) bars. To achieve this, experimental data from multiple sources were compiled, resulting in a curated dataset of 525 data points. Twelve input parameters were selected for model development. Notably, several critical variables—such as the ratio of eccentricity to the diameter or height of the cross-section (e/D), the column height (H), and the mechanical properties, type, and spacing of transverse reinforcement—were incorporated. These variables have not been explicitly addressed in previously published empirical equations.

The relationships between the input variables and their influence on the output were examined using a correlation matrix. The analysis revealed that column cross-sectional area had the strongest positive correlation with compressive strength, while eccentricity exhibited the most significant negative effect. The distribution of input variables across various intervals was also analyzed, both numerically and as percentages.

Four machine learning algorithms were applied: Artificial Neural Network (ANN), Multiple Linear Regression (MLR), Group Method of Data Handling (GMDH), and Gene Expression Programming (GEP). The models were trained and tested using standard performance metrics. Among them, ANN and GMDH yielded the most accurate predictions, with R2 values of 0.981 and 0.973 for the training set and 0.960 and 0.949 for the test set, respectively. The GEP model also demonstrated satisfactory performance, achieving R2 values above 0.93 for both datasets. In contrast, the MLR model exhibited the weakest predictive capability. Further analysis of error distribution and standard deviation showed that the ANN model achieved the highest overall accuracy, with an R2 of 0.974 and a standard deviation of 10.22. Notably, approximately 90% of the predictions generated by the ANN model fell within an error range of 0–20%.

In addition, several existing empirical formulas for estimating the compressive strength of FRP-reinforced columns were reviewed. Since most of these equations do not directly consider the effect of eccentricity, only the subset of data involving concentric loading was used for comparison. The compressive strength values obtained from both the existing formulas and the proposed models were compared for this subset.

To develop generalized prediction equations, the GMDH, GEP, and MLR methods were employed. Among these, the equations derived from the GMDH and GEP methods demonstrated relatively high accuracy, with R2 values of 0.966 and 0.942, respectively. These proposed equations were benchmarked against those available in existing design codes and published studies. The comparison confirmed that the GMDH- and GEP-based equations outperformed all other existing formulas in terms of predictive accuracy. Importantly, these models are capable of estimating both axial load capacity and moment strength for a wide range of input parameters, offering a substantial improvement over current empirical equations. The proposed equations are also straightforward to implement in engineering practice.

Given that the models were trained on 525 data points within defined parameter ranges, future work can improve their generalizability and precision by expanding the dataset and broadening the scope of the input variables. As one of the first studies to directly incorporate eccentricity effects in modeling the compressive strength of FRP-reinforced columns, this work provides a valuable reference for researchers aiming to develop more advanced models—not only for FRP-reinforced members but also for those reinforced with conventional steel.