Introduction

With the deepening of underground mineral resource development, especially in the mining process of deep mines, mining operations are facing more and more severe challenges. Factors such as high ground stress, high ground temperature and complex geological environment not only increase the technical difficulty of mining operations, but also increase the risk of potential safety hazards1. Therefore, dynamic disasters in mines, especially rock bursts, occur frequently, seriously threatening the life safety of operators and the normal operation of mining equipment. In order to cope with these severe safety issues, microseismic monitoring technology has gradually become an indispensable and important means for underground mine safety monitoring2. Microseismic monitoring can realize real-time assessment of rock stability by collecting elastic wave signals released by rock mass during fracture or rupture through precision instruments3. By analyzing the characteristics of microseismic signals, especially parameters such as waveform, frequency and energy, early warning can be provided before disasters occur, thereby providing strong support for the safety management of mining operations. However, the application of microseismic monitoring technology still faces some challenges, especially the high similarity between microseismic signals and other geological disturbance signals in waveform and spectrum characteristics, which makes the accurate identification and classification of microseismic signals a major bottleneck in the implementation of the technology4. Therefore, improving the recognition accuracy of microseismic signals and blasting vibration signals and developing more accurate signal processing algorithms are important aspects for researchers to enhance the effectiveness and reliability of microseismic monitoring technology in deep mining applications.

In recent years, the rise of machine learning and deep learning has provided new research ideas for the classification and prediction of microseismic blasting signals. For example, Di et al.5 constructed a new prediction model for microseismic, acoustic emission and electromagnetic radiation data based on multiple neural networks. Field verification shows that the proposed model can provide quantitative prediction data for rockburst monitoring. Luo et al.6 proposed combined hierarchical information model (CHIM-Net) to improve the prediction accuracy of microseismic events in high-energy mining through deep learning technology. The model effectively improves the accuracy of time, space and intensity prediction by combining two prediction branch modules through data decomposition and partitioning. Experimental results show that CHIM-Net outperforms other models in multiple evaluation indicators. Basnet et al.7 compared the application of parametric and non-parametric machine learning models in short-term rockburst risks prediction, and used logistic regression and support vector machine (SVM) to evaluate data sets. The results show that the parameterized model performs better on the normally distributed data set and can effectively predict the short-term rockburst risk, especially when the amount of data is limited. Shu et al.8 combined the histogram of oriented gradients (HOG) and ML techniques for microseismic waveform recognition. They extracted HOG features from images of event waveforms and evaluated five distinct classifiers. The experimental outcomes demonstrated SVM has the highest accuracy of 97.1%. Goulet et al.9 developed a random forest model to predict the seismic response intensity using 379 mining development blasting data, and verified the model performance by analyzing an additional 100 cases. Otaíza et al.10 used a SVM model to classify earthquakes and blasting activities recorded by an open-pit copper mine seismic network. The experimental results showed that the proposed method showed extremely high overall accuracy.

The above studies show that the existing models have achieved high performance in microseismic blasting prediction. However, the prediction accuracy of these models still has room for further improvement. The microseismic monitoring system collects data characterized by high signal volume and a wide variety of complex features. Therefore, redundant features in the features may cause the prediction model to overfit. At the same time, irrelevant features in the data can interfere with the model’s learning process, ultimately reducing its prediction accuracy. Thus, it is necessary to perform feature selection (FS) on the collected microseismic blasting data, extract effective key features, and improve the prediction performance.

Metaheuristic algorithms have been applied to FS tasks in recent years due to their good global search capabilities and ability to find approximate optimal solutions in a limited time. These algorithms, such as genetic algorithms (GA)11, differential evolution (DE)12, sine cosine algorithm (SCA)13, particle swarm optimization (PSO)14, Harris hawks optimization (HHO)15, moss growth optimization (MGO)16, artemisinin optimization (AO)17, and polar lights optimization (PLO)18 can efficiently explore complex search spaces to identify the most discriminative feature subsets for prediction tasks. Kwakye et al.19 combined PSO with bald eagle search to improve the global search performance. The FS performance was verified using 27 classification data sets. Lee et al.20 studied the effectiveness of FS based on harmony search and GA in improving the performance of diagnosing sarcopenia. The experimental results show the potential of metaheuristic selection in advancing sarcopenia diagnosis based on its competitive advantage. Beheshti et al.21 proposed a fuzzy transfer function and applied it to binary PSO (FBPSO). They evaluated the effectiveness of both atomic and composite transfer functions within binary PSO, as well as several other binary metaheuristic algorithms, by comparing them to FBPSO. The experimental outcomes demonstrated that FBPSO outperformed the other algorithms, achieving the lowest error rates and the highest FS rates. Ali et al.22 combined evolutionary algorithms and local search techniques to propose a machine learning model that can identify network intrusions. The synthetic minority over-sampling technique is used to solve the data imbalance problem and improve the model accuracy. The proposed method performs well on the CSE-CIC-IDS2018 and KDD CUP 99 datasets and can effectively improve the ability of intrusion detection. Al-Thanoon et al.23 proposed to use Kruskal-Wallis test and one-way ANOVA to initialize binary metaheuristic algorithms. Experimental results show that these methods outperform the comparison methods in classification accuracy, number of FS and running time on high-dimensional datasets. Bagadi et al.24 introduced an innovative robust hybrid metaheuristic approach for FS aimed at enhancing the accuracy of speech emotion recognition (SER) tasks. Experimental results on SER datasets show that their approach achieved high-level emotion recognition rates of 94.35% and 96.78%, respectively. Armaghani et al.25 combined three metaheuristic algorithms with SVM to propose an efficient hybrid model for rockburst trend prediction. Experimental results on 259 rockburst examples show that these hybrid models are superior to traditional SVM models in both prediction accuracy and generalization ability.

The above studies show that the model’s prediction accuracy can be effectively improved by combining metaheuristic algorithms with ML models. However, in the field of microseismic blasting, few people have studied how to apply metaheuristic algorithms to FS to improve the performance of prediction models. To this end, this study proposes a FS method based on the improved HHO15 for microseismic blasting events prediction. HHO is a metaheuristic algorithm that exhibits a robust global search capability and achieves rapid convergence. However, the optimization process of the original HHO is mainly guided by the current optimal individual and is prone to fall into the local optimum. This study introduces a kernel search-based HHO variant (KSHHO), which incorporates kernel functions to enhance the algorithm’s search space and reduce the likelihood of getting stuck in local optima. The CEC 2022 benchmark function is used to qualitatively analyze KSHHO, and the performance of KSHHO is compared with 10 state-of-the-art (SOTA) algorithms. A binary version of KSHHO (bKSHHO) was then proposed, and the bKSHHO-KELM model was proposed based on bKSHHO and KELM for microseismic blasting prediction tasks. Experiments were conducted on the microseismic blasting dataset with 7 classic classifiers, and the results were compared using four evaluation metrics. The main contributions of this study are as follows:

  1. (1)

    A HHO variant is proposed utilizing a kernel search strategy. This strategy effectively improves the algorithm’s search capability and avoids the algorithm falling into local optimality.

  2. (2)

    bKSHHO-KELM microseismic blasting prediction is proposed. It aims to improve the recognition accuracy of microseismic and blasting events.

  3. (3)

    KSHHO’s optimization performance is evaluated using the CEC 2022 benchmark function.

  4. (4)

    Based on the features extracted from microseismic and blasting images, the bKSHHO-KELM model was employed to identify the most optimal subset of features, thereby successfully achieving high-precision recognition of microseismic blasting events.

The subsequent sections are structured as follows: section “Materials and methods” details the microseismic blasting dataset utilized in this study and also introduces the specific implementation details of the original HHO, KSHHO and bKSHHO-KELM in detail. Section “Experiment and results” analyzes the experimental results of KSHHO in optimization experiments and bKSHHO-KELM in prediction tasks. In section “Discussion”, a detailed discussion regarding the limitations of the present study is provided, as well as suggestions for future improvements. Section “Conclusion and future works” summarizes the whole study and gives future prospects.

Materials and methods

Description of microseismic and blasting data

The data used in this study were collected from the online microseismic monitoring system deployed in the Daikoutou mining section of the Jiuqu mining area at the Linglong Gold Mine. This system was jointly developed by Linglong Gold Mine and Central South University and employs IMS microseismic monitoring technology to conduct long-term monitoring of deep ground stress conditions. As mining depth has exceeded 900 m, ground stress has increased significantly, leading to a higher risk of geological hazards such as rock bursts. To obtain reliable microseismic data to support the construction of disaster prediction models, the research team collected a large volume of representative waveform data at several critical levels, including − 620 m, -670 m, and − 694 m. The raw data are in the form of waveform images captured by the microseismic monitoring system. Each time a microseismic or blasting event occurs, vibration signals are recorded by multiple triaxial sensors and converted into images using the system’s built-in software, TRACE, which also logs the arrival times of P-waves and S-waves, sensor trigger times, and trigger sequences.

Due to the fundamental differences in the generation mechanisms of blasting and microseismic events, their waveform representations exhibit distinct characteristics. Figures 1 and 2 show representative waveform images of a blasting event and a microseismic event, respectively. As shown in Fig. 1, the waveform of the blasting event demonstrates a sharp rise in amplitude concentrated at the front end, followed by rapid attenuation, with no clear S-wave phase, resulting in a waveform that is short and intense. In contrast, the waveform image of the microseismic event, shown in Fig. 2, exhibits a longer duration, more evenly distributed energy, clear transitions between P-wave and S-wave phases, and significantly lower amplitude compared to the blasting event. In this study, by manually screening and interpreting waveform images, the events were accurately classified, and a clearly labeled image dataset of microseismic and blasting events was constructed. Subsequently, color moments (CM)32, gray-level co-occurrence matrix (GLCM)33, and local binary patterns (LBP)34 were employed to extract image features from multiple perspectives, including global statistical properties, spatial structural relationships, and local texture details. These features formed the basis for the training and testing of subsequent machine learning models.

Fig. 1
Fig. 1
Full size image

Waveform image of a blast event.

Fig. 2
Fig. 2
Full size image

Waveform image of a microseismic event.

Harris hawks optimization

HHO is a population-based optimization algorithm inspired by the cooperative hunting behavior of Harris hawks when pursuing prey, such as rabbits15. By mimicking the natural predation mechanisms of Harris hawks, the algorithm converts complex optimization problems into a dynamic problem-solving framework. At its core, HHO models candidate solutions as Harris hawks that iteratively approach the global optimal solution through cooperative and adaptive search strategies. The HHO process consists of two phases: exploration and exploitation, corresponding to global search and local search. These phases alternate to balance the breadth and depth of the search process. The HHO algorithm contains specific formulas as shown in Table 1.

In the exploration phase (Eqs. 1–3), the algorithm enhances diversity by deploying a randomly distributed population of Harris hawks to widen the search range. This phase simulates hawks scouting for prey, including guarding random locations or moving to new ones. The main goal of this phase is to improve population diversity and avoid premature convergence to local optima. In the exploitation phase, the algorithm refines the search space to pinpoint the optimal solution with precision. Strategies in this phase mimic various scenarios of Harris hawks attacking prey, including soft besiege (Eqs. (10)-(13)) and hard besiege (Eqs. (4)-(9)), targeting prey with strong or weak escape capabilities, respectively. HHO has been widely applied to various optimization problems, showing excellent performance in benchmark tests26,27 and practical engineering scenarios28,29. Detailed algorithm implementation, including mathematical formulations and performance analysis, is provided in reference15.

Table 1 Specific formulas in HHO algorithm.

Harris hawks optimizer with kernel search

Addressing the challenges of nonlinear function optimization has long been a significant difficulty in modern optimization. Kernel methods, originally developed in machine learning, map data into high-dimensional feature spaces, transforming nonlinear problems—difficult in lower dimensions—into linearly separable problems in higher dimensions.

Kernel search optimization (KSO) algorithm utilizes a kernel mapping function to transform a multidimensional nonlinear function into a linear form, simplifying the optimization process30,31. Specifically, given an objective function \(y=f\left( x \right)\),\(x=\left( {{x_1},~{x_2},~ \cdots ,~{x_n}} \right)\), KSO applies the kernel mapping function \(u=\varphi \left( x \right)\),\(~u=\left( {u,~{u_2},~ \cdots ,~{u_n}} \right)\), to convert it into a linear form.

$$y = f\left( x \right) = w^{T} \cdot u + b = \varphi \left( a \right) \cdot \varphi \left( b \right) + b = K\left( {a,~x} \right) + b$$
(14)

where w denotes the mapping from the original space vector \(a=\left( {{a_1},~{a_2},~ \cdots ,~{a_n}} \right)\) to a high-dimensional space vector, while \(\varphi \left( a \right)\) represents the normal vector of the hyperplane in the high-dimensional space. The kernel function \(K\left( {a,~x} \right)\) implicitly computes the similarity in the high-dimensional space through the inner product.

To address nonlinear problems more effectively, the KSO algorithm utilizes the radial basis function (RBF) as its kernel. The RBF kernel is widely used due to its strong generalization ability and adaptability to nonlinear data. Its form is given by:

$$y=f\left( x \right)=K\left( {a,~x} \right)+b={e^{\frac{{{{\left\| {x - a} \right\|}^2}}}{\sigma }}}+b$$
(15)

The KSO algorithm introduces two sets of random vectors \(\left( {x_{1}^{\prime },~x_{2}^{\prime },~ \cdots ,~x_{n}^{\prime }} \right)\) and \(\left( {x_{1}^{{''}},~x_{2}^{{''}},~ \cdots ,~x_{n}^{{''}}} \right)\) to estimate the parameters a and \(\sigma\) of the kernel function. The algorithm replaces the \(jth\) component of the vector \(\left( {{x_1},~{x_2},~ \cdots ,~{x_n}} \right)\) with the \(jth\) components of the random vectors, yielding the corresponding target values \(y_{j}^{\prime }\) and \(y_{j}^{{''}}\). By comparing the objective function values at different points, the algorithm iteratively adjusts \(\:a\) and \(\:\sigma\:\). The specific computation formula is as follows:

$$\sigma =\frac{{x_{j}^{\prime } - x_{j}^{{''}}}}{{\ln \left( {\frac{{y - b}}{{y_{j}^{\prime } - b}}} \right)/\left( {{x_j} - x_{j}^{\prime }} \right) - \ln \left( {\frac{{y - b}}{{y_{j}^{{''}} - b}}} \right)/\left( {{x_j} - x_{j}^{{''}}} \right)}}$$
(16)
$${a_i}=\frac{1}{2}\left[ {{x_i}+x_{i}^{\prime } - \sigma \ln \left( {\frac{{y - b}}{{y_{j}^{\prime } - b}}} \right)/\left( {{x_j} - x_{j}^{\prime }} \right)} \right]$$
(17)

Based on the above formula, the candidate optimal solution \({x_{candi}}\) for minimizing the objective function is determined by the following condition:

$${X_{candi}}=\left\{ {\begin{array}{*{20}{c}} {\frac{{LB,~\sigma <0~and~a \geqslant \left( {UB+LB} \right)/2}}{{UB,~\sigma <0~and~a<\left( {UB+LB} \right)/2}}} \\ {LB,~\sigma >0~and~a<LB} \\ {\frac{{a,~\sigma >0~and~LB \leqslant a \leqslant UB}}{{UB,~\sigma >0~and~a>UB}}} \end{array}} \right.$$
(18)

where \(UB\) and \(LB\) represent the upper and lower bounds of the search space, respectively.

By employing the concept of kernel mapping to address the inherent limitations of HHO, such as slow convergence and susceptibility to local optima, the position of the current candidate solution is updated after determining the kernel function parameters as follows:

$${X_i}\left( {t+1} \right)=\left\{ {\begin{array}{*{20}{c}} {{X_{candi}},~if~f\left( {{X_{candi}}} \right)<f\left( {{X_{rabbit}}\left( t \right)} \right)} \\ {{X_{rabbit}}\left( t \right)+rand \cdot {e^{ - FEs/MaxFEs}} \cdot \left( {{X_{candi}} - {X_{rabbit}}\left( t \right)} \right),~otherwise} \end{array}} \right.$$
(19)

where \(MaxFEs\) denotes the maximum number of evaluations, \(FEs\) represents the current evaluation number, \({X_{candi}}\) denotes the candidate solution obtained through kernel mapping, and \({X_{rabbit}}\left( t \right)\) represents the current optimal solution.

In summary, this study introduces a kernel-search HHO algorithm (KSHHO), with the algorithm’s specific steps outlined in Appendix A.1. The detailed process is as follows:

Step 1: Initialize two key parameters: population size and the maximum number of evaluations. Then, generate random initial position vectors for \(\:N\) hawks.

Step 2: In each optimization round, calculate the fitness value of each hawk in the current population.

Step 3: Set the position of the hawk with the highest fitness value in the current population as the position of the rabbit.

Step 4: Update the energy value \(\:E\) of the rabbit according to Eq. (3) and randomly generate the jump intensity \(\:J\). The energy value \(\:E\) decreases as the number of evaluations increases, simulating the energy expenditure of the prey during its escape process.

Step 5: Determine whether the algorithm should enter the exploration or exploitation phase based on the energy value \(\:E\). When \(\:\left|E\right|\ge\:1\), the algorithm enters the exploration phase, during which the hawks perform wide-range random searches. When \(\:\left|E\right|<1\), the algorithm enters the exploitation phase, where the hawks focus on searching around the known optimal solution.

Step 6: In the exploration phase, hawks update their positions based on Eq. (1) to expand the search range and explore new solution spaces. In the exploitation phase, hawks update their positions based on the encircling strategy in Eq. (4) to (13), gradually approaching the optimal solution.

Step 7: To prevent convergence to local optima and improve the diversity of candidate solutions, the kernel search method is applied to update the hawk’s position according to Eq. (19).

Step 8: After each iteration, verify if the stopping criteria are met. If not, repeat Steps 2 to 7.

Step 9: Return the position of the hawk with the best fitness in the current population as the global optimal solution of the algorithm.

Binary Harris hawks optimizer with kernel search

In the above content of this study, KSO is used to improve HHO and proposed KSHHO. In this section, a FS model based on KSHHO and KELM (bKSHHO-KELM) is proposed. In the FS problem, 1 or 0 is used to indicate the selection or non-selection of a feature. Therefore, the FS task is a binary problem. In this case, the KSHHO mentioned above for continuous problems needs to be modified to a version suitable for solving discrete problems. Thus, the transfer function (TF) is leveraged to update the position of the search individual and revise its position to the binary search space. Specific definitions of TF and the position correction function are as follows:

$$V=\left| {{\text{erf}}\left( {\frac{{\sqrt \pi }}{2}x} \right)} \right|$$
(20)
$${x_b}=\left\{ {\begin{array}{*{20}{c}} {\sim x,~if~r<V} \\ {x,~otherwise} \end{array}} \right.$$
(21)

where \(erf\left( \cdot \right)\) represents the error function and r is a random number in the range [0,1]. Equation (20) calculates the error function value V of the input search individual position applicable to continuous problems. Subsequently, by comparing the random number r, its corresponding position in the discrete space is determined.

In the process of finding the optimal feature subset, the fitness function is used to determine whether the current subset is optimal. The fitness function is defined in Eq. (22). The objective of FS is to identify a subset of features that best captures the data’s characteristics and exhibits high predictive performance. To achieve this goal, the fitness function consists of two parts: the classification error rate and the number of feature subsets. Two weight coefficients \({w_1}\) and \({w_2}\) are used to assign weights to these two parts. Therefore, in the process of searching for feature subsets, the goal is to minimize the fitness function used.

$$Fit\_fun={w_1} \times Err+{w_2} \times \frac{{ns}}{{nf}}$$
(22)

where \({w_1}\) and \({w_2}\) are 0.99 and 0.01, respectively.

Appendix A.2 shows the flowchart of bKSHHO-KELM. First, the feature extraction methods are used to extract features from the waveform image. This study uses three feature extraction methods: CM32, GLCM33, and LBP34 to extract texture, color, shape and other features of microseismic and blasting images. Specifically, CM extract the color features of an image by calculating the statistical moments of color distribution in the image. The GLCM constructs a statistical representation that characterizes the interactions between pairs of pixels with identical or similar gray levels at specified orientations and separations, thereby enabling the extraction of an image’s texture features. LBP generate a binary code by assessing the intensity of a central pixel relative to its surrounding neighbors. For each individual pixel, the grayscale value of its neighboring pixels is compared to that of the central pixel. LBP generates an encoding that identifies specific local features within the original image, including edges, corners, flat regions, and feature points.

Then, the FS process and prediction model training are carried out. The data set is divided into a training set and a test set. This study adopts ten-fold cross validation (CV) to avoid the influence of randomness on the model performance. During the training process, the population is first initialized and the parameters in the bKSHHO algorithm are set. Then, the bKSHHO algorithm searches for feature subsets by simulating the hunting behavior and kernel search strategy of Harris hawk. And the algorithm evaluates the fitness value of each feature subset (individual). The bKSHHO algorithm updates the position of the individual based on fitness value, thereby gradually optimizing the feature subset. The updating process continues until the maximum number of iterations is reached. And the current optimal feature subset is output. In the model training process, this study uses KELM35 as a classifier. KELM is a machine learning method based on extreme learning machine, which enhances its expressive power on nonlinear problems by introducing kernel functions.

Based on the selected optimal feature subset, the model is then evaluated using the test set and the corresponding evaluation indicators are calculated. After completing the evaluation of one fold, it is determined whether the training and verification process of all ten folds has been completed. If not all folds have been completed, the training and evaluation process of the next fold will be entered; if all folds have been completed, the average of the evaluation indicators of the 10 folds will be calculated as the final evaluation result of the model.

Experiment and results

Optimization capability validation via benchmark functions

Experimental setup

There are 12 test functions in CEC 2022 benchmark function36, including one unimodal function (F1), four basic functions (F2–F5), three hybrid functions (F6–F8), and four composition functions (F9–F12). Each function operates within a 20-dimensional space. Detailed descriptions of these test functions are presented in Table 2.

KSHHO will compare with ten SOTA algorithms, including 6 SOTA metaheuristic algorithm variants (HG_SMA37, ACWOA38, ALCPSO39, CGSCA40, ISNMWOA41, SRWPSO42 and 4 other related original metaheuristic algorithms (HHO43, RIME44, FATA45, PSO14, to evaluate the performance of KSHHO. Table 3 summarizes the main parameter settings of all algorithms. All parameter values of the comparative algorithms were taken directly from their respective original articles.

For fairness of comparison, all algorithms were limited to the maximum number of \(\:10000\times\:D\), where \(\:D\) is the problem dimension of the. Additionally, each algorithm on each test function was executed independently 30 times to reduce experimental errors stemming from the algorithms’ inherent randomness. Differences between algorithms were evaluated by analyzing the average optimal convergence value and standard deviation obtained from 30 independent runs. The significance of the differences between algorithms was compared by Wilcoxon signed-rank test (WSRT)46 and a Friedman test (FRT)47, with the significance level set at 0.05. The symbols “+∕-∕=” respectively indicate that KSHHO is better than, worse than or equal to the competing algorithms, while “b/w/s” counts the number of its corresponding test functions respectively. To visually demonstrate the algorithm’s convergence process, we plotted a convergence curve. Additionally, a box plot was used to represent the overall distribution of the optimal solutions obtained from 30 runs.

Table 2 The details of CEC 2022 benchmark test functions.
Table 3 Key parameters of different algorithms.

Qualitative analysis of KSHHO

This section will evaluate the optimization performance of KSHHO from three aspects: search history, trajectory result and fitness value. We selected four test functions to present the experimental results. These functions range from simple to complex and cover various optimization characteristics, providing a comprehensive demonstration of the search behavior and adaptability of the proposed algorithm across diverse problem structures. The results are shown in Fig. 3. The first column shows the three-dimensional visualization of the corresponding test function. The second column displays the search history based on x1 and x2, where red dots signify the optimal individual positions and black dots represent the search individuals. The third column illustrates the trajectory of the first search individual, while the fourth column displays the average fitness across all search individuals.

In the early stage of optimization, the trajectory curve of the search individual changes significantly, because at this time it is necessary to search the entire search space as much as possible. Therefore, it can also be found that in the corresponding search history graph, the search individuals basically cover the entire search space. As the number of iterations increases, the trajectory curve gradually becomes stable indicating that algorithm gradually transitions from global exploration to local exploitation, concentrates on exploitation within a certain range, and approaches the global optimal individual. In the composition function, the average fitness of all search individuals levels off for a period of time and then drops sharply, which shows that KSHHO successfully avoids the stagnation of particles in the local optimum.

The above results show that the optimization behavior based on the kernel search strategy in KSHHO can avoid the search individual falling into the local minimum and converge around the global optimum in the final stage.

Fig. 3
Fig. 3
Full size image

Qualitative analysis result of KSHHO. Each row corresponds to one test function \(\:{F}_{i}\).From left to right: (1) the 3D surface plot of \(\:{F}_{i}\), illustrating the function landscape; (2) the search history, where black dots represent visited candidate solutions and the red dot marks the best solution found; (3) the trajectories of agents over iterations, showing convergence behavior in the decision space; and (4) the average fitness curve of all agents, reflecting convergence in the objective space.

Comparison results analysis

Appendix A. 3 shows the average value (Avg) and standard deviation (Std) results of all algorithms. The table results indicate that KSHHO performs relatively well in basic and hybrid functions, among which the optimal solutions obtained in F4 and F8 are due to the solutions obtained by other compared algorithms. Although KSHHO’s performance in unimodal and composition functions is not so outstanding, its optimization ability in these two types of functions is also relatively good among these compared algorithms. Figure 4 illustrates the final ranking outcomes for each algorithm. The proposed KSHHO attained the highest average ranking of 3.42, ranking first, followed by RIME and SRWPSO.

Fig. 4
Fig. 4
Full size image

The average rank result based on Avg and Std between KSHHO and comparison algorithms.

Table 4 WSRT results of KSHHO and other algorithms on CEC 2022.

Table 4 shows the WSRT results of KSHHO in comparison with other algorithms. Results indicate that KSHHO significantly outperforms ACWOA, CGSCA, and FATA in optimization performance across all 10 test functions. Significantly better than HG_SMA, HHO, PSO in 9 test functions. It is worth noting that in the comparison with RIME, KSHHO’s performance is significantly better than RIME in only 4 test functions, and its performance is significantly weaker than RIME in 6 test functions. Table 5 shows the FRT results. The proposed algorithm achieved an average FRT ranking of 3.75, ranking first. RIME followed in second place, and SRWPSO attained the third rank. The above results verify the excellent optimization performance of KSHHO.

Table 5 FRT result of KSHHO with other algorithms.

Figure 5 plots the convergence curves of all algorithms on the test functions. The steeper the slope of the curve relative to the horizontal axis, the faster the convergence rate during the current evaluation interval. Note that most algorithms, such as HHO, ALCPSO, and ISNMWOA, show premature entrapment in local optimality in most test functions. Taking F4 as an example, although ALCPSO and SRWPSO show faster convergence speed at the beginning of the evaluation, ALCPSO falls into the local optimality at about 1000 evaluation times, and the curve tends to stabilize. The algorithm proposed in this paper continues to decline and eventually converges to the minimum value. While KSHHO does not achieve the highest convergence accuracy on every test function, its overall convergence performance surpasses that of the other algorithms.

Figure 6 presents the box plot of the optimal values obtained by all algorithms across 30 independent runs. It is evident that the shape for KSHHO is more compact, with fewer outliers. It shows that the optimal value found by KSHHO in 30 independent runs has a smaller range of variation and better stability. However, the box lengths of the comparison algorithms in different test functions are different and the stability is poor. In summary, the KSHHO has better optimization capabilities than other algorithms.

Fig. 5
Fig. 5
Full size image

Convergence curves of KSHHO and comparison algorithms in different test functions.

Fig. 6
Fig. 6
Full size image

Box plot of KSHHO and comparison algorithms in different test functions.

Case study on microseismic and blasting phenomena

Experiment setup

This study employs techniques such as CM, GLCM, and LBP to extract the shape features from waveform image data obtained by each sensor in microseismic and blasting signals. To improve the convergence of the model and eliminate the influence of different dimensions, the extracted features were normalized. All feature values were mapped to the range [-1, 1] to ensure equal contribution of each feature during model training. To ensure fairness and objectivity of the results, we use a stratified 10-fold CV method on the entire dataset to ensure fairness and objectivity. In each CV iteration, the data are randomly partitioned into 10 equal folds. One fold is held out as the test set, and the remaining nine folds form the training set. After cycling through all ten folds, we report the mean and standard deviation of the evaluation metrics across folds as the final result.

To evaluate the performance of the bKSHHO-KELM method in predicting microseismic and blasting signals, several commonly used classification metrics were applied, including accuracy, recall, precision, and F1 score. The detailed formulas for these metrics are provided in Table 6. The method was compared with several baseline models, including light gradient boosting machine (LightGBM)48, categorical boosting (CatBoost)49, random forest (RandomF)50, adaptive boosting (AdaBoost), fuzzy k-nearest neighbors (FKNN)51, extreme gradient boosting (XGBoost)52, and KELM35, as well as its counterparts, including binary specular reflection learning enhanced HHO with KELM (bHHOSRL-KELM)53, binary PSO with KELM (bPSO-KELM)54, binary moth-flame optimization with KELM (bMFO-KELM)55, binary ant colony optimization with KELM (bACO-KELM)56, bDE-KELM, binary artificial bee colony with KELM (bABC-KELM)57, and bHHO-KELM. Due to the FS technique, the algorithm’s dimensionality was adjusted to correspond to the number of features in the dataset, with a population size of 20. All implementations were executed in the same simulation environment.

Table 6 Model evaluation metrics.

Simulation results and analysis

This section aims to perform classification prediction using the extracted microseismic and blasting data, evaluating the effectiveness of the proposed bKSHHO-KELM method in real-world seismic forecasting. A comparative experiment was conducted between the bKSHHO-KELM method and traditional classification techniques (e.g., LightGBM, XGBoost, CatBoost, RandomF, FKNN), evaluating its accuracy, recall, precision, and F1 score.

Figure 7 illustrates the performance of the bKSHHO-KELM method and seven other classification models in predicting microseismic and blasting phenomena, with box plots showing the distribution of each evaluation metric. It is evident that the bKSHHO-KELM method outperforms the other models across all metrics, particularly in accuracy and F1 score, demonstrating its effectiveness in predicting microseismic and blasting phenomena. Notably, the LightGBM and XGBoost models also exhibited relatively good performance in predicting microseismic and blasting phenomena. Furthermore, the FKNN model performed relatively poorly, particularly in terms of stability, with significant fluctuations in its predictions.

Table 7 presents the performance metrics of the bKSHHO-KELM method and other models in predicting microseismic and blasting phenomena. The bKSHHO-KELM method clearly outperforms the other models in predicting microseismic and blasting phenomena. Its accuracy (95.625%) and recall (93.964%) are higher than those of LightGBM (94.197% and 90.033%) and XGBoost (94.197% and 90.061%), while its F1 score (0.931) also surpasses both (0.910). Additionally, the bKSHHO-KELM method has a smaller standard deviation, indicating more stable prediction results. In contrast, the FKNN model performed the worst, with an accuracy of only 90.326% and poor stability, as reflected by its larger standard deviation, further highlighting the superiority of bKSHHO-KELM.

Fig. 7
Fig. 7
Full size image

Box plots of the results of bKSHHO-KELM with other models predicting microseismic and blasting phenomena.

Table 7 Avg and Std results of bKSHHO-KELM versus other models for predicting microseismic and blasting phenomena performance metrics.

The above experiments demonstrate that the wrapper-based FS method, which combines the KELM classifier with the proposed bKSHHO algorithm, exhibits significant performance advantages in predicting microseismic and blasting phenomena compared to traditional classification methods, effectively addressing the limitations of individual classifiers. To evaluate the effectiveness of the bKSHHO-KELM method compared to other algorithms such as bHHOSRL, bPSO, bMFO, bACO, bDE, bABC, and bHHO, the goal is to determine whether the bKSHHO-KELM method outperforms these algorithms in predicting microseismic and blasting phenomena.

Figure 8 presents the performance of the bKSHHO-KELM method and various peer methods in predicting microseismic and blasting phenomena. Overall, all methods exhibited good predictive performance, but the bKSHHO-KELM method clearly outperformed the others across several performance metrics. According to the results in Table 8, bKSHHO-KELM outperforms other optimized KELM variants in terms of accuracy (95.625%) and recall (93.964%), leading bHHOSRL-KELM (accuracy 94.783%, recall 92.026%) and bPSO-KELM (accuracy 94.619%, recall 91.685%). Moreover, bKSHHO-KELM also outperforms all other methods in terms of precision (92.632%) and F1 score (0.931), particularly when compared to bABC-KELM (precision 90.526%, F1 score 0.902) and bDE-KELM (precision 90.526%, F1 score 0.908), demonstrating superior balance and overall performance.

Fig. 8
Fig. 8
Full size image

Box plots of the results of bKSHHO-KELM with peer methods predicting microseismic and blasting phenomena.

Table 8 Avg and Std results of bKSHHO-KELM versus peer methods for predicting microseismic and blasting phenomena performance metrics.

Table 9 presents the detailed experimental results of the bKSHHO-KELM method in predicting microseismic and blasting phenomena. The first column represents the fold numbers of the 10-fold cross-validation, and the second column shows the size of the optimal feature subset obtained by the bKSHHO method. The remaining columns show the accuracy, recall, precision, and F1 score used in the analysis.

From the experimental results, the bKSHHO-KELM method shows stable performance in 10-fold cross-validation, with an average accuracy of 95.625%, recall of 93.964%, precision of 92.632%, and F1 score of 0.931. The fluctuations between folds are minimal, with standard deviations of 0.019, 0.052, 0.043, and 0.028, indicating high stability. Overall, the prediction performance is excellent, demonstrating the effectiveness and reliability of the method in predicting microseismic and blasting phenomena.

Table 9 Ten-fold crossover results of the proposed bKSHHO-KELM method for predicting microseismic and blasting phenomena.

Discussion

The proposed bKSHHO-KELM model has shown promising performance in classifying microseismic events and blasting events in underground mining environments. While the experimental results are encouraging, several important aspects deserve further discussion to provide a more comprehensive understanding of the model’s strengths, limitations, and future research directions.

Although the bKSHHO-KELM model achieves high accuracy, recall, precision, and F1 scores while using a relatively small subset of features, several limitations need to be acknowledged. First, the feature selection process is based on bKSHHO, which is sensitive to the randomness of the initial population. This can lead to slight variability in selected feature subsets and classification performance across different runs. To address this, future studies could explore more robust initialization strategies, such as opposition-based learning or adaptive sampling techniques, to improve consistency and solution quality58,59. Second, the optimization performance of KSHHO is limited on unimodal and composition benchmark functions, as demonstrated in our evaluation using the CEC 2022 test suite. This indicates that the algorithm may lack sufficient global exploration or local exploitation in specific problem landscapes. Future research could consider hybridizing bKSHHO with local search techniques or incorporating adaptive parameter control mechanisms to enhance both global and local search capabilities. Moreover, the current model focuses only on classifying two types of events (microseismic and blasting), and its performance on detecting or identifying other rare or unknown seismic events remains unexplored. Expanding the classification framework to include multi-class capabilities60,61 or anomaly detection models62 could improve its applicability in more diverse mining scenarios.

In real underground mining environments, seismic monitoring systems inevitably capture a wide variety of signals beyond the intended microseismic and blasting events. These additional signals may originate from drilling operations, machinery movement, transportation of equipment, and other mechanical systems such as ventilation fans or conveyor belts63,64. Such signals often exhibit relatively stable frequency characteristics or low-frequency continuous vibrations, which can overlap with the time-frequency features of genuine seismic events, leading to potential misclassification by automated systems65. The presence of these non-target signals poses a significant challenge to the robustness and reliability of seismic event classification. Without careful preprocessing and discrimination, the model may incorrectly label operational noise or man-made disturbances as seismic phenomena, thereby reducing its practical effectiveness in safety monitoring. Addressing this issue requires the integration of advanced signal processing techniques capable of capturing subtle distinctions in frequency content and temporal patterns. Approaches such as time-frequency analysis66, wavelet transform67, empirical mode decomposition68, and even deep learning-based feature extraction69,70 could provide richer representations of the signals, enhancing the separability of different source types. Additionally, the adoption of unsupervised learning71 or clustering methods72 could enable the automatic detection of unknown or mixed signal types, further improving the adaptability of the classification system. To strengthen the model’s resilience in real-world deployment, future work should focus on constructing a more comprehensive and representative dataset that encompasses a broader spectrum of seismic and non-seismic sources encountered in mining environments. Such efforts would facilitate the development of more generalizable models capable of maintaining high classification accuracy even under complex and noisy conditions.

In recent years, deep learning methods such as convolutional neural networks, long short-term memory networks, and transformer-based models have been successfully applied to seismic signal classification, fault detection, and event localization73,74,75. These models are capable of learning complex spatial-temporal patterns from raw data, often achieving high performance in large-scale applications. However, deep learning models typically require large labeled datasets, high computational resources, and may lack interpretability, which poses challenges in underground mining environments where real-time decision-making and limited data are common constraints. In contrast, our bKSHHO-KELM model offers the advantage of achieving high prediction performance using fewer features and lower computational cost, making it more suitable for practical deployment in resource-constrained settings. In future work, it may be beneficial to explore hybrid approaches that combine the interpretability and efficiency of bKSHHO-KELM with the powerful feature extraction capabilities of deep learning models.

The proposed bKSHHO-KELM model demonstrates strong potential for microseismic and blasting event classification using compact and efficient feature subsets. This makes it well-suited for deployment in underground mines where computational resources are limited, and timely decision-making is crucial. However, to fully validate the generalizability of the model, further testing is needed across multiple mining sites, geological conditions, and varying noise environments. Future studies could also apply this methodology to other geophysical applications, such as earthquake early warning, structural health monitoring, or landslide detection, where reliable event classification is critical. Additionally, incorporating online learning, adaptive signal separation, and anomaly detection into the current framework could further enhance the system’s robustness and broaden its application scope.

Conclusion and future works

This study introduces a microseismic blasting prediction model that integrates an HHO variant with KELM. Firstly, a kernel search strategy is employed to strengthen the search efficiency of the original HHO, effectively avoiding local optima and enhancing the algorithm’s convergence accuracy. CEC 2022 benchmark function is used to evaluate the performance of KSHHO comparing with 10 SOTA algorithms. The experimental results show that the optimization performance of the proposed KSHHO is better than the comparison algorithm. Subsequently, the bKSHHO-KELM prediction model is proposed. Experimental comparisons are carried out with classical classifiers on microseismic blasting data. The experimental results show that the proposed bKSHHO-KELM performs well, with accuracy, recall, precision, and F1 of 95.625%, 93.964%, 92.632%, and 0.931 respectively.

There are several potential directions for future research regarding our proposed method. First, from the optimization experimental results, it can be found that KSHHO lacks the ability to find the best in unimodal and composition functions. Therefore, in the future, we can consider adopting more effective strategies to improve the optimization ability of the algorithm in these two types of test functions. Secondly, the quality of the optimal solution produced by the algorithm is affected by the randomness of the initial solution generation. To solve this, a novel initialization strategy could be developed in the future to improve the quality of the optimal solutions. Finally, we can apply the proposed model to other optimization application fields to further verify and expand the application scope of the proposed model.