Abstract
Randomization-based neural networks have gained wide acceptance in the scientific community owing to the simplicity of their algorithm and generalization capabilities. Random vector functional link (RVFL) networks and their variants are a class of randomization-based neural networks. RVFL networks have shown promising results in classification, regression, and clustering problems. For real-world applications, learning algorithms that can train with new samples over previous results are necessary because of to the constant generation of problems related to large-scale datasets. Various online sequential algorithms, commonly involving an initial learning phase followed by a sequential learning phase, have been proposed to address this issue. This paper presents a training algorithm based on multiple online sequential random vector functional link (OS-RVFL) networks for large-scale databases using a shared memory architecture. The training dataset is distributed among p OS-RVFL networks, which are trained in parallel using p threads. Subsequently, the test dataset samples are classified using each trained OS-RVFL network. Finally, a frequency criterion is applied to the results obtained from each OS-RVFL network to determine the final classification. Additionally, an equation was derived to reasonably predict the total training time of the proposed algorithm based on the learning time in the initial phase and the time scaling factor compared to the sequential learning phase. The results demonstrate a drastic reduction in training time because of data distribution and an improvement in accuracy because of the adoption of the frequency criterion.
Similar content being viewed by others
Introduction
The random vector functional link (RVFL) network is an artificial neural network (ANN) that belongs to the family of randomization-based feed-forward neural networks1,2. This algorithm was proposed by Pao et al.3,4 and is characterized by direct links from the input layer to the output layer, with randomly assigned weights and biases in the hidden layer. Because of these features, particularly the direct links from the input layer to the output layer, the overall performance of the RVFL network is better than that of other such networks without direct links5. The popularity and acceptance of the RVFL network has increased among the scientific community because of the simplicity of the model and its capacity for generalization to classification, regression, and clustering problems. However, the large database of RFVL networks poses a challenge as the training times are considerably longer and high-cost computational architectures are required6.
Various architectures and variants of the RVFL network, including RVFL for imbalance learning, kernelized RVFL, RVFL for semisupervised learning, online RVFL, ensemble learning, robust RVFL, ensemble deep RVFL, hybrid RVFL, and deep RVFL, are reported in literature7. The focus of the present study was on sequential online algorithms and ensemble models. Liang et al.8 proposed a sequential online algorithm for RVFL without direct links. This algorithm was proposed for applications where training data are entered one-by-one or chunk-by-chunk. The algorithm updates the training using new training data and the previous results without utilizing all the accumulated training data. The results showed that these algorithms are faster than other sequential algorithms. Other sequential online algorithms without direct connections in RVFL have been reported in literature. Matias et al.9 proposed a sequential online algorithm based on recursive least squares, while Mirza et al.10 introduced a voting-based weighted online sequential algorithm for imbalanced multiclass classification. Recently, Gelvez-Almeida et al.11,12 proposed a parallel training approach for a set of online sequential algorithms tailored for large-scale datasets using a fingerprint dataset. Additionally, Wibawa et al.13 used a model predictive control approach to modify the standard online sequential model. Other recent contributions to online sequential RVFL (OS-RVFL) include the works by Chen and Li14, Zhang et al.15, Zha et al.16, Kale et al.17, and Polat et al.18.
Several ensemble models have been developed19. Lan et al.20 introduced an ensemble of OS-RVFL without direct links; this ensemble model was more stable and accurate than the standard online sequential model proposed by Liang et al.8 In this ensemble model, multiple online sequential networks are trained, and the average of their outputs is used as the performance of the network. Liu and Wang21,22, and Wei et al.23 proposed an ensemble-based RVFL network that incorporates cross validation and a criterion based on the norm of network output weights. Zhai et al.24 developed an algorithm for integrating the standard OS-RVFL into classification problems with large datasets. Alhamdoosh and Wang25 employed RVFLs as the base components and combined them with the negative correlation learning strategy to construct neural network ensembles. Their technique was more effective and efficient than other ensemble techniques. Subsequently, Mirza et al.26 presented an ensemble of a subset of OS-RVFL for addressing class imbalance and concept drift. This proposal processes the minority classes with multiple classifiers, while the majority classes are processed in a round-robin fashion. Latter, Ling et al.27 proposed an improved ensemble of RVFL networks based on particle swarm optimization with a double optimization strategy, and Huang et al.28 proposed a parallel ensemble method based on MapReduce for large-scale learning.
Next, Rakesh and Suganthan29 proposed an ensemble of kernel ridge regression using the RVFL network to generate training samples for multiclass classification. Zhang and Suganthan30 introduced an efficient co-trained kernel ridge regression method and presented an ensemble of RVFL networks. Li et al.31 proposed a parallel one-class approach based on the Bayesian approach. Katuwal and Suganthan32 proposed an ensemble of RVFL networks by incorporating additional regularization or randomization through Dropout and DropConnect techniques. Then, Li et al.33 proposed a novel ensemble that initializes its base learners using different distributions to enhance their diversity. Huet al.34 introduced an adaptive ensemble variant of the RVFL network. Malik et al.35 combined the rotation forest and RVFL classifiers into an ensemble method for classification problems. Tanveer et al.36 proposed ensemble classifiers with RVFL using multiple SVD models. Shi et al.37 proposed deep-learning frameworks based on the RVFL network.
Meanwhile, online sequential algorithms and ensemble models based on RVFL have been used several real-world applications, such as online adaptive humidity monitoring38, industrial processes39, eye-tracking-based situation awareness recognition40, diagnosis of Alzheimer’s disease41, short-term electric-load forecasting42, landslide displacement prediction43, drought index forecasting44, turbofan engine direct thrust control45, low-resolution real-time face recognition46, cross-person and cross-position activity recognition47, laminar cooling water supply system for hot rolling mills driven by digital twin for energy-saving48, lane-changing control of vehicle platoon49, sediment transport in sewer pipes50, stock index trend prediction51, battery state of health estimation and remaining useful life prediction52, etc53,54,55,56,57,58,59,60,61,62.
Based on the previous works, we present an ensemble of OS-RVFL networks for classification problems with large-scale bases on a frequency criterion (EOS-RVFL-FC). The proposed model involves training several OS-RVFL networks in parallel via multithread computing, classifying the samples tested with all trained OS-RVFL networks, and finally, selecting from the individual results the classes with the highest frequency. This model is more accurate than the original online sequential algorithm. Further, the training time decreases when the training data are distributed and remains constant when they are replicated. To evaluate the efficiency of our proposal, we used a balanced fingerprint dataset. In addition, we used five datasets that have been widely used to validate randomized neural networks algorithms: MNIST, image segmentation, Adult, satellite image, and Mushroom. Four of these datasets are balanced. The main contributions of our article are as follows:
-
We proposed a model that leverages the advantages of multithreaded programming to train several OS-RVFL networks in parallel, thus reducing the training time when training data are distributed. A frequency criterion is used to improve accuracy in classification problems.
-
We experimentally demonstrated that the proposed method can effectively improve the computational time of the standard OS-RVFL network, increasing the accuracy of the testing data for all the databases. Thus, applying this method in other randomization-based neural networks will be a major scientific contribution.
-
We derived an equation that can reasonably estimate the behavior of our model based on the threads to be used. The parameters required are the execution time in the initial phase and its relationship with each chunk of the sequential learning phase, the total training samples, and the training samples in each chunk.
The rest of this paper is organized as follows: Section 2 briefly introduces the preliminary concepts, namely the RVFL network and its sequential online proposal. Section 3 describes our proposed model, including its algorithm and a graphical overview. Section 4 presents the experimental aspects, including a description of the databases, hyperparameter estimation, and the results. Finally, the conclusions and future studies are presented in Section 5.
Preliminaries
In this section, we present relevant previous works. We provide a brief description of the mathematical frameworks of RVFL networks introduced by Pao et al.3,4, as well as the sequential online models proposed by Liang et al.8
Random vector functional link network
RVFL is a single-layer feed-forward neural network that randomly assigns weights and biases to the hidden layer and analytically calculates the weights of the output layer. Let \({\textbf{Z}}\) be an arbitrary training set \({\textbf{Z}} = \{ ({\textbf{x}}_i, {\textbf{y}}_i) | {\textbf{x}}_i \in {\mathbb {R}}^d, {\textbf{y}}_i \in {\mathbb {R}}^c \}\) with \(i = 1, \ldots , N\), where \({\textbf{x}}_i\) represents the i-th training sample; \({\textbf{y}}_i\), the i-th target; d, the features of each sample; c, the number of classes; and N, the total number of samples. In the training process of the standard RVFL network, the three layers are connected as shown in Fig. 1. The input layer and the output layer are connected through randomly assigned weights and biases, while the output layer is connected to the other layers through analytically calculated weights. The training algorithm of the standard RVFL network can be written as follows:
where \(\varvec{\mu }_k\) and \(\sigma _k\) are the k-th weights and bias of the hidden layer, respectively; \(\varvec{\beta }_k\), is the k-th weight of the output layer; \(\varvec{\mu }_k \cdot {\textbf{x}}_i\) represents the inner product of \(\varvec{\mu }_k\) and \({\textbf{x}}_i\); L is the number of neurons in the hidden layer.
Standard model of the random vector functional link network. The red lines show the connection between the input layer and the output layer; the blue lines represent the connection between the input layer and the hidden layer; the black line represents the connection between the hidden layer and the output layer. The weights \(\varvec{\mu }\) and biases \(\sigma\) are randomly assigned, while the weights \(\varvec{\beta }\) are computed analytically.
The regularized optimization problem for a standard RVFL network with L neurons in the hidden layer can be written as follows:
where \({\textbf{H}} = [{\textbf{D}} {\textbf{X}}]\) is the concatenation of hidden features and original features, and C is the regularization parameter. Here, \({\textbf{X}} = [{\textbf{x}}_1, {\textbf{x}}_2, \ldots , {\textbf{x}}_N]^T\) is the training dataset, \({\textbf{Y}} = [{\textbf{y}}_1, {\textbf{y}}_2, \ldots , {\textbf{y}}_N]^T\) is the target matrix, and the output matrix of the hidden layer \({\textbf{H}}\) is given as shown below:
The output layer weights \(\varvec{\beta } = [\varvec{\beta }_1, \varvec{\beta }_2, \ldots , \varvec{\beta }_{(d+L)}]^T\) are calculated analytically from
where \({\textbf{H}}^{\dagger }\) is the Moore–Penrose generalized inverse of the \({\textbf{H}}\) matrix. In RVFL networks, the Moore–Penrose generalized inverse matrix of \({\textbf{H}}\) is computed as follows:
where \({\textbf{H}}^T {\textbf{H}}\) and \({\textbf{H}} {\textbf{H}}^T\) are symmetric positive semidefinite matrices and \(C > 0\)7.
Online sequential random vector functional link
This algorithm is a variant of the RVFL networks for real-world applications as the data for training are received in a chunk-by-chunk or one-by-one (a special case of chunks) manner8. The training algorithm of this model involves an initial phase and a sequential learning phase. Fig. 2 shows a general outline of this algorithm.
Initialization phase
Let an initial chunk of training samples \({\textbf{Z}}_0 = \{ ({\textbf{x}}_i, {\textbf{y}}_i) | {\textbf{x}}_i \in {\mathbb {R}}^n, {\textbf{y}}_i \in {\mathbb {R}}^c \}\) with \(i = 1, \ldots , N_0\), and L neurons in the hidden layer with \(L \le N_0\), where \(N_0\) is the number of examples in the initial training chunk \({\textbf{Z}}_0\). Thus, random weight \(\varvec{\mu }_i\) and bias \(\sigma _i\) are assigned. Then, the initial output matrix \({\textbf{H}}_0\) of the hidden layer is calculated. Finally, the initial weights \(\varvec{\beta }_0\) of the output layer are computed as follows:
where \({\textbf{K}}_0 = {\textbf{H}}_0^T {\textbf{H}}_0 + {\textbf{I}}/{C}\), and \({\textbf{Y}}_0\) is the target matrix of the initial training chunk \({\textbf{Z}}_0\).
Sequential learning phase
Let us consider the second chunk of training samples \({\textbf{Z}}_1 = \{ ({\textbf{x}}_i, {\textbf{y}}_i) | {\textbf{x}}_i \in {\mathbb {R}}^n, {\textbf{y}}_i \in {\mathbb {R}}^c \}\) with \(i = N_0 + 1, \ldots , N_0 + N_1\). Here, \(N_1\) is the number of training samples in the second training chunk \({\textbf{Z}}_1\). Here, the weights of the output layer \(\varvec{\beta }_1\) are updated as follows:
where \({\textbf{K}}_1 = {\textbf{K}}_0 + {\textbf{H}}_1^T {\textbf{H}}_1\)8.
In general, the sequential learning phase presents the \((k + 1)\)-th training chunk \({\textbf{Z}}_{k+1} = \{ ({\textbf{x}}_i, {\textbf{y}}_i) | {\textbf{x}}_i \in {\mathbb {R}}^n, {\textbf{y}}_i \in {\mathbb {R}}^c \}\) with \(i = (\sum _{j=0}^k N_j) + 1, \ldots , \sum _{j=0}^{k+1} N_j\), where \(N_{k+1}\) is the number of training samples in the \((k + 1)\)-th training chunk. Then, the partial hidden layer output matrix \({\varvec{H}}_{k + 1}\) for the \((k + 1)\)-th training chunk is calculated. Finally, the weights \(\varvec{\beta }_{k+1}\) of the output layer are computed using the following equation:
where \({\textbf{P}}_k = {\textbf{K}}_k^{-1}\)8. The sequential learning phase ends when \(\varvec{\beta }_{k+1}\) is computed with the last chunk.
Proposed parallel ensemble method
Various ensemble-based RVFL models are reported in literature7. In this paper, we propose a model that combines the advantages of ensemble models, sequential online algorithms, RVFL networks, and high-performance computing. Our model involves training multiple OS-RVFL networks in parallel by assigning the training phases of each OS-RVFL to a thread through a shared memory architecture. Additionally, we implement data distribution to reduce the training time, considering that we are using a large-scale database. The OS-RVFL version used in this research is the one proposed by Liang et al.8, which derives from the RVFL variant without direct link between the input and output layer63. Finally, we use a frequency criterion for the final classification of the testing data to consider the results obtained from each neural network. Fig. 3 shows an overall framework of our parallel ensemble proposal. The algorithm for our parallel ensemble proposal is presented as Algorithm 1, which can be summarized in the following steps:
-
Step 1. The training data are divided into p subsets, which are processed in parallel by independent threads. Mathematically, this distribution can be represented as follows:
$$\begin{aligned} \bigcup _{i=1}^{p} {\textbf{Z}}_i = {\textbf{Z}}, \end{aligned}$$(9)where \({\textbf{Z}}_i\) denotes each training subset, and \({\textbf{Z}}\) represents the complete training set.
-
Step 2. Each subset is used to train an individual OS-RVFL network. Each neural network operates independently with its own set of training data and randomly initialized weights and biases.
-
Step 3. Each OS-RVFL network performs the classification of the same set of test data. As each neural network operates independently, the individual accuracy may vary, leading to potential variation in the results.
-
Step 4. A frequency criterion is used to analyze the outputs obtained in step 3. The criterion involves selecting the output with the highest occurrence frequency among each neural network result.
-
Step 5. After applying the frequency criterion in step 4, the final classification is obtained.
This approach is suitable for both distributed and replicated data scenarios. When replicated data are used, each OS-RVFL network is trained using the exactly same training data. Each OS-RVFL network is trained with randomly assigned weights, rendering each network independent and exhibiting varying accuracies, thus making the frequency criterion effective. Replicated data are particularly advantageous when dealing with a limited number of samples. In the realm of computational complexity, OS-RVFL is predominantly shaped by the quantity of examples scrutinized during the training phase64. Analogous to standard RVFL and its variants, OS-RVFL equally showcases a low training complexity, notwithstanding the necessity for the algorithm to execute many iterations65,66,67. Our model upholds the computational efficiency of OS-RVFL, as the training of distinct networks is undertaken autonomously and parallel.
Experiments
In this section, we provide details about the dataset used and outline the hyperparameter estimation procedure. All experiments were conducted on a server equipped with 2\(\times\)Intel(R) Xeon(R) Gold 6238R CPUs @ 2.20 GHz and with 128 GB RAM. The implementation was written in C++ programming language, using OpenMP to enable parallel processing with shared memory.
Description of the datasets
In this study, we used a balanced dataset composed of synthetic fingerprint descriptors. The fingerprint descriptors were generated using a feature extractor based on FingerCode, singularities, and pseudo ridges described in a previous work68. The dataset has five distinct categories69, namely, arch, left loop, right loop, tented arch, and whorl. These categories are shown in Fig. 4. Each category has a different frequency of occurrence within the total population. However, for the purpose of this work, we used a dataset where an equal number of descriptors were available for each class, ensuring a balanced dataset. Each descriptor in the dataset consists of a vector of 202 double-precision type values representing its characteristics, along with the corresponding target. The dataset comprises a total of 210,000 samples, divided into three sets: 200,000 samples for training, 10,000 samples for testing, and an additional 60,000 samples for hyperparameter estimation.
Fingerprint patterns in a population70. Fingerprints exhibit distinct patterns that can be classified into five major types: arches, loops, whorls, tented arches, and radial loops. These patterns occur in different proportions within the total population, reflecting the unique distribution and prevalence of each fingerprint type.
To further evaluate the performance of our model under diverse conditions, we conducted experiments using five publicly available datasets. These datasets were selected to provide a comprehensive evaluation of our approach:
-
MNIST: This dataset is a widely used benchmark in the field of image classification. It contains a large collection of handwritten digit images, with a training set comprising 60,000 samples and a separate testing set consisting of 10,000 samples. Each image in the dataset is represented as a gray-scale image of dimensions \(28 \times 28\) pixels. To facilitate analysis and processing, the images are vectorized, yielding a feature vector with 784 attributes. The MNIST dataset serves as an excellent test bed for evaluating the performance of our proposed model on the task of digit recognition. The distribution of samples varies, with the most represented class having 6265 samples and the least represented class having 5421 samples. Despite this variation, each class is well represented, ensuring a balanced evaluation of the model’s performance across different digits.
-
Image segmentation: This dataset is a collection of images annotated and labeled for segmentation tasks. Each image in the dataset represents a real-world scene and is accompanied by corresponding segmentation masks, which indicate the pixel-level boundaries of different objects or regions within the image. Similarly to the MNIST dataset, this dataset is well-balanced, containing 300 samples for each class.
-
Adult: This dataset is a comprehensive collection of demographic and socioeconomic information of individuals. It encompasses a wide range of features, such as age, education level, occupation, marital status, and income. This dataset offers valuable insights into the various factors that influence the socioeconomic dynamics of a population. In contrast to the previously mentioned datasets, this dataset exhibits class imbalance, with one class representing 76.07% of the data and the other class 23.93%. This imbalance poses a significant challenge for classification algorithms, highlighting the need for robust model evaluation and selection strategies.
-
Satellite image: This image dataset comprises a diverse collection of high-resolution satellite images captured from different regions across the globe. Each image represents a specific area or landscape, showcasing various geographical features such as urban areas, forests, agricultural regions, and more. This dataset offers invaluable insights into land cover analysis, vegetation patterns, and the overall dynamics of the Earth’s surface. The dataset exhibits an unequal distribution of samples among its classes. The three most frequent classes account for 72.8% of the dataset, with each class representing approximately 24.3% of the samples. In contrast, the three least frequent classes comprise 33.1% of the dataset, suggesting a less balanced distribution among these classes.
-
Mushroom: This dataset is a comprehensive collection of data on diverse species of mushrooms. It encompasses essential features, such as cap shape, cap color, gill size, odor, and habitat. This dataset is popular in the field of classification tasks, specifically in the domain of mushroom identification and toxicity prediction. This dataset serves as a valuable resource for studies on mycology, fungal taxonomy, and the development of intelligent systems for mushroom identification and safety assessment. Importantly, the Mushroom dataset is balanced, ensuring an equitable representation of various mushroom species for classification tasks.
Four of the datasets, including the fingerprint dataset, are perfectly balanced. While our proposal primarily targets balanced databases, we have also included two imbalanced databases to assess the general performance of our approach across varying class distributions. This decision allows for a more comprehensive evaluation of our method’s robustness and effectiveness in handling different dataset characteristics. Table 1 presents an overview of the fundamental characteristics of fingerprints, with detailed information about the additional datasets used in this study.
Hyperparameter estimation
We conducted an extensive hyperparameter estimation process to optimize the accuracy of OS-RVFL. The hyperparameters under consideration were the number of neurons in the hidden layer and the regularization parameter, denoted as C. For this estimation, we used a fingerprint dataset containing 60,000 samples. The dataset was further divided into three subsets: 60% for training (36,000 samples), 20% for validation (12,000 samples), and 20% for testing (12,000 samples).
With regard to the hidden-layer neurons, we performed training experiments using 500–5000 neurons, in increments of 500 neurons in each experiment. Further, we explored a wide range of the C parameter, from \(10^{-20}\) to \(10^{20}\), incrementing the exponent by 1. In each phase of OS-RVFL, we used 9,000 samples for both the initial phase and each subsequent chunk in the sequential learning phase. Fig. 5 shows the results obtained for the different combinations of hyperparameters. The accuracy improved remarkably when the regularization parameter C was in the range of \(10^{-10}\)–\(10^{10}\). Based on these results, we selected 2000 neurons in the hidden layer and set \(C = 10\) as the regularization parameter for the experiments conducted in this work.
Results of hyperparameter tuning in an OS-RVFL neural network11. The accuracy improves when the regularization parameter is between \(10^{-10}\) and \(10^{10}\), while there is no significant change when the number of neurons in the hidden layer are increased (36,000 samples for training, 12,000 for validation, 12,000 for testing, and sigmoid activation function).
For the public datasets, we used the entire dataset to estimate the hyperparameters. We conducted experiments ranging from 100 neurons to 3,000 neurons in the hidden layer, with increments of 100 neurons in each experiment. Notably, each dataset achieved the highest accuracy with a different number of neurons in the hidden layer (see Table 1). By following the approach of previous authors8,20, we used \(L+100\) samples for the initial phase and each subsequent chunk of the sequential learning phase to ensure effective training.
Result using fingerprint dataset
We conducted a comparative analysis between the results obtained using OS-RVFL and our distributed model using the fingerprint dataset described in Table 1. For comparison, we evaluated the training time, training accuracy, and testing accuracy as performance metrics (see Table 2). Additionally, we analyzed the speed-up and efficiency of our algorithm using the distributed training data. The speed-up was calculated as follows:
where S(p) is the speed-up with p threads, and Time(1) is the training time with 1 thread and Time(p) with p threads. Next, we derived the corresponding efficiency \(E(p)=S(p)/p\). In our experiments, we used the sigmoid activation function and set the number of neurons in the hidden layer to 2000, with \(C=10\) as the regularization parameter. These configurations were selected based on the results of the previous hyperparameter estimation.
The results in Table 2 demonstrate that the training accuracy of EOS-RVFL-FC-D is comparable to that of the standard OS-RVFL network. However, the testing accuracy improves with an increase in the number of threads. The highest testing accuracy achieved by EOS-RVFL-FC-D was 93.05% when using 10 threads. By applying the frequency criterion, the accuracy was further enhanced, while the training time decreased drastically. Furthermore, the training time of the EOS-RVFL-FC-D decreased as the number of threads increased as the workload was distributed across the available threads.
Meanwhile, Fig. 6 shows the speed-up and efficiency of our parallel algorithm in comparison to the sequential version and provides a comprehensive overview of the results. Fig. 6 depicts the noteworthy observations obtained for EOS-RVFL-FC-D. The speed-up exceeds the number of threads (as shown in Fig. 6a), and the efficiency surpasses one (as depicted in Fig. 6b). This phenomenon is attributed to the distribution of samples as the threads increase, allowing each thread to be trained on a smaller subset of samples. Additionally, our experimental results reveal that the computation of \(\varvec{\beta }_0\) is approximately 8.8 times faster than in the case of the sequential learning phase \(\varvec{\beta }_{k+1}\). A graphical representation illustrating this behavior is shown in Fig. 7. The observed behavior can be effectively modeled using the following equation:
where \(Time_0\) is the computation time of \(\varvec{\beta }_0\) with the \(chunk_1\); \(\rho\) is the difference in computation time between \(\varvec{\beta }^0\) and \(\varvec{\beta }^{k+1}\); Q represents the chunks for training; and p indicates the number of threads used for parallel training with \(p \in {\mathbb {N}}\) and \(p = 1, \ldots , Q\).
(a) Speed-up and (b) efficiency of our model with regard to the training time when using a distributed (EOS-RVFL-FC-D) dataset (200,000 samples for training; sigmoid function activation; 2000 neurons in the hidden layer; and \(C = 10\) as the regularization parameter). The speed-up was calculated over the standard OS-RVFL network using a single thread.
Graphical representation showing the distribution of the dataset and the training time in relation to the number of threads utilized for parallel training. In this graph, Q represents the number of chunks used for training; p denotes the number of threads; and \(\rho\) represents the difference in computation time between \(\varvec{\beta }^0\) and \(\varvec{\beta }_{k+1}\).
Results using a large-scale dataset
We conducted experiments with a large-scale fingerprint database. The database consists of 1,000,000 training samples, 10,000 testing samples, and the same characteristics as those listed in Table 1 (classes and attributes). These experiments used 6–46 threads with a power-of-two growth. Table 3 lists the results related to training time, training accuracy, and testing accuracy. Additionally, we compared the speed-up and efficiency of the proposed network to those of the standard OS-RVFL neural network. The results demonstrate that the testing accuracy increases as the number of threads increases, reaching up to 93.19% for 48 threads. Although the improvement in accuracy is not significant, the converse is true for the training time. As shown in Table 3, the training time decreases significantly when 48 threads are used.
Furthermore, we compared the experimental results and those obtained using (11). The parameters employed in (11) were computed based on preceding experiments summarized in Table 2 and Fig. 6. The computation time “\(Time_0\)”, associated with “\(Chunk_1\),” was measured to be 494 seconds, while the time difference \(\rho\) between \(Time_0\) and the other Time values was found to be \(8.8\times\). Here, Q represents the ratio of N to \(N_0\), where N corresponds to the total samples present in the dataset, and \(N_0\) represents the number of samples allocated to each training Chunk. Further, p denotes the number of threads employed during the parallel training. Fig. 8 shows a comparison of the speed-up and efficiency between the experimental results and the results using (11).
(a) Speed-up and (b) efficiency of our proposal for training time using distributed (EOS-RVFL-FC-D) dataset (1,000,000 samples training, sigmoid function, 2000 neurons in the hidden layer, and \(C = 10\) as the regularization parameter). The speed-up was calculated over the OS-RVFL using a single thread.
The speed-up (Fig. 8a) and efficiency (Fig. 8b) achieved by our distributed proposal surpass those of a conventional parallel algorithm, considering the characteristics of sequential training depicted in Fig. 7. These results highlight the effectiveness of our proposal in terms of training time. Furthermore, the outcomes obtained using (11) agree closely with our experimental results. By using the parameters specified in (11), we can estimate the behavior of our distributed proposal when using multiple threads for parallel training. Notably, the accuracy of (11) can be further enhanced by incorporating the behaviors of other factors, such as hardware considerations.
Results using the replicated dataset
The experimental results we obtained so far can be achieved when working with databases that have sufficient samples to distribute across multiple threads. However, when the database has a limited number of training samples, the frequency criterion proposed herein can be applied by replicating the training samples in each thread (EOS-RVFL-FC-R). To test this approach with replicated data, we adopted widely used public databases to evaluate artificial intelligence algorithms, particularly neural networks for classification problems (see Table 1). It is important to highlight that each OS-RVFL network trains with randomly assigned weights, ensuring that each network is independent. This property guarantees that the frequency criterion remains effective even when training with a replicated database. As a result, the diversity among the independently trained networks contributes to robustness in the final classification decision, as it accounts for different perspectives captured by each network. Table 4 Table 4 lists the results of training time, training accuracy, and testing accuracy, obtained with our proposed method with replicated data. In these experiments, we used 10 threads, while the number of neurons in the hidden layer was obtained from the literature.
With regard to the training time, we can see from Table 4 that more time is required when the database has a larger number of training samples. However, when the number of samples is very small, the difference in training time is insignificant, as seen in the case of the Mushroom and Image Segmentation datasets. Meanwhile, the accuracy improves in all databases, particularly in the MNIST and Image Segmentation datasets. In the other databases, the improvement in accuracy is minimal. Overall, in applications with small databases, our proposal can enhance accuracy without drastically affecting the training time. However, when the number of samples is larger, the increase in accuracy does not justify the increase in training time. In these cases, it is preferable to evaluate the distribution of samples across multiple threads.
To compare our model with other ensemble OS-RVFL models reported in the literature, we compare the reported results with ours. The results presented by Lan et al.20, Liu et al.22, and Wei et al.23 show that their ensemble models increase accuracy in a similar range to ours. However, the training time significantly increases compared to standard OS-RVFL as the number of networks increases. Results presented by Huang et al.28 show a considerable increase in training time, even though their approach is parallel and based on MapReduce. Among the databases used by the authors are Mnist, Image Segmentation, Adult, and Satellite Image, making their results comparable to ours. Therefore, our proposal offers significant advantages over these models, especially concerning training time.
Conclusions
In this paper, we introduce a frequency criterion in a parallel ensemble algorithm for sequential online RVFL network (OS-RVFL-FC) in large-scale classification problems. We validated our proposed network using a synthetic fingerprint database and five widely used public databases. The parallel ensemble approach involves training multiple OS-RVFL networks by distributing or replicating the database samples and then applying a frequency criterion to the outputs of all the neural networks. The frequency criterion selects the most frequent output among the results obtained from all OS-RVFL networks. We used two methodologies: (1) distributed samples (EOS-RVFL-FC-D) for large-scale databases, and (2) replicated samples (EOS-RVFL-FC-R) for small-scale databases.
The results with EOS-RVFL-FC-D demonstrate that the accuracy increases when trained with multiple threads, while the training time significantly decreases; the achieved speed-up and efficiency exceed those of a conventional parallel program. This improvement originates from the substantial difference between the execution time of the initial phase and the sequential learning with each chunk. Considering this fact, we introduced an equation that can reasonably predict the speed-up and efficiency of our proposal based on the execution time in the initial phase, its relationship with the sequential learning in each chunk, the total number of training samples, and the size of the chunks in both phases.
With regard to EOS-RVFL-FC-R, the results demonstrate that the accuracy increases for all databases, though the improvement is drastically small in some cases. The difference in training time is negligible when the databases have few samples. However, this difference becomes significantly large as the size of the database increases, making the method impractical for large-scale databases. In general, the proposed model with distributed data is suitable for large-scale databases as it significantly reduces the training time as the number of threads increases. On the other hand, for small databases, the proposed model with replicated data can improve the overall accuracy of the neural network. However, when the number of samples is larger, it is more viable to consider the proposed model with distributed data.
In future work, we will continue investigating ensemble methods in randomization-based online sequential neural networks to further improve the accuracy and training time. We plan to include more datasets with imbalanced class distributions to evaluate the effectiveness of future proposals under such conditions. Additionally, we believe it is important to implement this proposed methodology in real-world applications, considering the substantial reduction in training time. Furthermore, we will continue to work on proposals that incorporate statistical improvements in the frequency criterion.
Data availability
The public datasets used and/or analyzed during the current study are available in the UCI Machine Learning Repository, [https://archive.ics.uci.edu/datasets]. The fingerprint datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.
References
Zhang, L. & Suganthan, P. N. A survey of randomized algorithms for training neural networks. Inf. Sci. 364–365, 146–155. https://doi.org/10.1016/j.ins.2016.01.039 (2016).
Suganthan, P. N. & Katuwal, R. On the origins of randomization-based feedforward neural networks. Appl. Soft Comput. 105, 107239. https://doi.org/10.1016/j.asoc.2021.107239 (2021).
Pao, Y.-H. & Takefuji, Y. Functional-link net computing: Theory, system architecture, and functionalities. Computer 25, 76–79. https://doi.org/10.1109/2.144401 (1992).
Pao, Y.-H., Park, G.-H. & Sobajic, D. J. Learning and generalization characteristics of the random vector functional-link net. Neurocomputing 6, 163–180. https://doi.org/10.1016/0925-2312(94)90053-1 (1994).
Zhang, L. & Suganthan, P. N. A comprehensive evaluation of random vector functional link networks. Inf. Sci. 367, 1094–1105. https://doi.org/10.1016/j.ins.2015.09.025 (2016).
Gelvez-Almeida, E. et al. Parallel methods for linear systems solution in extreme learning machines: An overview. J. Phys. Conf. Ser. 1702, 012017. https://doi.org/10.1088/1742-6596/1702/1/012017 (2020).
Malik, A. K., Gao, R., Ganaie, M., Tanveer, M. & Suganthan, P. N. Random vector functional link network: Recent developments, applications, and future directions. Appl. Soft Comput. 143, 110377. https://doi.org/10.1016/j.asoc.2023.110377 (2022).
Liang, N.-Y., Huang, G.-B., Saratchandran, P. & Sundararajan, N. A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans. Neural Netw. 17, 1411–1423. https://doi.org/10.1109/TNN.2006.880583 (2006).
Matias, T., Souza, F., Araújo, R., Gonçalves, N. & Barreto, J. P. On-line sequential extreme learning machine based on recursive partial least squares. J. Process Control 27, 15–21. https://doi.org/10.1016/j.jprocont.2015.01.004 (2015).
Mirza, B., Lin, Z., Cao, J. & Lai, X. Voting based weighted online sequential extreme learning machine for imbalance multi-class classification. In 2015 IEEE International Symposium on Circuits and Systems (ISCAS) 565–568. https://doi.org/10.1109/ISCAS.2015.7168696 (IEEE, 2015).
Gelvez-Almeida, E., Barrientos, R. J., Vilches-Ponce, K. & Mora, M. Parallel training of a set of online sequential extreme learning machines. In: 41st International Conference of the Chilean Computer Science Society (SCCC), Santiago. Chile 1–4, 2022. https://doi.org/10.1109/SCCC57464.2022.10000361 (21-25 November 2022).
Gelvez-Almeida, E., Barrientos, R. J., Vilches-Ponce, K. & Mora, M. Parallel model of online sequential extreme learning machines for classification problems with large-scale databases. In XI Jornadas de Cloud Computing, Big Data & Emerging Topics, La Plata, Argentina 19–23. http://sedici.unlp.edu.ar/handle/10915/155423 (27–29 June 2023).
Wibawa, I. P. D., Machbub, C., Rohman, A. S. & Hidayat, E. Modified online sequential extreme learning machine algorithm using model predictive control approach. Intell. Syst. Appl. 18, 200191. https://doi.org/10.1016/j.iswa.2023.200191 (2023).
Chen, Y. & Li, M. An effective online sequential stochastic configuration algorithm for neural networks. Sustainability 14, 15601. https://doi.org/10.3390/su142315601 (2022).
Zhang, X., Ma, H., Zuo, W. & Luo, M. Adaptive control of discrete-time nonlinear systems using ITF-ORVFL. IEEE/CAA J. Autom. Sin. 9, 556–563. https://doi.org/10.1109/JAS.2019.1911801 (2022).
Zha, L., Ma, K., Li, G., Yang, J. & Fang, Q. An improved extreme learning machine with self-recurrent hidden layer. Adv. Eng. Inform. 54, 101736. https://doi.org/10.1016/j.aei.2022.101736 (2022).
Kale, A. P., Sonawane, S., Wahul, R. M. & Dudhedia, M. A. Improved genetic optimized feature selection for online sequential extreme learning machine. Ingénierie des Systèmes d’Informationhttps://doi.org/10.18280/isi.270519 (2022).
Polat, Ö. & Kayhan, S. K. GPU-accelerated and mixed norm regularized online extreme learning machine. Concurr. Comput. Pract. Exp. 34, e6967. https://doi.org/10.1002/cpe.6967 (2022).
Ren, Y., Zhang, L. & Suganthan, P. N. Ensemble classification and regression-recent developments, applications and future directions. IEEE Comput. Intell. Mag. 11, 41–53. https://doi.org/10.1109/MCI.2015.2471235 (2016).
Lan, Y., Soh, Y. C. & Huang, G.-B. Ensemble of online sequential extreme learning machine. Neurocomputing 72, 3391–3395. https://doi.org/10.1016/j.neucom.2009.02.013 (2009).
Liu, N. & Wang, H. Ensemble based extreme learning machine. IEEE Signal Process. Lett. 17, 754–757. https://doi.org/10.1109/LSP.2010.2053356 (2010).
Liu, Y. et al. Particle swarm optimization based selective ensemble of online sequential extreme learning machine. Math. Problems Eng.https://doi.org/10.1155/2015/504120 (2015).
Wei, L., Wang, L., Li, Y. & Duan, S. Ensemble of online sequential extreme learning machine based on cross-validation. J. Phys. Conf. Ser. 1550, 032156. https://doi.org/10.1088/1742-6596/1550/3/032156 (2020).
Zhai, J., Wang, J. & Wang, X. Ensemble online sequential extreme learning machine for large data set classification. In 2014 IEEE International Conference on Systems, Man, and Cybernetics (SMC) 2250–2255. https://doi.org/10.1109/SMC.2014.6974260 (IEEE, 2014).
Alhamdoosh, M. & Wang, D. Fast decorrelated neural network ensembles with random weights. Inf. Sci. 264, 104–117. https://doi.org/10.1016/j.ins.2013.12.016 (2014).
Mirza, B., Lin, Z. & Liu, N. Ensemble of subset online sequential extreme learning machine for class imbalance and concept drift. Neurocomputing 149, 316–329. https://doi.org/10.1016/j.neucom.2014.03.075 (2015).
Ling, Q.-H., Song, Y.-Q., Han, F., Yang, D. & Huang, D.-S. An improved ensemble of random vector functional link networks based on particle swarm optimization with double optimization strategy. Plos ONE 11, e0165803. https://doi.org/10.1371/journal.pone.0165803 (2016).
Huang, S. et al. Parallel ensemble of online sequential extreme learning machine based on MapReduce. Neurocomputing 174, 352–367. https://doi.org/10.1016/j.neucom.2015.04.105 (2016).
Rakesh, K. & Suganthan, P. N. An ensemble of kernel ridge regression for multi-class classification. Procedia Comput. Sci. 108, 375–383. https://doi.org/10.1016/j.procs.2017.05.109 (2017).
Zhang, L. & Suganthan, P. N. Benchmarking ensemble classifiers with novel co-trained kernel ridge regression and random vector functional link ensembles [research frontier]. IEEE Comput. Intell. Mag. 12, 61–72 (2017).
Li, Y., Zhang, S., Yin, Y., Xiao, W. & Zhang, J. Parallel one-class extreme learning machine for imbalance learning based on Bayesian approach. J. Ambient Intell. Humaniz. Comput.https://doi.org/10.1007/s12652-018-0994-x (2018).
Katuwal, R. & Suganthan, P. N. Dropout and dropconnect based ensemble of random vector functional link neural network. In 2018 IEEE Symposium Series on Computational Intelligence (SSCI) 1772–1778. https://doi.org/10.1109/SSCI.2018.8628640 (2018).
Liu, Y. et al. Ensemble neural networks with random weights for classification problems. In 2020 3rd International Conference on Algorithms, Computing and Artificial Intelligence 1–5. https://doi.org/10.1145/3446132.3446147 (2020).
Hu, M., Shi, Q., Suganthan, P. N. & Tanveer, M. Adaptive ensemble variants of random vector functional link networks. In Neural Information Processing: 27th International Conference, ICONIP 2020, Bangkok, Thailand, November 18–22, 2020, Proceedings, Part V 27 30–37. https://doi.org/10.1007/978-3-030-63823-8_4 (2020).
Malik, A. K., Ganaie, M., Tanveer, M. & Suganthan, P. N. A novel ensemble method of RVFL for classification problem. In 2021 International Joint Conference on Neural Networks (IJCNN) 1–8. https://doi.org/10.1109/IJCNN52387.2021.9533836 (2021).
Tanveer, M., Ganaie, M. & Suganthan, P. N. Ensemble of classification models with weighted functional link network. Appl. Soft Comput. 107, 107322. https://doi.org/10.1016/j.asoc.2021.107322 (2021).
Shi, Q., Katuwal, R., Suganthan, P. N. & Tanveer, M. Random vector functional link neural network based ensemble deep learning. Pattern Recognit. 117, 107978. https://doi.org/10.1016/j.patcog.2021.107978 (2021).
Dong, Q., Zhou, Y., Lian, J. & Li, L. Online adaptive humidity monitoring method for proton exchange membrane fuel cell based on fuzzy c-means clustering and online sequence extreme learning machine. Electrochim. Acta 429, 141059. https://doi.org/10.1016/j.electacta.2022.141059 (2022).
Yao, Y., Ding, J., Zhao, C., Wang, Y. & Chai, T. Data-driven constrained reinforcement learning for optimal control of a multistage evaporation process. Control Eng. Pract. 129, 105345. https://doi.org/10.1016/j.conengprac.2022.105345 (2022).
Li, R., Gao, R., Cui, J., Suganthan, P. & Sourina, O. Advanced ensemble deep random vector functional link for eye-tracking-based situation awareness recognition. In 2022 IEEE Symposium Series on Computational Intelligence (SSCI) 300–307. https://doi.org/10.1109/SSCI51031.2022.10022019 (2022).
Ganaie, M. & Tanveer, M. Ensemble deep random vector functional link network using privileged information for Alzheimer’s disease diagnosis. IEEE/ACM Trans. Comput. Biol. Bioinform.https://doi.org/10.1109/TCBB.2022.3170351 (2022).
Gao, R., Du, L., Suganthan, P. N., Zhou, Q. & Yuen, K. F. Random vector functional link neural network based ensemble deep learning for short-term load forecasting. Expert Syst. Appl. 206, 117784. https://doi.org/10.1016/j.eswa.2022.117784 (2022).
Yu, X. et al. Selective ensemble deep bidirectional RVFLN for landslide displacement prediction. Nat. Hazards 112, 725–745. https://doi.org/10.1007/s11069-021-05202-w (2022).
Jamei, M. et al. Development of wavelet-based kalman online sequential extreme learning machine optimized with boruta-random forest for drought index forecasting. Eng. Appl. Artif. Intell. 117, 105545. https://doi.org/10.1016/j.engappai.2022.105545 (2023).
Zhou, X., Huang, J., Lu, F., Zhou, W. & Liu, P. A novel compound fault-tolerant method based on online sequential extreme learning machine with cycle reservoir for turbofan engine direct thrust control. Aerosp. Sci. Technol. 132, 108059. https://doi.org/10.1016/j.ast.2022.108059 (2023).
Rajpal, A., Sehra, K., Mishra, A. & Chetty, G. A low-resolution real-time face recognition using extreme learning machine and its variants. Imaging Sci. J.https://doi.org/10.1080/13682199.2023.2183544 (2023).
Xu, Q., Wei, X., Bai, R., Li, S. & Meng, Z. Integration of deep adaptation transfer learning and online sequential extreme learning machine for cross-person and cross-position activity recognition. Expert Syst. Appl. 212, 118807. https://doi.org/10.1016/j.eswa.2022.118807 (2023).
Wang, F., Song, Y., Liu, C., He, A. & Qiang, Y. Multi-objective optimal scheduling of laminar cooling water supply system for hot rolling mills driven by digital twin for energy-saving. J. Process Control 122, 134–146. https://doi.org/10.1016/j.jprocont.2023.01.004 (2023).
Yu, L., Bai, Y. & Li, K. Lane-changing control of vehicle platoon based on OS-ELM environmental parameter identifier. IEEE Trans. Veh. Technol.https://doi.org/10.1109/TVT.2022.3214935 (2023).
Kouzehkalani Sales, A., Gul, E. & Safari, M. J. S. Online sequential, outlier robust, and parallel layer perceptron extreme learning machine models for sediment transport in sewer pipes. Environ. Sci. Pollut. Res. 30, 39637–39652. https://doi.org/10.1007/s11356-022-24989-0 (2023).
Samal, S. & Dash, R. Developing a novel stock index trend predictor model by integrating multiple criteria decision-making with an optimized online sequential extreme learning machine. Granul. Comput. 8, 411–440. https://doi.org/10.1007/s41066-022-00338-x (2023).
Duan, W. et al. Battery SOH estimation and RUL prediction framework based on variable forgetting factor online sequential extreme learning machine and particle filter. J. Energy Storage 65, 107322. https://doi.org/10.1016/j.est.2023.107322 (2023).
Shi, Q., Hu, M., Suganthan, P. N. & Katuwal, R. Weighting and pruning based ensemble deep random vector functional link network for tabular data classification. Pattern Recognit. 132, 108879. https://doi.org/10.1016/j.patcog.2022.108879 (2022).
Li, R. et al. A spectral-ensemble deep random vector functional link network for passive brain-computer interface. Expert Syst. Appl. 227, 120279. https://doi.org/10.1016/j.eswa.2023.120279 (2023).
Li, R. et al. An enhanced ensemble deep random vector functional link network for driver fatigue recognition. Eng. Appl. Artif. Intell. 123, 106237. https://doi.org/10.1016/j.engappai.2023.106237 (2023).
Xiao, S., Zhang, F. & Huang, X. Online thickness prediction of hot-rolled strip based on ISSA-OSELM. Int. J. Interact. Design Manuf. (IJIDeM) 16, 1089–1098. https://doi.org/10.1007/s12008-021-00833-6 (2022).
Li, Y., Zhang, J., Zhang, S. & Xiao, W. Dual ensemble online modeling for dynamic estimation of hot metal silicon content in blast furnace system. ISA Trans. 128, 686–697. https://doi.org/10.1016/j.isatra.2021.09.018 (2022).
Ren, Q., Li, M., Kong, T. & Ma, J. Multi-sensor real-time monitoring of dam behavior using self-adaptive online sequential learning. Autom. Constr. 140, 104365. https://doi.org/10.1016/j.autcon.2022.104365 (2022).
Zhou, Z., Ji, H. & Zhu, Z. Online sequential fuzzy dropout extreme learning machine compensate for sliding-mode control system errors of uncertain robot manipulator. Int. J. Mach. Learn. Cybern. 13, 2171–2187. https://doi.org/10.1007/s13042-022-01513-x (2022).
Ali, M. et al. Coupled online sequential extreme learning machine model with ant colony optimization algorithm for wheat yield prediction. Sci. Rep. 12, 5488. https://doi.org/10.1038/s41598-022-09482-5 (2022).
He, J. et al. Ensemble deep random vector functional link for self-supervised direction-of-arrival estimation. Eng. Appl. Artif. Intell. 120, 105831. https://doi.org/10.1016/j.engappai.2023.105831 (2023).
Gao, R., Li, R., Hu, M., Suganthan, P. N. & Yuen, K. F. Significant wave height forecasting using hybrid ensemble deep randomized networks with neurons pruning. Eng. Appl. Artif. Intell. 117, 105535. https://doi.org/10.1016/j.engappai.2022.105535 (2023).
Huang, G.-B., Zhu, Q.-Y. & Siew, C.-K. Extreme learning machine: A new learning scheme of feedforward neural networks. In 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, vol 2, 985–990. https://doi.org/10.1109/IJCNN.2004.1380068 (25–29 July 2004).
Safaei, A., Wu, Q. M. J., Akilan, T. & Yang, Y. System-on-a-chip (SoC)-based hardware acceleration for an online sequential extreme learning machine (OS-ELM). IEEE Trans. Comput.-Aided Design Integr. Circuits Syst. 38, 2127–2138. https://doi.org/10.1109/TCAD.2018.2878162 (2018).
Rosato, A., Altilio, R. & Panella, M. On-line learning of RVFL neural networks on finite precision hardware. In 2018 IEEE International Symposium on Circuits and Systems (ISCAS) 1–5. https://doi.org/10.1109/ISCAS.2018.8351399 (IEEE, 2018).
Decherchi, S., Gastaldo, P., Leoncini, A. & Zunino, R. Efficient digital implementation of extreme learning machines for classification. IEEE Trans. Circuits Syst. II: Express Briefs 59, 496–500. https://doi.org/10.1109/TCSII.2012.2204112 (2012).
Xia, M., Wang, J., Liu, J., Weng, L. & Xu, Y. Density-based semi-supervised online sequential extreme learning machine. Neural Comput. Appl. 32, 7747–7758. https://doi.org/10.1007/s00521-019-04066-3 (2020).
Hong, J.-H., Min, J.-K., Cho, U.-K. & Cho, S.-B. Fingerprint classification using one-vs-all support vector machines dynamically ordered with naï ve bayes classifiers. Pattern Recognition 41, 662–671. https://doi.org/10.1016/j.patcog.2007.07.004 (2008).
Henry, E. R. Classification and uses of finger prints (His Magjesty’s Stationery Office, London, 1922).
Zabala-Blanco, D., Mora, M., Barrientos, R. J., Hernández-García, R. & Naranjo-Torres, J. Fingerprint classification through standard and weighted extreme learning machines. Appl. Sci. 10, 4125. https://doi.org/10.3390/app10124125 (2020).
Zabala-Blanco, D., Mora, M., Hernández-García, R. & Barrientos, R. J. The extreme learning machine algorithm for classifying fingerprints. In 2020 39th International Conference of the Chilean Computer Science Society (SCCC) 1–8. https://doi.org/10.1109/SCCC51225.2020.9281232 (IEEE, 2020).
Deng, L. The MNIST database of handwritten digit images for machine learning research [best of the web]. IEEE Signal Process. Mag. 29, 141–142. https://doi.org/10.1109/MSP.2012.2211477 (2012).
Dua, D. & Graff, C. UCI machine learning repository. University of California, Irvine, School of Information and Computer Sciences. http://archive.ics.uci.edu/ml (2017).
Acknowledgements
This work was funded by the National Agency for Research and Development (ANID)/Scholarship Program/BECAS DOCTORADO NACIONAL/2020-21201000. The paper’s authors also thank the Research Project ANID FONDECYT REGULAR 2020 No. 1200810 “Very Large Fingerprint Classification Based on a Fast and Distributed Extreme Learning Machine,” Government of Chile. Elkin Gelvez-Almeida is also appreciative of the licenses for doctoral studies for the “Fund for Teacher and Professional Development” of the Universidad Simón Bolívar, Colombia.
Author information
Authors and Affiliations
Contributions
All authors contributed equally to this study. E.G.-A. Conceptualization, Formal analysis, Funding acquisition, Investigation, Writing - original draft. R.J.B. Conceptualization, Methodology, Project administration, Supervision, Validation, Visualization, Writing - review & editing. K.V.-P. Methodology, Visualization, Writing - review & editing. M.M. Conceptualization, Funding acquisition, Methodology, Project administration, Supervision, Validation, Writing - review & editing.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Gelvez-Almeida, E., Barrientos, R.J., Vilches-Ponce, K. et al. Parallel ensemble of a randomization-based online sequential neural network for classification problems using a frequency criterion. Sci Rep 14, 16104 (2024). https://doi.org/10.1038/s41598-024-66676-9
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-024-66676-9











