Mobile QR Code QR CODE

2024

Acceptance Ratio

21%


  1. ( Baode Coal Mine of Guoneng Shendong Coal Group Co., Ltd., Xinzhou 036600, China 200516666@ceic.com (Peng Gao), 10036711@ceic.com (Jinlin Ruan), 10020983@ceic.com (Yuan Sun))
  2. ( Research Institute of Mine Big Data, Chinese Institute of Coal Science, Beijing 100013, China hateif@163.com (Hao Li))
  3. ( Beijing Technology Research Branch, Tiandi Science & Technology Co., Ltd., Beijing 100013, China sangcongCCRI@163.com (Cong Sang))



Machine learning, Prediction model, Support vector machine, Belt transportation system, Neural network

1. Introduction

Given the widespread application of interpretable machine learning algorithms, the prediction of system performance has emerged as a prominent research area. Belt conveyor system (BCS) is a commonly used material conveying system that plays a wide role in modern industry. The design quality directly affects production efficiency and product quality [1]. High performance prediction is functional to strengthen the transportation capacity and efficiency of BCS systems. The commonly used performance prediction algorithm is the neural network (NN) algorithm of machine learning. This study proposes a comprehensive NN optimization method by integrating the initial weight threshold (IWT) optimization of NNs and the partition optimization of training and testing sets. Moreover, the support vector machine (SVM) algorithm is improved through the fish swarm algorithm (FSA) [2], which improves the interpretability and optimization ability of the algorithm and establishes a machine learning model. The purpose is to achieve high-performance prediction of BCS. The structure of the paper is as follows. Firstly, it elaborates on the research significance and background of predicting high-performance BCS, as well as the purpose of optimizing machine learning algorithms. The latter part precisely elucidates the underlying principles and research methodologies of two approaches for optimizing NNs, alongside delineating the process and innovative aspects of optimizing support vector machines. The third part fully elaborates on the application of the designed algorithm in the experimental design of predictive performance, as well as the quantitative statistics and analysis of experimental results. The fourth part describes the experimental conclusions, as well as the shortcomings of this research design and the directions for further in-depth research.

2. Related Works

A NN is a mathematical model composed of a large number of simple processing units connected, which adjusts the inter-connectivity between them to achieve the purpose of processing information. It has the ability to self-learning and adaptive. Belt conveyor is a commonly used conveying equipment, mainly used for conveying various materials, such as coal, ore, sand, fertilizer, grain, etc. The study of its performance is one of the key points in improving transportation efficiency, and many scholars have put forward their own suggestions for this. Choi and his team proposed an unsupervised intelligent system for predicting the operation status of truck transportation systems during open-pit ore transportation. This system combined Harris Hawks Optimization (HHO) and SVM to obtain the HHO-SVM model. They developed Random Forest (RF) and BP Neural Network (BPNN) models and compared together, indicating that BPNN, RF, raw SVM, and HHO-SVM models were potential intelligent for predicting ore production [3]. Belt deviation (BD) is the common fault of belt conveyors, which is significant for ensuring the transportation system runs safely and efficiently. In response to the shortcomings of existing technologies in detection speed, Zhang's team proposed a new conveyor BD monitoring method grounded on deep learning. This method was implemented by improving the output results of the universal object detection network, thereby enhancing the network's ability to detect straight lines rather than bounding boxes. It exactly solved the fast feature extraction and deviation judgment on the edge of conveyor belts in complex backgrounds, balanced detection accuracy and speed, and showed good real-time performance [4]. Bhaskaran et al. adopted a laboratory scale fluid transport system based on a distributed control system, where flow rate and pressure are the most influential parameters during pipeline risk occurrence. The results showed that the system improved monitoring performance, and the monitoring and data collection platform combined with the Internet of Things by utilizing the Narrow Band Internet of Things (NB IoT) module with advanced engineering interfaces [5]. Daoudr et al. proposed to derive a clustering algorithm based on weighted algorithm for head allocation process, emphasizing parallel processing techniques based on various wavelet baby functions. His research had achieved performance improvements in data transmission and processing speed [6].

Belt conveyors are widely used for short and long distance material transportation, and the failure of individual components can lead to fatal consequences. Li et al. used machine learning for timely fault diagnosis to ensure the safe running of belt conveyors and proposed a grey wolf optimization method. The application of this diagnostic model to the fault diagnosis of underground BCS showed that the combined classification model had good performance in intelligent fault diagnosis [7]. In the optimization of belt mechanics, Bae and Kim introduced the structural design and mechanical performance evaluation results of a lightweight belt used in high-rise elevators. Structural design and performance evaluation of its ropes: Compared with traditional steel wire ropes, the belt had sufficient performance [8]. Liu and his companions used acoustic signal based methods to detect faults in the conveyor rollers, extracting Mel frequency cepstrum coefficients as features from the obtained sound signals. A gradient enhancement decision tree model was established and trained to validate the proposed fault detection method [9]. Li proposed a novel NN architecture with attention mechanism for developing building energy prediction based on Recurrent Neural Network (RNN). This study was effective in improving the interpretability of RNN models for building cooling load prediction [10].

In summary, scholars have studied faults, structures, and algorithms in the performance research of belt transportation systems. However, there is still little research on the use of neural networks for interpretation and performance prediction, mainly manifested in the improvement and enhancement of one or several of their performance, such as predictive performance, explanatory performance, data transmission, processing speed, etc., without comprehensive research. Therefore, in order to predict the problem of high-performance belt transportation systems, neural networks and vector machine algorithms in interpretable machine learning are adopted. Corresponding improvements are first made, and optimization is carried out based on the performance of the transportation system to improve its performance.

3. Predictive High-Performance BCS Applications based on Interpretive Machine Learning Models

Machine learning is the process of programming a computer to solve a given problem using sample data or past experience. NNs, as two excellent methods in the field of machine learning, have been widely applied in research topics of performance prediction in recent years. Moreover, the optimization problem of this method in the application process has become a constantly developing new topic.

3.1 An Interpretable Machine Learning Model Based on Improved NNs

Although NNs have good utility in processing sample information of BCS, there is still some optimization space. In practical applications, due to different network structures, the prediction accuracy of NNs is also different, and even over-fitting problems occur. That is, due to unreasonable initial state selection and excessive training of the network, the weight threshold falls into the local minimum value [11]. Therefore, to optimize the selection of IWT for NNs, this study proposes a method based on genetic algorithm (GA) to divide training set and testing set (TrS & TeS). A comprehensive new method for optimizing the IWT of the network and optimizing dataset partitioning is proposed by integrating the optimization methods of dividing the TrS & TeS designed in the research.

Building a NN prediction model has several steps: input and output parameter selection, normalization processing, selection of network structure, and network training. In the selection of input and output parameters, based on the parameters and performance of BCS, the operating conditions and environmental parameters of BCS can be ignored when establishing a prediction model. The input parameters for selecting the model include speed, operating mode, conveying capacity, conveying distance, etc. [12,13].

NNs are a parallel processing system that achieves interpretability of parameters through normalization processing. This study selected $[-1$, $1]$ normalization in machine learning to normalize the parameters, and the expression function is Eq. (1).

(1)
$ \bar{x}=-1+\frac{2\left(x-x{}_{\min } \right)}{x_{\max } -x_{\min } }. $

In Eq. (1), $x_{\max } $ and $x_{\min } $ are the max and min values of the sample. $x$ and $\bar{x}$ represent vectors before and after normalization. The selection of NNs directly affects their predictive performance. In the study of BCS performance, the speed, operation mode, conveying capacity, conveying distance, environmental temperature and other influencing factors of belt conveyors vary. Therefore, before using a belt conveyor, it is necessary to have a detailed understanding of the performance parameters of the equipment in order to ensure its normal operation and safe use.

The three-layer feedforward network structure of NNs includes input, output, and hidden layers, demonstrating good performance. This three-layer network structure is selected as a prediction study for the performance of BCS [14]. Based on the machine performance and emission data that can be collected by the belt transportation system equipment, the input parameters of the neural network prediction model are ultimately determined to be: speed, power, rail pressure, and fuel injection timing; The output parameters to be predicted include: fuel consumption rate, maximum burst pressure, burst pressure angle, turbocharger speed, front vortex temperature, and smoke level. Due to the fact that the neurons in hidden layer (HLN) are an essential structural parameter affecting network performance, analysis and research were conducted on distinctive quantity of neurons. The input layer is equipped with $m$ input nodes, and any input signal is expressed as $x_{i} $. The hidden layer has l hidden neurons, and any of output layer neuron (OLN) is denoted by t. The output layer owns $n$ output neurons. The activation transfer function is the key of NNs and determines their utility, as shown in Eq. (2).

(2)
$ f(t)=\left\{\begin{aligned} &1,&&t\ge 0,\\ &0,&&t<0. \end{aligned}\right. $

In Eq. (2), $f(t)$ is a threshold function that reflects the excitation or inhibition of neurons. The network training adopts an error backpropagation algorithm, sets a set of samples, and calculates the connection weight matrix (CWM) in the layer of input and hidden, as shown in Eq. (3).

(3)
$ W^{1} =\left[\!\!\begin{array}{ccccc} {w_{11}^{1} } & {w_{12}^{1} } & {w_{13}^{1} } & {...} & {w_{1m}^{1} } \\ {w_{21}^{1} } & {w_{22}^{1} } & {w_{23}^{1} } & {...} & {w_{2m}^{1} } \\ {...} & {...} & {...} & {...} & {...} \\ {w_{l1}^{1} } & {w_{l2}^{1} } & {w_{l3}^{1} } & {...} & {w_{lm}^{1} } \end{array}\!\!\right] . $

The CWM between the output and the hidden layers is calculated as shown in Eq. (4).

(4)
$ W^{2} =\left[\!\!\begin{array}{ccccc} {w_{11}^{2} } & {w_{12}^{2} } & {w_{13}^{2} } & {...} & {w_{1m}^{2} } \\ {w_{21}^{2} } & {w_{22}^{2} } & {w_{23}^{2} } & {...} & {w_{2m}^{2} } \\ {...} & {...} & {...} & {...} & {...} \\ {w_{l1}^{2} } & {w_{l2}^{2} } & {w_{l3}^{2} } & {...} & {w_{lm}^{2} } \end{array}\right] . $

The thresholds $\theta ^{1} $ and $\theta ^{2} $ of HLN and OLN are calculated, as shown in Eq. (5).

(5)
$ \begin{align} & \theta ^{1} =\left[\theta _{1}^{1}~~ \theta _{2}^{1}~~\theta _{3}^{1}~~ ...~~\theta _{l}^{1} \right],\nonumber\\ & \theta ^{2} =\left[\theta _{1}^{2} ~~\theta _{2}^{2}~~ \theta _{3}^{2} ~~...~~\theta _{l}^{2} \right] . \end{align} $

The output of HLN is obtained from the forward propagation of NN working signals, as shown in Eq. (6).

(6)
$ \begin{align} &O_{j} =f\left(\sum _{i=1}^{m}w_{ji}^{1} -\theta _{j}^{1} \right)=f\left(net_{j} \right),\nonumber\\ &j=1,~2,~3,~...,~l. \end{align} $

In Eq. (6), $l$ represents the amount of HLN. The expected error based on the OLN output is obtained as shown in Eq. (7).

(7)
$ \begin{align} E&=\frac{1}{2} \sum _{k=1}^{n}\Bigg\{y_{k} -g\Bigg[\sum _{j=1}^{l}w_{kj}^{2} f\left(\sum _{i=1}^{m}w_{ji}^{1} -\theta _{j}^{1} \right)\nonumber\\ &\quad -\theta _{k}^{2} \Bigg]\Bigg\}^{2} . \end{align} $

In Eq. (7), $\theta _{j}^{1} $ and $\theta _{k}^{2} $ represent the thresholds of the corporate HLN and OLN. Next, the error signal is backpropagated and the link values are modified layer by layer. Finally, the partial derivative of the connecting weight value to the error is obtained, as shown in Eq. (8).

(8)
$ \left\{\begin{aligned} & \theta _{j}^{1} (t+1)=\theta _{j}^{1} (t)+\Delta \theta _{j}^{1} \\ &\hskip 3.1pc =\theta _{j}^{1} (t)-\eta ^{1} \frac{\partial E}{\partial \theta _{j}^{1} }\\ &\hskip 3.1pc =\theta _{j}^{1} (t)+\eta ^{1} \delta _{j}^{1},\\ & \theta _{j}^{2} (t+1)=\theta _{j}^{2} (t)+\Delta \theta _{j}^{2}\\ &\hskip 3.1pc =\theta _{j}^{2} (t)-\eta ^{2} \frac{\partial E}{\partial \theta _{j}^{2} }\\ &\hskip 3.1pc =\theta _{j}^{2} (t)+\eta ^{2} \delta _{j}^{2}. \end{aligned}\right. $

Due to overtraining, the network falls into local minima, resulting in overfitting and affecting generalization ability. Fig. 1 shows the NN process.

The ``timely termination'' is adopted to optimize the process to define target errors to a small extent and provide sufficient training space. This method monitors the timely error changes of the test set. When the network suffers from a local min and the test set error cannot jump out of the local minimum within the specified cycles numbers, the training will be terminated and the optimal state of the set will be output.

Fig. 1. The training process of neural networks.

../../Resources/ieie/IEIESPC.2025.14.2.205/image1.png

3.2 Comprehensive Improvement of NNs and SVM for Predicting High Performance BCS

Due to changes in network structure, the predictive performance of NNs varies. Therefore, it is necessary to randomly define the IWT and divide the TrS & TeS according to a certain proportion, and repeat the training of the network multiple times to analyze its performance [15]. The performance parameters of commonly used evaluation and prediction models are shown in Eqs. (9) and (10).

(9)
$ R^{2} =\frac{\left(n\sum\limits _{i=1}^{n}\hat{y}_{i} y_{i} -\sum\limits _{i=1}^{n}\hat{y}_{i} \sum\limits _{i=1}^{n}y_{i} \right)}{\left(n\sum\limits _{i=1}^{n}\hat{y}_{i} {}^{2} -\left(\sum\limits _{i=1}^{n}\hat{y}_{i} \right)^{2} \right)\left(\sum\limits _{i=1}^{n}\hat{y}_{i} {}^{2} -\left(\sum\limits _{i=1}^{n}y_{i} \right)^{2} \right)} . $

The calculation of mean square error (MSE) and mean absolute percentage error (MAPE) is Eq. (10).

(10)
$ \left\{\begin{aligned} & MSE=\frac{1}{n} \sum _{i=1}^{n}\left(\hat{y}_{i} -y_{i} \right)^{2},\\ & MAPE=\frac{1}{n} \sum _{i=1}^{n}\left(\left|\frac{\hat{y}_{i} -y_{i} }{y_{i} } \right|\right)\times 100. \end{aligned}\right. $

In Eqs. (9) to (10), $n$ represents the samples. $\hat{y}_{i} $ and $y_{i} $ represent the target and predicted values of the $i$-th sample. NNs provide direction for performance prediction of BCS. However, NNs are all influenced by certain factors and exhibit certain instability, so it is necessary to optimize them [16]. When using the "timely termination" method to solve the overfitting problem of neural networks, it is necessary to divide the training samples into a training set and a validation set. The training set is used for training network weights and thresholds, while the validation set is used to detect changes in training errors, thereby determining the optimal number of training cycles. Therefore, the division of the training set and validation set will also affect the predictive performance of the trained network. Therefore, a genetic algorithm based optimization algorithm for dividing the training and validation sets is proposed in this study. This chapter will compare and study the existing optimization methods for initial weight thresholds of neural networks using genetic algorithms with the neural network training and validation set optimization algorithm proposed in this paper. By integrating the two optimization algorithms, a new comprehensive optimization algorithm for initial weight values of neural networks and optimization algorithms for training and validation sets will be proposed. In traditional GA, a fitness function is constructed based on the objective function of the problem to be optimized, an initial population is created, followed by genetic coding, evaluation, selection, and calculation of the already encoded population, and multiple iterations are performed. After the above steps, the optimal adaptive individual is the optimal solution of the problem [17]. Fig. 2 shows the flowchart of the basic GA.

Fig. 2. Genetic algorithm flowchart.

../../Resources/ieie/IEIESPC.2025.14.2.205/image2.png

Firstly, to generate an initial population randomly, calculate the fitness of each individual, and then evaluate their ability to adapt to the environment through a defined fitness function. Selecting individuals based on fitness and reproduce the next generation, followed by genetic operations, including crossover and mutation, to ultimately test the termination conditions of the algorithm. Optimizing the division of the TrS & TeS of the NN, using binary encoding to achieve individual encoding. On this basis, this study proposes a sample partitioning method based on chromosome gene information. The specific optimization population initialization method, encoding, and genetic operation steps are as follows. Initializing the population, marking the sorted training samples with an index number, and dividing them into a TrS & TeS proportionally according to random sorting, to obtain a random quantity of individuals in the given population. To assign the index numbers of the obtained individuals and generate chromosomes. 1 represents the training set sample, and 0 represents the test set sample. The new genetic operation involves adjusting the mutation and crossover operators, as shown in Fig. 3.

Fig. 3. Genetic manipulation.

../../Resources/ieie/IEIESPC.2025.14.2.205/image3.png

The crossover step first decodes the parent chromosome to obtain the set to which the sample belongs, and then integrates the test set samples of the decoded two individuals. Randomly to select a certain number of samples as new test set samples, while the rest are training set samples. Based on the new TrS & TeS samples, recoding and obtaining offspring chromosomes. The mutation operation involves randomly selecting a certain amount of training and testing samples for equal sample exchange [18]. Based on the optimization of the IWT and test set partitioning of NNs mentioned above, this study proposes a comprehensive NN optimization method combining the two, and establishes a high-performance prediction model for BCS. By comparing the initial weight threshold and optimization methods for training and validation sets of neural networks based on genetic algorithms, and considering that both optimization methods can have a certain optimization effect on the performance of the network. Therefore, a comprehensive neural network optimization algorithm is proposed, which combines the optimization objectives of the above two methods. In the main line of the optimization methods for the training and validation sets with obvious optimization effects, the initial weight threshold optimization of each individual's network is added to achieve comprehensive optimization of the neural network, in order to find the best network state to predict the response of the belt transportation system. The comprehensive process of the algorithm is Fig. 4.

Fig. 4. Comprehensive optimization algorithm flowchart.

../../Resources/ieie/IEIESPC.2025.14.2.205/image4.png

Firstly, an initial population with different combinations of training and validation sets is generated. Then, before optimizing the training and validation sets, optimization of the initial network weight threshold for each individual is added. After optimizing the initial weight threshold of each individual in the network, the optimal network performance for each training and validation set individual is obtained, and then genetic evolution is performed on them to generate a new population. Because the partition of the training set and validation set is different, the corresponding optimal initial weight threshold of the network is also different. Therefore, before the evolution of each generation of new populations, the network initial weight and threshold will also be optimized. The termination condition for the final evolution is still the maximum evolution generation. Once the maximum evolution generation is reached, a network with the best partition of training and validation sets is output, and the initial weight threshold of the network is also its corresponding optimal value.

Both optimizations in Fig. 4 are completed by GA, and the fitness function is shown in Eq. (11).

(11)
$ \left\{\begin{aligned} & Fitness=R^{2} =1-SSE/SS{\rm T},\\ & SSE=\sum _{i=1}^{m}\left(y_{i} -\hat{y}_{i} \right)^{2},\\ & SST=\sum _{i=1}^{m}\left(y_{i} -\bar{y}\right)^{2}. \end{aligned}\right. $

In Eq. (11), $\hat{y}_{i} $ is the estimated value of $y_{i} $, $\bar{y}$ is the average value of $y$, $SSE$ is the sum of squared residuals, and $SST$ is the sum of squared total deviations.

After optimization, the system performance was predicted and SVM parameters were selected in combination with FSA. FSA is an intelligent optimization algorithm based on fish behavior, which gradually optimizes by updating its own position during each iteration process [19]. The current state of the $i$-th fish is set to $X_{i} $, with a fitness of $Y_{i} $. Each fish selects a state at random, as shown in Eq. (12).

(12)
$ \left\{\begin{aligned} & X_{v} =X_{i} +Rand()\times Visual,\\ & X_{i\mid next} =X_{i} +Rand()\times Step\times \frac{X_{v} -X_{i} }{\left\| X_{v} -X_{i} \right\| } ,\\ & X_{i\mid next} =X_{i} +Rand()\times Step. \end{aligned}\right. $

In Eq. (12), $Rand()$ is a random number generated between 0 and 1. $X_{i} $ represents the current state of the $i$-th fish. $\left\| X_{v} -X_{i} \right\| $ represents the distance between $X_{v} $ and $X_{i} $. $X_{i\mid next} $ represents the next state of the $i$-th fish. FSA mimics the way fish flock, allowing them to move towards the center while limiting crowding and avoiding overcrowding, as shown in Eq. (13).

(13)
$ \begin{align} &S_{j} =\left\{X_{i} \left\| X_{j} -X_{i} \right\| \right\}\le Visual,\nonumber\\ &j=1,~2,~...,~i+1,~...,~N. \end{align} $

In Eq. (13), with one's own position as the center, the number of fish within the perception range is $N_{j} $, forming a set $S_{j} $. If set $S_{j} $ is an empty set, indicating that there are fish within the perception range of the $i$-th fish, then calculate the center position $X_{center} $ of the set according to Eq. (14) and calculate the fitness value $Y_{center} $ of the center position.

(14)
$ X_{center} =\frac{\sum _{j=1}^{N_{f} }X_{j} }{N_{f} } . $

If tail chasing behavior occurs, it indicates that each fish is pursuing the closest fish with the highest fitness [20]. If $X_{\min } $ is the center, the number of fish $N_{f} $ within the perception range satisfies Eq. (15).

(15)
$ Y_{\min } <Y_{i}~~AND~~N_{f} ,~Y_{\min } <\delta ,~Y_{i},~~(\delta >1). $

In Eq. (15), $X_{i} $ searches for the fish $X_{\min } $ with the best fitness within its perception range, and its fitness value is $Y_{\min } $. Then, it is determined that if $Y_{\min } >Y_{i} $, foraging behavior will be performed, and vice versa, Eq. (15) is satisfied. If it indicates that the location has good adaptability and is not too crowded, then execute Eq. (16) to move forward in the direction of fish $X_{\min } $ with the best adaptability. Otherwise, perform foraging behavior.

(16)
$ X_{i|next} =X_{i} +Rand()\times Step\times \frac{X_{\min } -X_{i} }{\left\| X_{\min } -X_{i} \right\| } . $

In Eq. (16), $\left\| X_{v} -X_{i} \right\| $ represents the distance between $X_{v} $ and $X_{i} $. Fig. 5 shows FSA process.

When SVM is used for BCS performance prediction, its prediction performance may vary due to differences in kernel function parameters and penalty factors. And for different prediction targets, the distribution also varies [21]. When using FSA to optimize SVM, the algorithm can accurately and efficiently optimize, and the optimization is stable. Regardless of the initial population distribution state, it can quickly converge to the optimal region. The fish species in each iteration are concentrated towards the optimal region. The algorithm can obtain a more detailed search in the optimal region [22]. However, to avoid the algorithm falling into local optima, restrictions have been placed on the crowding degree of the fish school.

Fig. 5. Flowchart of fish school algorithm.

../../Resources/ieie/IEIESPC.2025.14.2.205/image5.png

3.3 Overview of Integration of Multiple Technologies

In order to optimize the selection of initial weight thresholds for neural networks, a study was conducted on the partitioning of training and testing sets based on genetic algorithms, combined with the research designed optimization methods for partitioning training and testing sets. Firstly, optimize the initial weight threshold and dataset partitioning of the network. The establishment of a neural network prediction model includes several steps, including the selection of input and output parameters, normalization processing, selection of network structure, and network training. By randomly defining initial weight thresholds and dividing the training and testing sets into a certain proportion, the network is trained multiple times to analyze its performance; By comparing the initial weight thresholds and optimization methods of neural network training and validation sets based on genetic algorithms, it was found that both optimization methods have a certain optimization effect on the performance of the network. Therefore, combining the optimization objectives of the above two methods. In the main line of optimization methods, for the training and validation sets with obvious optimization effects, the initial weight threshold optimization of each individual network was added to achieve comprehensive optimization of the neural network, thereby finding the best network state to predict the response of the belt transportation system.

4. Verification of Predicting High Performance Transportation Systems Based on Interpretable Machine Learning Models

This study verified the performance prediction of the designed improved NN and SVM algorithm, exploring the feasibility and effectiveness of interpretive machine learning in predicting high-performance BCS. Moreover, the improvement effects were compared and corresponding data analysis was obtained.

4.1 Preparation and Design of Experimental Data

A comprehensive study on the optimization of IWT and training set test sets for NNs was conducted. 10 sets of IWT and ten sets of training set and validation set (Ts-Vs) partitioning combinations were randomly generated in the experiment. Using the optimization method of IWT to generate 10 initial populations. For the training set and validation set optimization methods, the IWTs are 10 sets of random IWTs, respectively. Each population has the same combination of 10 Ts-Vs. Based on this, the optimization effects of two optimization methods on NNs were compared. The optimization of NNs selects a network structure with relatively stable performance and four neurons in the hidden layer. The maximum evolutionary algebra set in the experiment is 100 generations, and the population size remains the same as the initial population size at 10. Table 1 shows the relevant parameters of the experiment.

Table 1. Correlation coefficients involved in the experiment.

Parameter name and serial number

1

2

3

System parameter

Transmission distance

Speed

Response time

Number of training sets

10

Number of test sets

10

Iterations

100

Number of hidden layer neurons

4

4.2 Analysis of Predictive Performance of Predictive Models

For the evaluation of the initial population, Fig. 6 shows the statistical results of the predictive performance for each initial population. Fig. 6(a) shows the IWT optimization network prediction performance distribution state. Fig. 6(b) shows the distribution status of network prediction performance with a combination of random training set and validation set partitioning. In Fig. 6(a), there is a significant difference in the distribution of the box line structure, indicating that both optimization methods will have a certain impact on the performance of the network. Fig. 6(b) shows the average and minimum values of each initial population, used to analyze the trend of changes between different populations. Compared to the data in Figs. 6(a) and Fig. 6(b), it can be observed that compared to groups with different IWTs, the differences between groups with different combinations of Ts-Vs partitioning are much greater. In Fig. 6(a), there is a significant difference between each box and the overall moving average. The box distribution in Fig. 6(b) is uniform, and there is not much difference between the average and minimum values. The minimum values of all populations are significantly lower than the overall moving average. The minimum values of all populations in the network are significantly lower than the overall moving average.

Fig. 6. Initial population performance evaluation of IWT optimization and training set test set optimization.

../../Resources/ieie/IEIESPC.2025.14.2.205/image6.png

Fig. 7 shows the optimization trajectories of different populations under two optimization methods. Fig. 7(a) shows the trajectory of optimizing the IWT of the network under different combinations of Ts-Vs partitioning. With the increase of evolutionary algebra, network performance gradually decreases, indicating the effectiveness of optimization algorithms. Dividing the Ts-Vs is more effective than optimizing the IWT. Due to the differences between the initial group of Ts-Vs, there may be significant deviations in the optimization process of the IWT. Therefore, the optimization process usually takes longer to converge. Fig. 7(b) shows the optimization trajectories of different populations under different IWTs by dividing the Ts-Vs. As shown in Fig. 7(b), optimizing populations with different initial network weight thresholds through Ts-Vs helps to quickly converge to the optimal value and ensure relative stability in the later evolution stage. All evolutionary processes reach their optimal values within 17 generations.

Fig. 7. Optimization trajectories of different populations under two optimization methods.

../../Resources/ieie/IEIESPC.2025.14.2.205/image7.png

Fig. 8 shows the synthesis of the final optimization amplitude of two optimization algorithms on network performance, and compares the optimization effects of the two optimization methods on network performance. Fig. 8(a) represents the optimization amplitude of two methods for different initial populations. Fig. 8(b) is the overall average of the two optimization methods relative to the initial population. Fig. 8(a) indicates that both optimization methods have improved the predictive performance of the network to varying degrees. There is a significant difference between the second and eighth red columns in Fig. 8(b). This indicates that the predictive ability of the network after optimizing through weight threshold has not yet reached the average level of all individuals in the initial population. Compared to the average level of all individuals in the initial population, Ts-Vs optimization can more effectively improve the performance of the network than optimizing the IWT.

Fig. 9 shows the evolutionary trajectory of the proposed comprehensive optimization algorithm for 100 generations. The optimization algorithm stably converges to the optimal value through only 14 generations of evolution, resulting in an optimal network prediction MSE of 0.011678. Previously, the average network prediction MSE of all individuals in the initial population was 0.016845. Therefore, using this comprehensive optimization algorithm to optimize the prediction performance of the network reaches 29.6%, and the optimization effect of the NN is superior to both optimization algorithms.

Fig. 8. Comparing the optimization effects of two optimization methods on network performance.

../../Resources/ieie/IEIESPC.2025.14.2.205/image8.png

Fig. 9. Evolutionary trajectory of comprehensive optimization algorithm.

../../Resources/ieie/IEIESPC.2025.14.2.205/image9.png

Table 2 shows the time occupied by the four calculation methods. The training of the prediction model takes the longest time, and the parameters of it were optimized during the establishment of the model. The optimization process requires repeated training and evaluation of the prediction model, which takes up a long time. However, the total time required for optimizing the performance prediction of BCS is only 5.991 seconds, which is relative to the highest speed of 1750 r/min.

Table 2. Time occupation in the four calculation methods.

Calculation work

Data processing

Predictive Model Training

Fitness function generation

Optimization calculation

Holding time

0.948

2.514

0.684

1.845

Total elapsed time

5.991

Occupancy percentage

15.82%

41.96%

11.42%

30.80%

After obtaining a set of optimization results, in order to ensure the accuracy and precision of the optimization results, the simulation model will be used again to verify the accuracy of the optimization results. If the accuracy does not reach the set goal, the new data obtained from the simulation model can be added to the previously obtained database, and all the obtained simulation data can be used to reinforce the support vector machine prediction model again. Then, the support vector machine after reinforcement training is used again to optimize and validate the diesel engine, until the optimization results that meet the accuracy requirements are obtained. In this way, reinforcement learning and optimization of the prediction model can be synchronized, ensuring optimization accuracy while reducing the number of simulation calculations and avoiding waste caused by excessive data. Fig. 10 shows the combination of SVM reinforcement learning and BCS operation parameter optimization process. Fig. 10(a) describes the variation of SVM response prediction accuracy with each BCS reinforcement training. The overall prediction accuracy steadily improves with the increase of reinforcement training. Overall, the proposed reinforcement training method can significantly improve the accuracy of SVM and ensure the accuracy of optimization results. Fig. 10(b) shows the continuously improving performance of BCS simulation throughout the entire optimization process. After the 10th reinforcement training, the total number of simulations was 65, and the prediction error percentage (PEP) of SVM decreased to 0.99%.

Fig. 10. The combination of SVM reinforcement learning and the optimization process of operating parameters in BTS.

../../Resources/ieie/IEIESPC.2025.14.2.205/image10.png

5. Conclusion

NNs in interpretable machine learning are widely used in system performance prediction. However, the performance prediction accuracy and speed of existing NNs are not sufficient to meet the requirements. This manuscript improved the predictive performance of the NN by optimizing the IWT and partitioning the Ts-Vs. Optimizing the partitioning of both the Ts-Vs could effectively improve the predictive performance of the network. Moreover, the improved SVM algorithm combined with FSA improved the accuracy of prediction. The experiment showed that the total time required for optimizing the performance prediction of BCS was only 5.991 seconds, and the response speed was fast. After the 10th reinforcement training, the PEP of SVM decreased to 0.99%, and the total number of simulations was equal to 65. The comprehensive optimization algorithm converged stably to the optimal value through 14 generations of evolution, resulting in an optimal network prediction MSE of 0.011678. The optimization degree of its prediction performance on the network reached 29.6%, and the optimization effect on the NN was good. It could be proved that the accuracy and feasibility of the proposed comprehensive optimization NN in predicting system performance were very obvious. However, the study of BCS parameters only considered some parameters, so further analysis and comprehensive and sufficient parameter considerations are needed in the future.

REFERENCES

1 
J. Smith-Colin, A. A. Amekudzi-Kennedy, and G. Kingsley, ``Role of inputs, processes, and relations in transportation system performance management: Case study of regional transportation collaborations,'' Journal of Management in Engineering, vol. 37, no. 1, pp. 1943-5479, 2021.DOI
2 
A. L. Alfeo, M. G. C. A. Cimino, and G. Vaglini, ``Degradation stage classification via interpretable feature learning,'' Journal of Manufacturing Systems, vol. 62, no. 3, pp. 972-983, 2021.DOI
3 
Y. Choi, H. Nguyen, X. N. Bui, and T. T. Nguyen, ``Optimization of haulage-truck system performance for ore production in open-pit mines using big data and machine learning-based methods,'' Resources Policy, vol. 75, 102522, 2022.DOI
4 
M. Zhang, K. Jiang, Y. Cao, M. Li, N. Hao, and Y. Zhang, ``A deep learning-based method for deviation status detection in intelligent conveyor belt system,'' Journal of Cleaner Production, vol. 363, 132575, 2022.DOI
5 
P. E. Bhaskaran, C. Maheswari, S. Thangavel, M. Ponnibala, T. Kalavathidevi, and N. S. Sivakumar, ``IoT based monitoring and control of fluid transportation using machine learning,'' Computers and Electrical Engineering, vol. 89, no. 34, 106899, 2020.DOI
6 
O. Daoudr, Q. Hamarsheh, and A. Damati, ``Enhancing the LTE-based intelligent transportation system's performance,'' Digital Signal Processing, vol. 99, 102677, 2020.DOI
7 
X. Li, Y. Li, Y. Zhang, F. Liu, and Y. Fang, ``Fault diagnosis of belt conveyor based on support vector machine and grey wolf optimization,'' Mathematical Problems in Engineering, vol. 2020, 1367078, 2020.DOI
8 
S. Y. Bae and Y. H. Kim, ``Structural design and mechanical performance evaluations of flexible and lightweight belt by CFRP,'' Modern Physics Letters B, vol. 34, no. 7/9, 2040034, 2020.DOI
9 
X. Liu, D. Pei, G. Lodewijks, Z. Zhao, and J. Mei, ``Acoustic signal based fault detection on belt conveyor idlers using machine learning,'' Advanced Powder Technology, vol. 31, no. 7, pp. 2689-2698, 2020.DOI
10 
A. Li, F. Xiao, C. Zhang, and C. Fan, ``Attention-based interpretable neural network for building cooling load prediction,'' Applied Energy, vol. 299, 117238, 2021.DOI
11 
M. Carletti, M. Terzi, and G. A. Susto, ``Interpretable anomaly detection with DIFFI: Depth-based feature importance of isolation forest,'' Engineering Applications of Artificial Intelligence, vol. 119, 105730, 2023.DOI
12 
G. Mitrentsis and H. Lens, ``An interpretable probabilistic model for short-term solar power forecasting using natural gradient boosting,'' Applied Energy, vol. 309, 118473, 2022.DOI
13 
S. Shin, Y. Kim, and J. W. Yoon, ``A new approach to training more interpretable model with additional segmentation,'' Pattern Recognition Letters, vol. 152, pp. 188-194, 2021.DOI
14 
T. Bikmukhametov and J. Jschke, ``Combining machine learning and process engineering physics towards enhanced accuracy and explain ability of data-driven models,'' Computers and Chemical Engineering, vol. 138, 106834, 2020.DOI
15 
S. Pawar, O. San, A. Nair, A. Rasheed, and T. Kvamsdal, ``Model fusion with physics-guided machine learning: Projection-based reduced-order modeling,'' Physics of Fluids, vol. 33, no. 6, 067123, 2021.DOI
16 
T. Dong, Q. Yang, N. Ebadi, R. Luo, and P. Rad, ``Identifying incident causal factors to improve aviation transportation safety: Proposing a deep learning approach,'' Journal of Advanced Transportation, vol. 2021, 5540046, 2021.DOI
17 
C. Moreira, Y. L. Chou, M. Velmurugan, C. Ouyang, R. Singhgatta, and P. Bruza, ``LINDA-BN: An interpretable probabilistic approach for demystifying black-box predictive models,'' Decision Support Systems, vol. 150, 113561, 2021.DOI
18 
M. Bunyakitanon, A. P. D. Silva, X. Vasilakos, R. Nejavati, and D. Simeoidou, ``Auto-3P: An autonomous VNF performance prediction & placement framework based on machine learning,'' Computer Networks, vol. 181, 107433, 2020.DOI
19 
R. K. C. Chan, M. Y. Lim, and R. Parthiban, ``A neural network approach for traffic prediction and routing with missing data imputation for intelligent transportation system,'' Expert Systems with Applications, vol. 171, no. 2, 114573, 2021.DOI
20 
Y. Kuo, N. Chan, and J. Leung, ``An integrated approach of machine learning and systems thinking for waiting time prediction in an emergency department,'' International Journal of Medical Informatics, vol. 139, 104143, 2020.DOI
21 
X. Yang, Y. Zou, J. Tang, J. Liang, and M. Ijaz, ``Evaluation of short-term freeway speed prediction based on periodic analysis using statistical models and machine learning models,'' Journal of Advanced Transportation, vol. 2020, 9628957, 2020.DOI
22 
Y. Guo, Z. Mustafaoglu, and D. Koundal, ``Spam detection using bidirectional transformers and machine learning classifier algorithms,'' Journal of Computational and Cognitive Engineering, vol. 2,, no. 1, pp. 5-9, 2023.DOI

Author

Peng Gao
../../Resources/ieie/IEIESPC.2025.14.2.205/author1.png

Peng Gao was born in February 1994, male, native to Shenmu City, Shaanxi Province, Han ethnicity. He obtained a bachelor's degree in intelligent building and electrical engineering from Xi'an University of Architecture and Technology in 2017 and a master's degree in intelligent building engineering from Xi'an University of Architecture and Technology in 2020. His research focuses on wireless communication and sensing networks. Work experience: From 2020 to present, he has been the intelligent supervisor of Guoneng Shendong Baode Coal Mine, mainly responsible for the technical management of coal mine intelligent construction. Academic situation: Published 2 EI papers, participated in 1 of the top 10 key scientific research projects of the National Energy Group, and has accepted 13 invention patents.

Jinlin Ruan
../../Resources/ieie/IEIESPC.2025.14.2.205/author2.png

Jinlin Ruan was born in February 1987, male, native to Shenmu City, Shaanxi Province, Han ethnicity, obtained a bachelor's degree in electrical engineering and automation from China University of Mining and Technology in 207. Work experience: From July 2011 to November 2013, worked as an electrician in the second excavation anchor team of Baode Coal Mine in Shendong Coal Mine; From December 2013 to October 2016, worked as a clerk at the Mechanical and Electrical Information Center of Baode Coal Mine in Shendong Coal Mine; From November 2016 to August 2019, worked as an electromechanical technician in the anchor excavation team of Baode Coal Mine in Shendong Coal Mine; From September 2019 to September 2020, served as the mechanical and electrical technician and deputy secretary of the Shendong Coal Baode Coal Mine Comprehensive Mining Team; From October 2020 to October 2022, served as the Secretary of the Third Party Branch of Baode Coal Mine, Deputy Director of the Mechanical and Electrical Management Office, and Leader of the Intelligent Execution Group; From October 2023 to present, Secretary and Director of the Party Branch of the Mechanical and Electrical Management Office of Baode Coal Mine, Shendong Coal. Academic situation: Published 3 scientific and technological papers since work, participated in 2 of the top 10 key scientific research projects of the National Energy Group, authorized 8 patents, including 2 invention patents.

Yuan Sun
../../Resources/ieie/IEIESPC.2025.14.2.205/author3.png

Yuan Sun was born in December 1995, male, native to Xinzhou City, Shanxi Province, Han ethnicity. In July 2018, he obtained a bachelor's degree in electrical engineering and automation from the College of Information, Shanxi Agricultural University, with a research focus on coal mine intelligence. Work experience: From September 2019 to July 2020, worked as an electrician in the second excavation anchor team of Guoneng Shendong Baode Coal Mine; From August 2020 to November 2021, worked as an inspector for the second anchor excavation team of Guoneng Shendong Baode Coal Mine; From December 2021 to September 2022, worked as a data engineer for the intelligent operation and maintenance team of Guoneng Shendong Baode Coal Mine; From October 2022 to present, a staff member of the Mechanical and Electrical Management Office of Guoneng Shendong Baode Coal Mine. Academic situation: Participated in one of the top ten key scientific research projects of the National Energy Group, and has accepted 11 invention patents.

Hao Li
../../Resources/ieie/IEIESPC.2025.14.2.205/author4.png

Hao Li was born in July 1983, male, native of Handan City, Hebei Province, Han ethnicity. He obtained a bachelor's degree in electronic information engineering from University of Science and Technology Beijing from September 2002 to July 2006, a master's degree in electronic information and computer engineering from the University of Nottingham in the UK from August 2006 to January 2008, and a Ph.D. degree in mechanical and electronic engineering from China University of Mining and Technology (Beijing) from September 2013 to December 2019. Research direction: Intelligent perception and control. Work experience: From March 2012 to July 2018, worked as a system software engineer in the unmanned project department of Beijing Tiandi Marco Electric Hydraulic Control System Co., Ltd; From July 2018 to August 2019, served as the Deputy Manager of the Information Management Department of Beijing Tiandi Marco Electric Hydraulic Control System Co., Ltd; August 2019 to August 2021: Backbone scientific researchers of the Mining Big Data Research Institute of the General Institute of Coal Science Research; From August 2021 to present, Vice President of Ningxia Research Institute of Coal Science Research Institute Co., Ltd. Academic situation: More than ten papers have been publicly published, of which 8 have been indexed by SCI/EI; Responsible for participating in and completing 8 national, provincial, and ministerial level scientific research projects in the past 5 years, with 1 project ranking first and 2 projects ranking second; In the past 5 years, 7 invention patents have been authorized, and 1 second prize of National Science and Technology Progress Award and 3 first prizes of provincial and ministerial level have been won.}

Cong Sang
../../Resources/ieie/IEIESPC.2025.14.2.205/author5.png

Cong Sang was born in October 1987, male, native of Tai'an, Shandong, Han ethnicity, obtained a bachelor's degree in safety engineering from North China Institute of Science and Technology from September 2006 to July 2010, and a master's degree in safety technology and engineering from Liaoning Technical University from September 2010 to January 2013. My research direction is intelligent ventilation and safety in mines. Work experience: From August 2013 to July 2021, served as the business supervisor of the Safety Branch of Coal Science and Technology Research Institute Co., Ltd; From August 2021 to present, I have been a research backbone of the Mining Big Data Research Institute of Coal Science Research Institute Co., Ltd. Academic situation: More than 20 papers have been publicly published, nearly 10 national and provincial-level scientific research projects have been completed, 11 invention patents have been authorized, and 4 provincial-level science and technology awards have been won.