Mobile QR Code QR CODE




Enterprise information management, Evolutionary game, Fireworks algorithm, Levenberg-Marquardt algorithm, Neural network, Principal component analysis method, Performance simulation

1. Introduction

The globalization of the economy is developing at a speed that cannot be ignored, and with it, research and application of economic game theory are becoming more and more extensive. Many scholars and experts gradually apply the concept of game theory to enterprise information management in this new era to solve the dilemma and problems of enterprise operation management [1,2]. Management of the enterprise actually maximizes the interests of the enterprise. Enterprise information management refers to digitizing important business transactions such as process production, capital processing, business processing, customer communication, etc., processing this information into new resources through various information exchange systems and networks [3,4]. Enterprise information management belongs to a strategic management category of enterprise development, but due to various restrictions, such as economic and era development, the construction of enterprise information management systems is not accurate enough [5]. In current computer technology, the construction of an enterprise information management and game system by using various artificial intelligence algorithms has gradually become the focus of research in the industry. With the development and application of machine learning technology, the back propagation (BP) neural network has been widely used in many fields, obtaining a lot of research results. In addition, the fireworks algorithm (FWA), a recent group intelligence optimization algorithm, has garnered attention being successfully merged with data sensing by scholars to yield positive outcomes. In view of this, this study combines the FWA and BP algorithms. In the process, principal component analysis (PCA) is used to reduce the dimensions of the data, and the least square is added to optimize the model calculation process.

This study has two main innovations. The first is to introduce PCA to reduce the dimensions of the input model vector in order to improve data processing efficiency and output accuracy. The second is using the Levenberg-Marquardt (LM) algorithm to improve the accuracy of information data processing during system operation. This paper is divided into four main parts. Section~2 is a summary of the existing domestic and foreign research results. Section 3 discusses PCA and the LM algorithm to optimize model parameters, and how the FWA\hbox{-}BP algorithm builds the enterprise information management game system. Section 4 presents a performance evaluation of the research model, and Section 5 summarizes the paper.

2. Related Works

Enterprise information management plays a very significant role in the development of enterprises. It is a game of information between enterprises, which determines their prospects and development in different industries. Zhang et al. investigated unnatural control chart patterns and found they may be related to data in the production process. These data were different combinations of two or more basic patterns [6]. Aiming at the problem of combination, a hybrid CCP identification method was proposed, which integrated the support vector machine boosted by the fireworks algorithm and feature reduction. The final results showed that the proposed method can affect high precision and effectively cut down runtime. Gong et al. proposed predictive modeling that integrates the fireworks approach and long short-term memory to improve prediction accuracy and convergence speed. The data showed that compared with existing algorithm molds, the rendition of the proposed mold was significantly improved, and the mold error significantly reduced [7]. Sebbane and El Akchioui proposed a new diagnostic system for photovoltaics that integrated the inherent meaning of artificial neural networks and the fireworks algorithm. In a comparative test, the new hybrid model was completely superior to the traditional model in terms of high precision and convergence, and the system’s diagnostic assignment was more reasonable and simpler. The data showed that the mixed mode can obtain fidelity of 99.98% after iterating 241 times, which was significantly excellent [8]. Manivannan and Kavitha proposed a new evolutionary classification model to judge the type of diabetes better. The model combined a diabetes detection system with strengthened growth pyrotechnics as well as brain storm optimization [9]. Experiments compared the precision, recall, and decision-making accuracy of different algorithms, proving the superiority of the proposed algorithm. In order to better realize the CAFE of electric vehicles and improve the performance of electric motor charging, Wang et al. proposed applying an improved fireworks algorithm to optimize the controller parameters. The data showed that fuel economy improved by 10% in typical and real urban cycle driving, compared with an electric motor without algorithm model optimization [10].

de Pinho Pinheiro et al. proposed using deep learning and swarm intelligence technology for detection and model construction. The process joined convolutional neural networks and different biological image segmentation techniques to find cancer nodules. The superiority of the proposed method from aspects of accuracy, precision, sensitivity, and specificity was confirmed by experiments [11]. Y\i{}lmaz et al. proposed a pi-sigma artificial neural network to enrich neural network algorithms. The proposal was criticized through two number covers, and the mutation strategy differential evolution algorithm was used for experimental assistance in the process. The outcome showed that the proposed way had superior action [12]. To address the difficulty in accurately estimating probability integral parameters from measured data in geological mining, Wang et al. introduced an improved fireworks algorithm and integrated mining subsidence law to build a reverse deduction model of integral parameters. Experiments showed that the posed way had a very good showing against observation errors and geological mining [13]. Zhang et al.~proposed an adaptive fault diagnosis method based on the fireworks algorithm and the multi-vector support algorithm for linearity and multiphase characteristics of a multi-stage intermittent process. In that process, kernel entropy component analysis was used to realize adaptive segmentation, and the similarity of the corner structure was used as a statistic to construct the model [14]. In a verification experiment, the penicillin fermentation process was taken as an instance to prove its effectiveness. Yadav et al. proposed combining the fireworks algorithm with various heuristics to achieve efficient scheduling. To stack up the actions of the algorithms, various experiments were carried out, and a dual-objective optimization method was applied to achieve the minimum value of the manufacturing cycle and the cost factors [15]. The importance of the algorithm’s results over other algorithms was demonstrated with exhaustive simulation results.

From the above research at home and abroad, the FWA and the neural network algorithm are widely used in various fields, and good results have been achieved. Most of the research has focused on the fields of medicine, architecture, and electricity, but few scholars have applied the neural network or fireworks algorithm to information management and competition of enterprises to solve the problem of enterprise information management. Hence, a novel enterprise information management evolution game system based on the FWA-BP network was devised to enable model-based management and enhance the overall operational efficiency of enterprise management.

3. Enterprise Game System Model Construction via FWA-BP Neural Network

3.1 System Reconstruction Preprocessing

The BP neural network algorithm has super extensive applications, and the fusion effect of the established model is good and received a lot of attention. The neural network fully mines the nonlinear relationship between the enterprise and its own information data and the target object according to the sample data, continuously adjusts the process network parameters through the back propagation algorithm, and integrates the knowledge in the trained model. Stored in the network parameters, the new input data are fused according to fusion knowledge to obtain the final output fusion result [16,17]. Because enterprise information data are voluminous, and management levels are complex with multiple dimensions, this study employed the principal component analysis technique for sample dimension reduction processing of enterprise information data to minimize information correlation and to subsequently reduce the number of neural network input nodes, thus simplifying the network structure [18]. Then, the firework algorithm was used to adjust and optimize the parameters involved in the neural network. This step can improve the network's calculation speed, boost the fidelity of the mold fusion network’s calculation speed, and boost mound fusion's fidelity. The fusion flow chart from constructing the model is shown in Fig. 1.

Fig. 1. Dimensionality reduction steps in the PCA method.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig1.png
Fig. 2. Evaluation model based on the BP neural network.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig2.png

The flowchart in Fig. 2 shows the specific dimension reduction steps of PCA. For the data collected by a total of $g$ information management systems within the same time period in $n$ units of time, a sample matrix, $X$, can be composed of $n$ rows and $g$ columns, and formula (1) transforms this sample.

(1)
$ \mu _{ij}=\frac{x_{ij}-\overline{x_{j}}}{s_{j}},i=1,2,\cdots ,n;j=1,2,\cdots ,g $

In formula (1), $\mu _{ij}$ represents the parameter matrix obtained after sample matrix transformation; $x_{ij}$ represents the sample; $\overline{x_{j}}$ is the average sample. $\overline{x_{j}}=\frac{1}{n}\sum _{i=1}^{n}x_{ij}$ and $s_{j}^{2}=\frac{1}{n-1}\sum _{i=1}^{n}\left(x_{ij}-\overline{x_{j}}\right)^{2}$ can get the standardized matrix, $U$. The solution of the correlation coefficient matrix of the standardized matrix is shown in formula (2):

(2)
$ R=\frac{U^{T}U}{n-1} $

In formula (2), $R$ and its characteristic equation is solved. See formula (3) for details:

(3)
$ \left| R-\lambda E_{m}\right| =0 $

In formula (3), $m$ represents the eigenvalues obtained when the matrix equation is solved. Then, the selected eigenvalues are sorted in descending order, and the feature vectors corresponding to the former proper values, $p$, are picked and combined into a transformation array. The expression for $A^{T}$ is shown in formula (4):

(4)
$ A^{T}=\left(u_{1},u_{2},\cdots ,up\right),p<m $

In formula (4), $u$ represents different eigenvectors. The value for $p$ is determined from $\frac{\sum _{j=1}^{P}\lambda _{j}}{\sum _{j=1}^{m}\lambda _{j}}\geq 0.85$. That is to say, when the contribution rate of the information of the previous principal component $p$ is significantly greater than 85%, the former principal component $p$ can be used as the eigenvalue of the sample. In the process, to eliminate the adverse effects of different data on the prediction fidelity of the mold, it is necessary to normalize the sample number. So min-max normalization is used to linearly change the original data and control the data between [0,-1]. The specific expression is formula (5):

(5)
$ x'=\frac{x-\min \left(x\right)}{\max \left(x\right)-\min \left(x\right)} $

In formula (5), $x$ deputizes the basic number; $\min (x)$ deputizes the least number, $\max \left(x\right)$ deputizes the max in the basic number; $x'$ represents the valid data after normalization. During the running of the model, 90% of the normalized sample data is stochastic chosen as the workout sample, and the leftover 10% is used as a sample to be tested.

3.2 Construction of the FWA-BP Enterprise Game Machinery Bottom

After improving the BP algorithm by using the principal component analysis method, the BP neural network model still has problems such as local extreme values and a slow convergence rate. In order to better optimize and adjust the training parameters, this study introduces FWA to optimize it and improve the computing speed and fusion accuracy of the network. FWA will constantly produce fireworks during operation of the system. Each firework is regarded as a solution to the system problem in the computing space, while the process of firework explosion is regarded as the process of constantly seeking the optimal solution.

Combining FWA and BP algorithms can effectively enhance the system's global search capabilities and makes more efficient use of parallel computing resources. Compared with other algorithms, implementation of the FWA-BP neural network algorithm is simpler.

The algorithm first determines the number of starting fireworks, $N$, and the dimensions of different fireworks, $n$. The upper and lower limits of the fireworks dimensions are then determined, and fireworks are randomly generated within that range. Then, the explosion operation is carried out based on the strategy [19-21]. Sparks generated in the explosion process are randomly distributed, and the number and range of sparks are mainly determined according to the detonation semidiameter, $A_{i}$, and the date of detonation spark-over,$S_{i}$. The specific expression is formula (6).

(6)
$ \left\{\begin{array}{l} A_{i}=E_{r}*\frac{f\left(x_{i}\right)-y_{\min }+\varepsilon }{\sum _{i=1}^{N}\left(f\left(x_{i}\right)-y_{\min }\right)+\varepsilon }\\ S_{i}=E_{n}*\frac{y_{\max }-f\left(x_{i}\right)+\varepsilon }{\sum _{i=1}^{N}\left(y_{\max }-f\left(x_{i}\right)\right)+\varepsilon } \end{array}\right. $

In formula (6), $f\left(x_{i}\right)$ represents the fitness value for each firework; $y_{\min }$ and $y_{\max }$ represent the worst-error and optimal-error values for solving the objective function; $E_{r}$ represents the parameter that controls the range of sparks; $E_{n}$ represents the parameter that controls the number of sparks; and $\varepsilon $ is the coefficient to avoid 0 in the denominator. The explosion operation, $D_{select}=rand*D$, is carried out for each firework, randomly determining the number of dimensions, $x_{i}$, that needs to be offset. formula (7) shows the explosion sparks generated:

(7)
$ ex_{ik}=x_{ik}+h $

In formula (7), $x_{ik}$ means firework $i$ is selected and needs to be offset; $k$ denotes the explosion spark generated after firework $i$ explodes; $h$ is the random value generated within the explosion radius. The specific expression is formula (8):

(8)
$ h=A_{i}\times rand\left(1,-1\right) $

The introduction of Gaussian sparks in the FWA increases the difference in the study plant or animal community. The figure of Gaussian sparks is defined as $g$. Fireworks are randomly selected from the initial fireworks population, and a dimension multiplied by random number $e$, which conforms to the Gaussian distribution, is defined as shown in formula (9):

(9)
$ \overset{\wedge }{x}_{i,k}=x_{i,k}\times e $

In formula (9), the range of $e$ belongs to the Gaussian scatter, $N\left(1,1\right)$. For sparks beyond the boundary generated during the explosion, mapping rules are used to constrain them, as shown in formula (10):

(10)
$ \overset{\wedge }{x}_{ik}=x_{LB,K}+\left| x_{ik}\right| -\left\{\left| x_{ik}\right| /\left(x_{UB,K}-x_{LB,K}\right)\right\}\times \left(x_{UB,K}-x_{LB,K}\right) $

In formula (10), $x_{ik}$ represents the boundary line that firework $i$ exceeds in dimension $k$. Then, the initial fireworks group, the explosion fireworks group, and the Gaussian sparks group together form the fireworks group, which is defined as $K$. The one with the best fitness performance is picked from the plant or animal community, and then, one is picked from the remaining fireworks groups according to the selection strategy, $N-1$. The $N$ selected fireworks are used as the fireworks population for the next iteration [22,23]. The strategy formula is (11):

(11)
$ \left\{\begin{array}{l} p\left(x_{i}\right)=\frac{R\left(x_{i}\right)}{\sum _{j\in K}R\left(x_{j}\right)}\\ R\left(x_{i}\right)=\sum _{j=1}^{K}d\left(x_{i}-x_{j}\right)=\sum _{j=1}^{K}\left\| x_{i}-x_{j}\right\| \end{array}\right. $

In formula (11), $R\left(x_{i}\right)$ represents the sum of the European distances between other fireworks and current fireworks $x_{i}$. The range for Euclidean distance is [0,1]. In the set of all candidates, $p\left(x_{i}\right)$ represents the probability that fireworks $x_{i}$ are selected. When calculating the fitness value of the fireworks population generated in each iteration, if the optimal fitness satisfies the maximum number of iterations, the search for the optimal solution ends; otherwise, it jumps to the first step of the process. Combining the fireworks algorithm with the neural network, a BP neural network is created, the fireworks population is initialized, and the goal of optimization is actually all the network parameters in the neural network. The fitness worth of every firework is computed according to the above steps of the FWA. The sum of squared error (SSE) of the neural network is taken as the fitness function, and the expression for $f\left(x_{i}\right)$ is shown in (12):

(12)
$ f\left(x_{i}\right)=SSE=\sum _{i=1}^{S}\left(t_{i}-y_{i}\right)^{2} $

In formula (12), $t$ represents the anticipated export value under the BP approach, the quantity of nerve cells in the neural net's export layer, and the actual export value under the BP calculation. According to the steps, optimization of the fireworks population and judgment on the termination conditions are carried out, and the optimal individual satisfying the conditions is decoded and assigned to the neural network. The LM algorithm is then used to train and optimize the neural network weights and thresholds with higher precision. The process involves setting the target error function to the SSE, and specifying the number of training iterations and target error beforehand, followed by initiating the training process. When the target error or the max quantity of iterations is achieved during training, it means the training is over, and establishment of the optimal model is completed. If the conditions are not met, the previous step is conducted to continue the operation.

The input variables of the sample data define the nodes in the net. There is no specific method to be sure of the number of nodes in the concealing layer. Therefore, a common empirical formula is selected, and the number of nodes is determined according to the change in the selected formula to carry out the experiment. The expression is shown in (13):

(13)
$ n'=\sqrt{m+p}+d $

In (13), $n'$ is the number of nodes in the hidden layer in the neural network, $m'$ represents the number of nodes in the input layer, and $p$ represents the quantity of nodes in the output layer. When network performance no longer changes with an increase in the quantity, it could mean the number of conceal nodes at this moment is the optimal quantity of conceal nodes. After that, all indicators would be preprocessed accordingly, as seen in Fig. 2.

Fig. 3. Flow chart of the information game based on the fusion model for the FWA-BP neural network algorithm.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig3.png

In Fig. 3, all indicators are preprocessed first, and then BP neural network detection is performed through a series of operations. The neural network deals with enterprise information, and thus, conducts games among the enterprises. Before fusion, it is necessary to clarify the relationship between the data and the target object, which is not a simple linear relationship. Therefore, it is necessary to add nonlinear factors to the activation function in the hidden layer to improve the nonlinear expression power of the mold. The hidden layer usually adopts a nonlinear sigmoid function, and the output layer adopts a linear purelin effect. When comparing the constructed models, two or more models are usually compared, and the same test samples are used to test the effect fusion. Most researchers select the mean absolute error (MAE) and the root mean square error (RMSE) as test indexes. The specific expression is formula (14):

(14)
$ \left\{\begin{array}{l} MAE=\left(\sum _{i=1}^{N}\left| t_{i}-y_{i}\right| \right)/N'\\ RMSE=\sqrt{\frac{1}{N}\sum _{i=1}^{N}\left(t_{i}-y_{i}\right)^{2}} \end{array}\right. $

In formula (14), $N'$ is the test date of the pattern, and $y$ is the predicted output date. Using the above formula to build the enterprise game system bottom process in the FWA-BP fusion model is revealed in Fig. 3.

4. Simulation Experiment and Performance Analysis

4.1 Cumulative Contribution of Principal Components and Sample Training

There are many variables in the enterprise information management system, and the data from these variables are collected by different game systems. There are strong correlations among the data. If they are directly input to the neural network without processing, it may increase the complexity of the network, which will eventually affect the accuracy of the model. Therefore, in the early stages of model construction, PCA should be performed on the obtained combined dataset. The princomp function in MATLAB was called to carry out PCA processing [24,25]. The integrated proffer speed of the most important module of the variables in the combination is shown in Fig. 4.

Fig. 4. Cumulative contributions of principal components.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig4.png

In Fig. 4, the dataset was reduced to six principal components that accounted for 94% of the variance. This dimensionality reduction helped streamline the enterprise information management game system and simplify the network structure, leading to the construction of the FWA-BP fusion model. To showcase the effectiveness of the game system developed in this study, the program was executed using a MATLAB 2017 environment. First, BP was structured as 6-12-1, with the number of input variables determined to be six due to panel data; the quantity of panel points in the hidden layer was set at 12 according to model configurations. The output layer variables represented the information management of game system effectiveness, hence requiring a single output neuron. The hidden layer activation function in BP utilized logsig, while the output layer selected the purelin effect. The main parameters of FWA-BP were set as follows: population size, $N=10$; adjustment constant of the radius of the firework explosion, $E_{r}=5$; adjustment constant of the quantity of sparks in the fireworks detonation, $E_{n}=30$; the quantity of Gaussian variation sparks, $g=10$; and the maximum number of iterations, $T=100$.

After using the FWA to update the coarse precision of BP, it was necessary to use the practicing record to further train the mold. The number of training sessions was 300; the training function selected was trainlm; the learning rate of the algorithm was expressed as 0.01, and the target error of training was set at 0.01. The specific training situation is shown in Fig. 5.

As shown in Fig. 5, the number of iterations for training the constructed model was set to 300 in advance, but in practice, the model mostly reached the target error after 110 iterations. This showed the FWA was effective for parameter optimization of the BP neural network.

Fig. 5. The training situation of the sample.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig5.png

4.2 Comparison of Output Effects, and Evaluation Scores of Different Models

Then, the LM algorithm was used to train the model with higher precision, which was conducive to the model jumping from within the local min, improving practice fidelity and speeding up the training. The model was then tested, and the results are shown in Fig. 6.

Fig. 6. Test results from the FWA-BP mold.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig6.png

Based on the FWA-BP construction experiment performed on the enterprise game system, the results depicted in Fig. 7 demonstrate that while prediction deviation for some samples exceeded 10, the remaining model predictions proved consistent with the actual expected value, the predicted curve largely aligning with the actual curve. Generally speaking, the closer the accessories degree is to 1, the better the fitting effect, and the fit was 0.9254 in the $R^{2}$ results. This showed that the constructed fusion model can well represent information management of the enterprise game system, and can also meet the system’s precision requirements. To verify the feasibility and superiority of the model, the evaluation results from the constructed model were compared with the improved particle swarm optimization algorithm BP (IPSO-BP) model and the improved genetic algorithm BP neural network (IGA-BP) model. The comparison is shown in Fig. 7.

Fig. 7. Comparison of the effects obtained from different models.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig7.png

From Fig. 7, observe that the RMSE predicted by the FWA-BP neural network model is approximately 7.775; The prediction result of the IGA-BP neural network is approximately 7.941; The IPSO-BP neural network predicted a result of approximately 7.928. The results of the constructed model are not significantly different from the IGA-BP neural network, slightly smaller than the IGA-BP neural network, but slightly larger than the predicted results obtained by IPSO-BP. The reason the model produces such results may be from outliers in the predicted output that deviate too much from the true value. And due to the squared processing of the model error, when encountering outlier data, the error will be amplified, resulting in poorer RMSE results. However, in the evaluation of MAE indicators, the indicator effect of the FWA-BP neural network is superior to the other two models, with a score of 6.608. The difference between the indicator score of the IPSO-BP neural network and the proposed model’s score is relatively small, while the score of the IGA-BP neural network indicator was relatively large.

4.3 Simulation Comparison of the Enterprise Information Management Game System Model

To further verify the accuracy and effectiveness of the proposed method, a fitness index was used for analysis and comparison. The fitness curves of the FWA and IPSO algorithms are plotted in Fig. 8.

Fig. 8. Comparison curve of fitness between FWA-BP and IPSO-BP.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig8.png

Fig. 8 compares the number of iterations at which the algorithms reached convergence. Between them, the FWA-BP curve reached convergence earlier, and the final value was small, whereas the IPSO-BP curve achieved convergence after 150 iterations and fell into a local minimum. The search for solutions stopped prematurely. However, FWA-BP continued to calculate downwards to find the optimal solution at the same time the number of iterations reached equilibrium. The comparison of the two phases showed that global optimization of the FWA-BP was stronger, and the best point can be found faster. This also showed that the research and construction model could quit the local optimum more easily in the process of calculation and operation, and accuracy was higher. Then, the degree of fit for the game system was analyzed, as shown in Fig. 9.

Fig. 9. Fitness curve of the game system model.
../../Resources/ieie/IEIESPC.2024.13.3.285/fig9.png

Fig. 9 shows that the error analogy result of the IGA-BP neural network was significantly higher than the FWA-BP algorithm: 2.612 versus 0.625, respectively. The maximum error value IGA-BP could reach was 4.3, which was relatively large; while the maximum error between the FWA-BP neural network and the target result was 0.01. Comparison of the two showed that FWA-BP had a very good fit, which can be applied to system construction of enterprise information management and can effectively ensure its stability and consistency. After passing the above tests, a construction enterprise was taken as an experimental object to test the performance of the information management system during construction of the enterprise. The duration of the test was 24 hours. The specific test is shown in Table 1.

Table 1. Data from the construction information management performance test of an enterprise.

Test run time (h)

CPU utilization rate (%)

System memory usage (%)

3

20.1

60.1

6

21.3.

59.9

9

20.9

61.1

12

19.9

60.3

15

20.1

60.5

18

20.2

60.0

21

21.1

60.6

24

20.2

60.8

Table 1 shows the system stability and use of the information management system of a construction enterprise. As time passed, the utilization rate of the CPU and system memory changed little. When the test time reached 6h, maximum CPU utilization was 21.3%; At 12h, minimum CPU utilization was 19.9%. At 24h, the system memory utilization rate had a maximum of 60.8%, but at 6h, system memory utilization had a minimum value of 59.9%. In the overall process, the difference between CPU and memory utilization rates of the built system was not more than ${\pm}$ 0.5, and stability can be guaranteed.

5. Conclusion

The traditional information management system has problems such as slow information management and an inability to share information resources easily, which cannot meet the requirements of enterprises and managers. To upgrade the comprehensive ability of enterprise word management, this research introduced the fireworks algorithm and the neural network algorithm and combined them into the FWA-BP algorithm. By evaluating the indicators of different models, the RMSE predicted by the FWA-BP model was about 7.775, which was between the forecast outcomes of BP and PSO-BP. In the evaluation of MAE indicators, the FWA-BP neural network model demonstrated performance superior to other models, with a score of 6.608. At the same time, comparing the fitness of the FWA-BP and PSO-BP algorithms, FWA-BP could achieve convergence faster, and the convergence value was small; the FWA-BP algorithm continued to reach a balance every time. Calculations were carried out until the optimal value was found. In addition, the simulation curve fitting of the proposed enterprise game system model showed that the error simulation of the FWA-BP algorithm was evidently below that of BP, and the max error between it and the target result was only 0.01. The performance test of the building model applied to a construction company showed that the overall CPU and memory utilization of the system was no more than ${\pm}$ 0.5, which was stable. This showed that the model constructed by the FWA-BP algorithm had better evaluation results for the enterprise game, could jump out of the local minimum more quickly and accurately, and achieved high accuracy. Furthermore, the other algorithms introduced in this process can prove useful in constructing enterprise game systems, and are of significant importance to the development of businesses within the country and throughout the global economy. However, the model built in the experiment needs to rely on a large amount of sample data for learning and training. Mining the connections for management of a large number of enterprises is a prerequisite for high integration accuracy. Therefore, the collected enterprise management information must be abundant and comprehensive. This is what subsequent experiments must focus on.

REFERENCES

1 
J. Gao, “Analysis of enterprise financial accounting information management from the perspective of big data”. International Journal of Science and Research (IJSR), Vol. 11, pp. 1272-1276, 2022.URL
2 
V. I. Osipov, N. A. Gorbunova, N. F. Kolesnik, E. G. Sarantseva, O. V. Shibileva, “Improving the Training of Experts in Economic Specialties in the Field of Enterprise Information Management”. Revista Geintec-Gestao Inovacao E Tecnologias, Vol. 11, pp. 1421-1430, 2021.URL
3 
L. Ying, X. Liu, M. Li, L. Sun, P. Xiu, & J. Yang, “How does intelligent manufacturing affects enterprise innovation”? The mediating role of organisational learning. Enterprise Information Systems, Vol. 16, pp. 630-667, 2022.DOI
4 
M. Zdravković, H. Panetto, G. Weichhart. “AI-enabled enterprise information systems for manufacturing”. Enterprise Information Systems, Vol. 16, pp. 668-720, 2022.DOI
5 
S. Chatterjee, A. K. Kar, S. Z. Mustafa, “Securing IoT devices in smart cities of India: from ethical and enterprise information system management perspective”. Enterprise Information Systems, Vol. 15, pp. 585-615, 2021.DOI
6 
M. Zhang, Y. Yuan, R. Wang, et al. “Recognition of mixture control chart patterns based on fusion feature reduction and fireworks algorithm-optimized MSVM”. Pattern Analysis and Applications, Vol. 23, pp. 15-26, 2020.DOI
7 
C. Gong, X. Wang, A. Gani, et al. “Enhanced long short-term memory with fireworks algorithm and mutation operator”. The Journal of Supercomputing, Vol. 77, pp. 12630-12646, 2021.DOI
8 
S. Sebbane, N. El Akchioui, “A Novel Hybrid Method Based on Fireworks Algorithm and Artificial Neural Network for Photovoltaic System Fault Diagnosis”. International Journal of Renewable Energy Research (IJRER), Vol. 12, pp. 239-247, 2022.DOI
9 
D. Manivannan, M. Kavitha, “A Diabetic Monitoring System using Learning Automata based Fireworks Algorithm and Dynamic Brain Storm Classifier”. International Journal of Intelligent Systems and Applications in Engineering, Vol. 10, pp. 213-225, 2022.URL
10 
W. Wang, K. Liu, C. Yang, et al. “Cyber physical energy optimization control design for PHEVs based on enhanced firework algorithm”. IEEE Transactions on Vehicular Technology, Vol. 70, pp. 282-291, 2020.DOI
11 
C. A. de Pinho Pinheiro, N. Nedjah, L. de Macedo Mourelle, “Detection and classification of pulmonary nodules using deep learning and swarm intelligence”. Multimedia Tools and Applications, 79, pp. 15437-15465, 2020.DOI
12 
O. Yılmaz, E. Bas, E. Egrioglu. “The training of Pi-Sigma artificial neural networks with differential evolution algorithm for forecasting”. Computational Economics, Vol. 59, pp. 1699-1711, 2022.DOI
13 
L. Wang, K. Jiang, T. Wei, et al. “Estimation of parameters of probability integral method model based on improved fireworks algorithm”. Survey Review, Vol. 53, pp. 366-382, 2021.DOI
14 
M. Zhang, R. Wang, Z. Cai, et al. “Phase partition and identification based on kernel entropy component analysis and multi-class support vector machines-fireworks algorithm for multi-phase batch process fault diagnosis”. Transactions of the Institute of Measurement and Control, Vol. 42, pp. 2324-2337, 2020.DOI
15 
A. M. Yadav, K. N. Tripathi, S. C. Sharma, “A bi-objective task scheduling approach in fog computing using hybrid fireworks algorithm”. The Journal of Supercomputing, Vol. 78, pp. 4236-4260, 2022.DOI
16 
Q. Ke, J. Siłka, M. Wieczorek, Z. Bai, & M. Woźniak, “Deep neural network heuristic hierarchization for cooperative intelligent transportation fleet management”. IEEE Transactions on Intelligent Transportation Systems, Vol. 23, pp. 16752-16762, 2022.DOI
17 
M. S. Jassim, G. Coskuner, M. Zontul, “Comparative performance analysis of support vector regression and artificial neural network for prediction of municipal solid waste generation”. Waste Management & Research, Vol. 40, pp. 195-204, 2022.DOI
18 
X. Yi, E. K. Latch, “Nonrandom missing data can bias principal component analysis inference of population genetic structure”. Molecular Ecology Resources, Vol. 22, pp. 602-611, 2022.URL
19 
L. Wang, K. Jiang, T. Wei, C. Jiang, J. Zhao, & S. Chi. “Estimation of parameters of probability integral method model based on improved fireworks algorithm”. Survey Review, Vol. 53, pp. 366-382, 2021.DOI
20 
C. Gong, X. Wang, A. Gani, & H. Qi, “Enhanced long short-term memory with fireworks algorithm and mutation operator”. The Journal of Supercomputing, Vol. 77, pp. 12630-12646, 2021.DOI
21 
S. Sebbane, and Nabil El Akchioui. “A Novel Hybrid Method Based on Fireworks Algorithm and Artificial Neural Network for Photovoltaic System Fault Diagnosis”. International Journal of Renewable Energy Research (IJRER), Vol. 12, pp. 239-247, 2022.DOI
22 
D. Manivannan, M. Kavitha. “A Diabetic Monitoring System using Learning Automata based Fireworks Algorithm and Dynamic Brain Storm Classifier”. International Journal of Intelligent Systems and Applications in Engineering, Vol. 10, pp. 213-225, 2022.URL
23 
X. D. Li, Y. Yang, and W. W. Li. “ANN model of preparation of energetic materials by spray drying”. Propellants, Explosives, Pyrotechnics, Vol. 46, pp. 1405-1414, 2021.DOI
24 
O. Yılmaz, B. Eren, and E. Erol. “The training of Pi-Sigma artificial neural networks with differential evolution algorithm for forecasting”. Computational Economics, Vol. 59, pp. 1699-1711, 2022.DOI
25 
S. Das, S. C. Nayak, B. Sahoo. “Towards crafting optimal functional link artificial neural networks with RAO algorithms for stock closing prices prediction”. Computational Economics, Vol. 60, pp. 1-23, 2022.DOI
Lianqin Zhu
../../Resources/ieie/IEIESPC.2024.13.3.285/au1.png

Lianqin Zhu majored in Mathematical Statistics at Hangzhou University in 1998 (bachelor) and Quantitative Economics at Business University in 2007 (master's degree). Presently, she is a lecturer in accounting at Business & Tourism Institute of Hangzhou Vocational and Technical College. Led or participated in multiple municipal level projects, and published multiple papers in China. Her areas of interest include economic modeling, big data analysis, and more.