Research and Application of Improved Multiple Imputation Based on R Language in Fire Prediction

: An improved multiple imputation based on R language is proposed to deal with the miss of data in a ﬁre prediction model, which can affect the accuracy of the prediction results. Hazard and operability (HAZOP) is used to accurately ﬁnd the data related to the research purpose, and exclude data with a missing rate greater than 80% and small differences in characteristics. Then, by changing the m value in the mice package under the R language (R-mice), the relevant parameters of the complete ﬁlling factor set under different m values are obtained. The value of m is determined after observing and comparing the parameters. The proposed method fully considers the randomness of ﬁlling and the difference between the generated dataset. Taking Hubei Province as an example, the data processed by this method are used as the input of the Bayesian network, and the ﬁre trend is used as the output. The results show that the improved multiple imputation based on R-mice can solve the problem of missing data very well, and have a high prediction effect (AUC = 94.0800). In addition, the results of the predictive reasoning and sensitivity analysis show that the government’s supervision has a vital inﬂuence on the trend of ﬁres in Hubei Province.


Introduction
Missing data are a common problem in research, and the main reasons for the deletion are measurement errors, data corruption, and equipment failure [1]. The treatment of missing data can be roughly divided into three categories: not processing, direct deletion, and filling. Among them, non-processing refers to the direct application of incomplete data by machine learning such as Bayesian networks [2] and artificial neural networks [3], they are widely employed to estimate the fire risk of human casualties [4]. However, the requirements for operators are high and the error is large. Direct deletion refers to the deletion of data objects or attributes that contain missing values, but the deletion method applies to completely randomly missing datasets with a missing percentage of less than 5%. Otherwise, the loss of critical information affects the results of the study [5].
The filler method is increasingly recognized by scholars as it preserves key data and obtains more accurate research results [6]. The main methods are the maximum expected value algorithm (EM) [7], regression fill [8], cluster fill, and multiple fill [9] etc., and methods of improvement based on their various theories (e.g., KEMI [10], MIDA [11] etc.). Among them, the multiple-fill method is the most commonly used [12], which can fill the missing data with the appropriate statistical methods according to the pattern of data loss, which not only ensures the correlation between the variables but also effectively solves the uncertainty of data loss [13]. However, most of its applications are used in clinical studies [14], and a few scholars have applied it to the field of fire prediction [15,16]. Most of the ways to deal with missing data in the fire field are direct deletion, such as Liu deletes data with missing key features [17]; Jin et al., retain data containing complete features [18]; and Sattari et al., remove cases where the information required for classification was missing [19]. However, with the rapid development of economy and society, although the risk of fire is decreasing year by year and tends to a stable state, the base is still large and cannot be underestimated. Fire situation assessment and prediction play an important role in reducing its risk, which requires that the error of the research results is as small as possible, and the deletion method may cause larger errors.
The coding of the R-based multi-fill method is detailed in Section 2.2, where the m value defaults to 5 [20], which can be taken as 3 to 5 [21], 5 to 10 [22], or 5 to 20, or even higher [23]; the larger the value, the better, but the size of the calculation amount should also be considered. Therefore, the value of m needs to be discussed on a case-by-case basis, and should not be too large or too small, which may make the filling inaccurate and affect the research results.
To solve the above problems, our work proposes an improved multi-imputation based on the R language. In recent years, the relevant national security departments and enterprises have accumulated a large amount of fire data. Therefore, firstly, the Hazard and Operability Study (HAZOP) is used to qualitatively filter out the categories of data that best affected the trend of fire occurrence from the fire statistics, thereby reducing noise, and then the improved multi-fill was applied. After filling it, it is used as the input of the prediction model and the fire occurrence trend is used as the output. The output can provide a basis for decision-makers to formulate countermeasures to reduce the adverse effects of fires. The idea of this article is shown in Figure 1. the uncertainty of data loss [13]. However, most of its applications are used in clinical studies [14], and a few scholars have applied it to the field of fire prediction [15,16]. Most of the ways to deal with missing data in the fire field are direct deletion, such as Liu deletes data with missing key features [17]; Jin et al., retain data containing complete features [18]; and Sattari et al., remove cases where the information required for classification was missing [19]. However, with the rapid development of economy and society, although the risk of fire is decreasing year by year and tends to a stable state, the base is still large and cannot be underestimated. Fire situation assessment and prediction play an important role in reducing its risk, which requires that the error of the research results is as small as possible, and the deletion method may cause larger errors. The coding of the R-based multi-fill method is detailed in Section 2.2, where the m value defaults to 5 [20], which can be taken as 3 to 5 [21], 5 to 10 [22], or 5 to 20, or even higher [23]; the larger the value, the better, but the size of the calculation amount should also be considered. Therefore, the value of m needs to be discussed on a case-by-case basis, and should not be too large or too small, which may make the filling inaccurate and affect the research results.
To solve the above problems, our work proposes an improved multi-imputation based on the R language. In recent years, the relevant national security departments and enterprises have accumulated a large amount of fire data. Therefore, firstly, the Hazard and Operability Study (HAZOP) is used to qualitatively filter out the categories of data that best affected the trend of fire occurrence from the fire statistics, thereby reducing noise, and then the improved multi-fill was applied. After filling it, it is used as the input of the prediction model and the fire occurrence trend is used as the output. The output can provide a basis for decision-makers to formulate countermeasures to reduce the adverse effects of fires. The idea of this article is shown in Figure 1.

Hazard and Operability (HAZOP)
HAZOP is a qualitative safety evaluation method, guided by keywords (e.g., more, less, part of, no, as well as, reserve, other than) [23], combined with the process parameters (e.g., temperature, pressure, etc.) to find out the deviation of the process state, and then analyze the causes and consequences of the deviation, and finally formulate the prevention and control measures [24], as shown in Figure 2.
The HAZOP analysis is conducted in the form of expert discussion. Before the analysis, the expert decomposes the system into several units according to the research purposes. Each unit has its related functional requirements. There are one or more related parameters, and each parameter corresponds to several guiding words. Select a unit to test the combination of parameters and guiding words that is a deviation. If there is a deviation, analyze its causes and consequences, conduct a risk assessment, and propose the measures to eliminate and control the dangers and reduce losses. After the analysis of

Hazard and Operability (HAZOP)
HAZOP is a qualitative safety evaluation method, guided by keywords (e.g., more, less, part of, no, as well as, reserve, other than) [23], combined with the process parameters (e.g., temperature, pressure, etc.) to find out the deviation of the process state, and then analyze the causes and consequences of the deviation, and finally formulate the prevention and control measures [24], as shown in Figure 2.

Multiple Fills
The multiple-fill method is based on Bayesian estimation theory, and the fills of the missing data are random and all derived from observations. The specific filling steps can be divided into three steps:  The HAZOP analysis is conducted in the form of expert discussion. Before the analysis, the expert decomposes the system into several units according to the research purposes. Each unit has its related functional requirements. There are one or more related parameters, and each parameter corresponds to several guiding words. Select a unit to test the combination of parameters and guiding words that is a deviation. If there is a deviation, analyze its causes and consequences, conduct a risk assessment, and propose the measures to eliminate and control the dangers and reduce losses. After the analysis of all the possible deviations of one unit is completed, transfer to another unit, and repeat the above steps until the analysis of all units is completed.

Multiple Fills
The multiple-fill method is based on Bayesian estimation theory, and the fills of the missing data are random and all derived from observations. The specific filling steps can be divided into three steps:

1.
Choose the appropriate padding method based on the data deletion pattern: monotonic or arbitrary [25], such as predictive mean matching, logistic regression imputation, and professional odds models, m-fill missing data to generate m-group complete datasets.
The Rubin method was used to integrate all the analysis results to produce a final inference.
Many pieces of software can implement multiple fillings, such as SOLAS, SAS, and R languages, among others [13]. Compared with other software, R is simple to operate, easy to understand and effective, and is more suitable for scholars to apply it. The steps to implement multiple fillings using the R language and their codes are [26,27]: library(mice) # Call into the function package x 1 <-read.spss("data location/file name.sav", to.data.frame = T) # Import data x 2 <mice(x 1 , m, meth) # Fill in the data fit <-with(x 2 , analysis) # Contains m individual statistical analysis results, analysis is used to set the statistical analysis method applied to m-filled datasets pooled <-pool(fit) # Contains m statistical analysis of the average results summary(pooled) # Summary

Bayesian Networks
In solving the practical problems, inferences and judgments need to be made from uncertain knowledge and information, and Bayesian networks (BN) can use a probability theory to deal with the uncertainties caused by conditional correlation [28]. Its network topology is a directed acyclic graph (DAG). The nodes in BN consist of the random variables X = {X 1 , X 2 , . . . , X n }, non-conditionally independent variables are connected by arrows, and the parent node points to the child node [29]. It has the flexibility to not only learn network structures and parameters through large amounts of data [30], but also to use expert knowledge and data to improve the performance of the model, thus helping us analyze some complex problems (uncertain or missing data) [31], and even support us in developing measures [32].
Let BN = (G, θ), G is the network structure, and θ is the network parameter. The joint probability distribution on X is Equation (1) [33][34][35]: where Pa(X i ) is all parent nodes of X i . For arbitrary variables, the joint probability distribution can also be expressed as Equation (2): Fire 2023, 6, 235 4 of 13 A simple Bayesian network: a person with a weak security awareness may be unconsciously littering unextinguished cigarette butts, and if there is combustible material around them, it is highly likely to cause a fire. If the fire is not detected and controlled in time, it will grow larger and quickly spread to the surrounding area, and then improper operation or untimely escape will cause injury or even death; in some cases, such as in Figure 3a, the network structure and condition probability table can be obtained in the input software, as shown in Figure 3b. Its joint probability distribution is P(X) = P(X A ) P(X C |X A ) P(X D |X A ) P(X B ) P(X E |X B , X C ) P(X D |X A , X E ). Change either condition, and the probability of being late will change accordingly. For example, when a person has an extremely weak security awareness (True = 100%, Figure 3c), the probability of injury or death will increase from 42% to 68%. Conversely, when a person has a good sense of security (False = 100%, Figure 3d), the probability of injury or death will reduce from 42% to 17%.
where Pa(Xi) is all parent nodes of Xi. For arbitrary variables, the joint probability distribution can also be expressed as Equation (2): A simple Bayesian network: a person with a weak security awareness may be unconsciously littering unextinguished cigarette butts, and if there is combustible material around them, it is highly likely to cause a fire. If the fire is not detected and controlled in time, it will grow larger and quickly spread to the surrounding area, and then improper operation or untimely escape will cause injury or even death; in some cases, such as in Figure 3a, the network structure and condition probability table can be obtained in the input software, as shown in Figure 3b. Its joint probability distribution is P(X) = P(XA) P(XC|XA) P(XD|XA) P(XB) P(XE|XB, XC) P(XD|XA, XE). Change either condition, and the probability of being late will change accordingly. For example, when a person has an extremely weak security awareness (True = 100%, Figure 3c), the probability of injury or death will increase from 42% to 68%. Conversely, when a person has a good sense of security (False = 100%, Figure 3d), the probability of injury or death will reduce from 42% to 17%.

Instance Calculations
Hubei Province is located in the east of China, with 17 cities and prefectures in total. This article will use an improved multiple imputation based on R-mice to construct a prediction model to predict fire trends x0 and provide a basis for decision-makers to formulate fire prevention measures. The specific analysis steps are as follows.

Instance Calculations
Hubei Province is located in the east of China, with 17 cities and prefectures in total. This article will use an improved multiple imputation based on R-mice to construct a prediction model to predict fire trends x 0 and provide a basis for decision-makers to formulate fire prevention measures. The specific analysis steps are as follows.

Collection and Processing of Fire Statistics
Through field investigation and consulting the China Fire Rescue Yearbook, we have obtained data on various fire scenarios in Hubei Province from 2016 to 2018, which can be roughly divided into the following 5 categories: fire loss, fire information, fire rescue, fire brigade, and fire equipment [36]. However, not all of them are what we need. Exclude some scenario data that have little influence on the research purposes, which can reduce the complexity and noise of the model, reduce the amount of calculation, and increase the readability and accuracy of the model [37]. Last but not least, all data need to be normalized to facilitate the subsequent analysis [38], as shown in Equation (3): The role of HAZOP in this article is to screen out the data we need. The advantage of the HAZOP method is simplicity and ease of operation. Experts from various fields gather together to express their opinions, influence, and inspire each other and discover more problems creatively. The specific process of using HAZOP to screen data is shown in Figure 4, that is, screen out some data irrelevant to the research purpose and data with unobvious characteristics through Condition 1 and Condition 2. The most important thing is in the analysis process, the primary and secondary indicators generated can construct a prediction model structure about the fire trend through correlation analysis, and the deviation can be used as the state of the model node, as shown in Table 1.
Through field investigation and consulting the China Fire Rescue Yearbook, we have obtained data on various fire scenarios in Hubei Province from 2016 to 2018, which can be roughly divided into the following 5 categories: fire loss, fire information, fire rescue, fire brigade, and fire equipment [36]. However, not all of them are what we need. Exclude some scenario data that have little influence on the research purposes, which can reduce the complexity and noise of the model, reduce the amount of calculation, and increase the readability and accuracy of the model [37]. Last but not least, all data need to be normalized to facilitate the subsequent analysis [38], as shown in Equation (3): The role of HAZOP in this article is to screen out the data we need. The advantage of the HAZOP method is simplicity and ease of operation. Experts from various fields gather together to express their opinions, influence, and inspire each other and discover more problems creatively. The specific process of using HAZOP to screen data is shown in Figure 4, that is, screen out some data irrelevant to the research purpose and data with unobvious characteristics through Condition 1 and Condition 2. The most important thing is in the analysis process, the primary and secondary indicators generated can construct a prediction model structure about the fire trend through correlation analysis, and the deviation can be used as the state of the model node, as shown in Table 1.

Fill Scenario Data
A total of 51 cases were obtained, with 1683 data and 312 missing data, with a missing rate of 18.54%. When the missing rate is about 20%, compared with the deletion method, mean (mode) method, and regression method, the R-mice filling effect is the best [11]. The specific steps are shown in Section 2.2. In this section, the value of m is determined and a complete dataset is obtained.
First, calculate the total variance t (Equation (4)) of the complete factor set x 1 , x 2 , x 9~x15 , x 17~x20 , and x 22 filled under different values of m through code 1 and code 2. The above value of t is smaller the better. The results are shown in Figure 5. The t value of each complete factor set under eleven different m values is presented in the form of a scatter plot. As the m value increases, the t value continues to decrease. Observing the black dashed line in Figure 5, it can be seen that when m ≥ 8, the change in the t value is small, or even no longer changes. That is, when m > 8, the value of m has little effect on the value of t. Therefore, the value of m is 8 in this work. code 1: data <-mice(x, m = I, maxit = 100, seed = 2020), i = 1, 2, . . . , n code 2: fit <-with(data, lm (x 0~x1 + x 2 + . . . + x 21 + x 22 )) where m refers to the number of interpolated datasets, maxit refers to the number of iterations, seed refers to the number of set seeds, and u refers to the mean squared error (MSE). Then, use code 3 to find eight complete datasets, use code 4 to analyze them, and find the residual standard error (on 57 degrees of freedom), the multiple R-squared, and the F statistic value. It is not difficult to see from the three dashed lines in Figure 6 that the parameters of the fourth dataset are the best in all datasets. Among them, dataset 9 is the dataset with the best parameters when m is 5. Therefore, this article chooses the fourth dataset as the input of the predictive model. Moreover, the results show that determining the value of m by the above method can make the filled data better fit the real data. The impact on the forecast results will be analyzed below. code 3: data y <-complete(data, y), y = 1, 2, 3, …, 8 code 4: fit y <-with(data y, lm (x0~x1 + x2 + … + x21 + x22)) where the code 3 is to fill in one of the eight filling datasets to the vacant position, and finally form a complete dataset.  Then, use code 3 to find eight complete datasets, use code 4 to analyze them, and find the residual standard error (on 57 degrees of freedom), the multiple R-squared, and the F statistic value. It is not difficult to see from the three dashed lines in Figure 6 that the parameters of the fourth dataset are the best in all datasets. Among them, dataset 9 is the dataset with the best parameters when m is 5. Therefore, this article chooses the fourth dataset as the input of the predictive model. Moreover, the results show that determining the value of m by the above method can make the filled data better fit the real data. The impact on the forecast results will be analyzed below. code 3: data y <-complete(data, y), y = 1, 2, 3, . . . , 8 code 4: fit y <-with(data y, lm (x 0~x1 + x 2 + . . . + x 21 + x 22 )) where the code 3 is to fill in one of the eight filling datasets to the vacant position, and finally form a complete dataset. dataset as the input of the predictive model. Moreover, the results show that determining the value of m by the above method can make the filled data better fit the real data. The impact on the forecast results will be analyzed below. code 3: data y <-complete(data, y), y = 1, 2, 3, …, 8 code 4: fit y <-with(data y, lm (x0~x1 + x2 + … + x21 + x22)) where the code 3 is to fill in one of the eight filling datasets to the vacant position, and finally form a complete dataset.  Based on the filled complete dataset and expert knowledge, the structure of the prediction model is learned, and then the dataset is learned according to the training ratio of 8:2, and the final fire occurrence trend prediction and prediction model is obtained as shown in Figure 7. Based on the filled complete dataset and expert knowledge, the structure of the prediction model is learned, and then the dataset is learned according to the training ratio of 8:2, and the final fire occurrence trend prediction and prediction model is obtained as shown in Figure 7.

Results and Discussion
This section considers the predictive reasoning of the Bayesian network, that is, forward propagation analysis [28,39]. By comprehensively using the fire scenario data and expert knowledge of Hubei Province in 2018, it is input into the prediction model to predict the trend of fire occurrence in Hubei Province in 2019, and compared with actual data to verify the accuracy of the model. This paper uses the area under the Receiver Operating Characteristic (ROC) curve, that is, the AUC value, to test the prediction effect. Among them, the AUC standard for judging the pros and cons of the prediction model is shown in Table 2 [40,41]. The classification threshold is calculated through the Youden index (Equation (4)). When the predicted probability is greater than the threshold, it is considered to happen; otherwise, it is considered not to happen [42].

Results and Discussion
This section considers the predictive reasoning of the Bayesian network, that is, forward propagation analysis [28,39]. By comprehensively using the fire scenario data and expert knowledge of Hubei Province in 2018, it is input into the prediction model to predict the trend of fire occurrence in Hubei Province in 2019, and compared with actual data to verify the accuracy of the model. This paper uses the area under the Receiver Operating Characteristic (ROC) curve, that is, the AUC value, to test the prediction effect. Among them, the AUC standard for judging the pros and cons of the prediction model is shown in Table 2 [40,41]. The classification threshold is calculated through the Youden index (Equation (4)). When the predicted probability is greater than the threshold, it is considered to happen; otherwise, it is considered not to happen [42].

Predictive Reasoning
Through the controlled variable method, three other sets of prediction models are constructed, as shown in Table 3, in which feature screening is to screen out features that have greater weight on the predictor variables from the original data, and the methods include filtering, wrapping, and embedded; this article uses the filtering, because it and HAZOP both screen the data before building the predictive model. According to the analysis and processing of the screening method and m value, the status of each scenario data of each city in Hubei Province in 2018 is obtained then input into the prediction model to find the fire occurrence trend in 2019. The prediction effect is tested using the ROC curve and AUC, as shown in Figure 8. include filtering, wrapping, and embedded; this article uses the filtering, beca HAZOP both screen the data before building the predictive model. According to ysis and processing of the screening method and m value, the status of each scen of each city in Hubei Province in 2018 is obtained then input into the prediction find the fire occurrence trend in 2019. The prediction effect is tested using the R and AUC, as shown in Figure 8. HAZOP m When m is 5 (the default and the most used), it is better to use HAZOP (3) the data compared with no screening (1) and using the filtering method (2). amount of calculation required is relatively small. Second, it can not only screen tive data, but also qualitative data, and the operation difficulty is relatively low. important thing is that part of the predictive model construction is complet screening process, that is, the state of the node, which greatly reduces the time c and improves the efficiency of screening. Finally, the forecasting effect is better > 78.59072 > 77.44641). Therefore, HAZOP is more suitable for data screening.
Compared with m = 5 based on the HAZOP method (3), the prediction effe based on the HAZOP method (4) is better (94.08003 > 82.52844). Although the ferences in other aspects, they are negligible and can be ignored. The results improved multiple imputation based on R language can be used for data proc fore constructing a prediction model, and it has a good prediction effect (AUC = Taking full account of the geographical location, population density, and tribution of each city in Hubei Province, combined with the prediction results in a fire risk map of Hubei Province is finally generated, as shown in Figure 9. It ca  When m is 5 (the default and the most used), it is better to use HAZOP (3) to screen the data compared with no screening (1) and using the filtering method (2). First, the amount of calculation required is relatively small. Second, it can not only screen quantitative data, but also qualitative data, and the operation difficulty is relatively low. The most important thing is that part of the predictive model construction is completed in the screening process, that is, the state of the node, which greatly reduces the time consumed and improves the efficiency of screening. Finally, the forecasting effect is better (82.52843 > 78.59072 > 77.44641). Therefore, HAZOP is more suitable for data screening.
Compared with m = 5 based on the HAZOP method (3), the prediction effect of m = 8 based on the HAZOP method (4) is better (94.08003 > 82.52844). Although there are differences in other aspects, they are negligible and can be ignored. The results show that improved multiple imputation based on R language can be used for data processing before constructing a prediction model, and it has a good prediction effect (AUC = 94.0800).
Taking full account of the geographical location, population density, and GDP distribution of each city in Hubei Province, combined with the prediction results in Figure 9, a fire risk map of Hubei Province is finally generated, as shown in Figure 9. It can be seen from Figure 9 that the eastern part of Hubei Province is economically developed, although the population is dense, the fire risk is low; the western region is relatively backward in the economy and sparsely populated, but the fire risk is high, especially in the southwest.

of 13
The reason for this phenomenon may be that compared with the eastern region, the western region has a higher terrain, mostly mountainous areas, and inconvenient transportation. In addition, the processing industry is relatively weak, the development of the high-precision tertiary industry is relatively lagging, the degree of industrialization is low, and the economic relevance is low. This has led to the relatively slow economic development of the western region and the inability to provide sufficient funds and talents for the development of fire rescue services. It is directly manifested in the lack of firefighting funds, weak government supervision, insufficient propaganda, and inadequate facilities and equipment.

Sensitivity Analysis
Sensitivity analysis refers to the study of the impact degree of changes in one node in the prediction model on other nodes from the perspective of quantitative analysis, which can identify errors in network structure or CPT [28,43], and provide the basis for risk analysis and control measures. In this paper, entropy reduction (Equation (6)) is used to quantify the sensitivity [44], and the nodes were sorted according to the impact of the results on x0 nodes. The results of the sensitivity analysis are shown in Table 4.
where H(Y) is the entropy and H(Y|X) is the conditional entropy. As shown in Table 4, y5 is the most significant variable, leading to the maximum entropy reduction of x0. y1 and y4 also have a large impact, with entropy reduction of more than 20%. However, the influence of y3 and y2 is relatively small, and the entropy reduction is less than 10%, so they are not suitable to be used as control variables in the prediction. Therefore, y5, y1, and y4 can be used as control variables. The reason for this phenomenon may be that compared with the eastern region, the western region has a higher terrain, mostly mountainous areas, and inconvenient transportation. In addition, the processing industry is relatively weak, the development of the high-precision tertiary industry is relatively lagging, the degree of industrialization is low, and the economic relevance is low. This has led to the relatively slow economic development of the western region and the inability to provide sufficient funds and talents for the development of fire rescue services. It is directly manifested in the lack of firefighting funds, weak government supervision, insufficient propaganda, and inadequate facilities and equipment.

Sensitivity Analysis
Sensitivity analysis refers to the study of the impact degree of changes in one node in the prediction model on other nodes from the perspective of quantitative analysis, which can identify errors in network structure or CPT [28,43], and provide the basis for risk analysis and control measures. In this paper, entropy reduction (Equation (6)) is used to quantify the sensitivity [44], and the nodes were sorted according to the impact of the results on x 0 nodes. The results of the sensitivity analysis are shown in Table 4.
where H(Y) is the entropy and H(Y|X) is the conditional entropy. As shown in Table 4, y 5 is the most significant variable, leading to the maximum entropy reduction of x 0 . y 1 and y 4 also have a large impact, with entropy reduction of more than 20%. However, the influence of y 3 and y 2 is relatively small, and the entropy reduction is less than 10%, so they are not suitable to be used as control variables in the prediction. Therefore, y 5 , y 1 , and y 4 can be used as control variables.
Compared to reality, the result of sensitivity is reasonable. Because if the government's supervision is insufficient, fire hazards cannot be managed in time, firefighting funds cannot be fully utilized, firefighting equipment and facilities will lag behind more and more complex fire situations, and even people's safety awareness will become weak, leading to urban fire risks. The rapid increase cannot be controlled by the emergency rescue capability alone, that is, it cannot improve the state of the target node alone. Because of the importance of supervision and management, the relevant departments should attach great importance to the development and soundness of efficient and resilient supervision mechanisms and systems.

Measures and Suggestions
The western region should develop processing industries, promote the development of leading industries from resource exporting to processing industries, and increase the added value of products; develop a tertiary industry with a high rate of return, increase the degree of relevance to surrounding industries, cultivate new economic growth points, and drive the development of other industries; and take advantage of talents, develop high-tech industries and emerging industries, and accelerate the pace of transforming scientific and technological achievements into productivity, so that there are surplus funds to develop the firefighting and rescue business, to meet the growing safety needs of the people.
Among them, cities of F, O, K, D, M, E, and L should strengthen supervision, promptly discover and supervise the elimination of fire hazards (especially electrical hazards), and archive major hazards and conduct regular inspections; timely check, repair, update, upgrade, and add rescue equipment to ensure that they play their due role in the disaster; and strengthen the publicity, innovative publicity ways, expand the scope of publicity, make fire prevention propaganda ubiquitous, and subtly improve people's safety awareness and skills. While strengthening supervision and propaganda, other cities also need to improve the legal system for fires, increase penalties for deliberate arson, and restrict them.

Conclusions
Aiming at the fire prediction model of firefighters, they can accurately obtain the relevant scenario data, predict the fire trend of the next year for the first time, and provide a quantitative basis for effective control measures to reduce the number of fires. This paper constructs a predictive model, analyzes, verifies, and applies the predictive model, and the results show that the model can predict the trend of fire occurrence well. It can not only propose specific measures for a certain city, but also provide a strong basis for a province to formulate accurate and effective fire protection planning, thereby reducing the number of