The impact of supply chain management on a company’s operation and decision based on the multidimensional data analysis of upstream and downstream industry market states

This study provides data analysis support for the entire enterprise procurement management process, thereby improving the management effectiveness of supply chain operations. It analyzes upstream and downstream industry market status data in the supply chain and various primary data in enterprise management activities. By utilizing the Delphi method to screen and verify multimode market status data indicators, which significantly impact upstream and downstream industries in multiple rounds, 28 types of market status data were selected for analysis. This analysis aimed to investigate the effect of supply chain management on operational decisions within the company. The data reduction method based on adaptive statistics was the most effective in revealing the market status and promoting efficient operation decision-making based on supply chain management. This study also suggests a brand-new technique for measuring supply chain performance based on the Levenberg-Marquardt Back Propagation (LMBP) algorithm, offering a more impartial manner of doing so. The performance evaluation results showed a maximum error level of less than 0.4% when paired with empirical analysis. The proposed optimization model provides strategic guidance for optimizing supply chain management and improving overall performance.


INTRODUCTION
The competition in the 21st century has changed from the match between enterprises to the competition between supply chains (Deng et al., 2020). As the transformation progresses, the company will use multidimensional data analysis to assess the state of the upstream and downstream industrial markets to manage the supply chain more efficiently. This will effectively aid the business in operational decisions and ensure the implementation of cost-cutting and efficiency-boosting measures (Zhang et al., 2021).
More businesses are beginning to value the impact of multidimensional data analysis on the market status of upstream and downstream industries, use data analysis techniques How to cite this article Ding J. 2023. The impact of supply chain management on a company's operation and decision based on the multidimensional data analysis of upstream and downstream industry market states. PeerJ Comput. Sci. 9:e1369 http://doi.org/10.7717/peerjcs.1369 to direct production and operations, and accomplish scientific decision-making as a result of the wave of data analysis (Rudskoi & Baurova, 2019). The research shows that the analysis based on the multidimensional data analysis of upstream and downstream industry market status plays an essential role in business understanding, performance measurement, improving customer relations, and creating business opportunities (Li & Li, 2016). For example, the management knowledge mined through data analysis can help enterprises identify the key factors that affect product quality in the supply chain, summarize the causal relationship between product quality and production processes, and thus avoid potential risks (Qu, Liu & Bao, 2021). It can also help enterprises understand customer needs, strengthen customer loyalty, maintain a good relationship with customers, and solve problems before crises (Mishra et al., 2018). In brief, adept utilization of practical operational data analysis and management knowledge mining can facilitate the acquisition of actionable, evidence-based insights and knowledge from internal and external operational data sources. This, in turn, empowers enterprises to make prompt and informed decisions, allowing managers to understand their organization's standing better, consequently enhancing work efficiency and uncovering novel business prospects (Bian, Shang & Zhang, 2016).
Many scholars have studied the positive impact of multimodal joint data analysis on promoting the company's operational decisions. Fan, Lau & Zhao (2015) empirically discussed the role of data analysis capability in group strategy, culture, and financial control by establishing a structural equation model. The study's findings indicate that data analysis resource integration capabilities most influence group strategy, culture, and financial control. Kiseľáková et al. (2018) highlighted the moderating impact of enterprise strategic orientation while examining the relationship between data analysis and enterprise performance; specifically, businesses should build a data-driven strategy from a macro viewpoint to receive twice the outcome with half the effort. When exploring the relationship between data analysis and enterprise performance, the intermediary role of dynamic capabilities was emphasized (Perfilieva et al., 2016). Making decisions about how a company operates is a key contributor to its performance and a crucial safeguard for its competitiveness. Wang, Wang & Huang (2017) pointed out in his research that the reduction of costs and the improvement of customer satisfaction are primarily due to enterprises' efficient operation decision-making ability. Kościelniak & Puto (2015) has also concluded that operational decision-making can promote an enterprise's customer service capabilities, increase customer service satisfaction, and thus improve other performance. Flynn, Koufteros & Lu (2016)'s research confirmed the conclusion that the external integration of the supply chain (customer integration and supplier integration) positively affects performance. Taking Chinese manufacturing enterprises as the survey object confirmed the intermediary role of supply chain integration in relationship and operation performance. Many scholars focus on different aspects of evaluating the performance of enterprise operations and decision-making. Cost, flexibility, service, quality, and other indicators are used to analyze and assess the enterprise operation performance. For example, Centobelli, Cerchione & Esposito (2018) used service level, delivery capacity, and financial status to evaluate the performance of enterprise operation decisions. Truong et al. (2017) categorized operational performance into cost, flexibility, and distribution for evaluation. Belekoukias, Garza-Reyes & Kumar (2014) argues that operational performance is manifested through the rapid response ability, production cost, delivery cycle, and customer satisfaction of enterprises. From this perspective, Belekoukias, Garza-Reyes & Kumar (2014) discusses the key indicators that enable enterprises to gain a competitive edge in the market. Rompho (2018) evaluated enterprise operational decision-making performance using four distinct criteria: new products, services, quality, and cost.
Most scholars have recognized the positive role of data analysis capability in improving enterprise performance, and data analysis is also considered an essential driver of innovation (Nasution, Fahmi & Prayogi, 2020). However, most of these studies focus on direct effects, and there are few kinds of literature discussing the role of other factors in the promotion path, let alone empirical research. The interpretive structure model is used in this study to decompose ten hierarchical relationships among 28 different types of market data to aid cross-border e-commerce decision-makers in understanding the rational relationships among indicators and building a valuable and efficient supply chain system. The LMBP feedback neural network model is used to automatically update the weight of each indicator and evaluate cross-border e-commerce enterprises.

RELATED RESEARCH WORKS
Because more competitors and clients have a more comprehensive range of needs in the upstream and downstream industrial marketplaces, businesses are forced to make more operational decisions due to the growing product diversity (Wu et al., 2021). To inform their operations and decision-making, businesses must employ scientific methods to study the upstream and downstream industrial markets and judiciously integrate multidimensional data analysis of the upstream and downstream industrial market status (Gardi et al., 2021). Nhemachena & Murimbika (2018) confirmed that data analysis capability could positively affect enterprise performance by improving supply chain management capability and company reputation. Chowdhury & Quaddus (2017) improved the value of the supply chain through data analysis, researched the main influencing factors of the multimodal joint data of market status, and proposed the dynamic capability theory. Kaur et al. (2020) established a procurement model based on the supply chain to promote data analysis on the whole supply chain process. Wang et al. (2018) analyzed the crisis and opportunities faced by traditional supply chain cost control methods in the e-commerce industry. By utilizing the ease of information sharing, Craighead, Ketchen Jr & Darby (2020) emphasized that businesses should incorporate the supply chain management theory into their cost management practices based on the multidimensional data analysis of the upstream and downstream industry market status. This will improve the overall level of cost control for businesses. Reda et al. (2020) discussed using supply chain data analysis to make up for traditional cost management methods and studied new supply chain cost control strategies and enterprise product pricing models. Schmidt & Wagner (2019) studied supply chain management for analysis and research and analyzed the functional role, scenario mode, and performance evaluation of supply chain management. -Junior & Carpinetti (2020) positioned the relationship between enterprises and developed a new supply chain performance evaluation method. The author graded the supply chain management and proposed that the lead time of goods and the service level of retailers are crucial to the optimal decision-making and pricing of enterprise operations.

Lima
Data analysis based on supply chain management is a descriptive, predictive, and normative analysis of multimodal joint data based on upstream and downstream industries' market status and operation data (Tiwari, Wee & Daryanto, 2018). In addition to statistical analysis, multimodal joint data analysis can combine data collection, storage, knowledge management, and analysis to extract useful information from many upstream and downstream industry market status data, providing a reference for decision-making (Tavasszy, 2018). In order for decision-makers to understand the changing trend of the enterprise's external environment and to facilitate their analysis, Ouedraogo, Montarnal & Gourc (2022) believes that data analysis based on supply chain management can efficiently sort out upstream and downstream enterprise data as well as user-generated content. Goodarzian et al. (2021) proposed a demand framework for the supply chain of the medical industry based on the data analysis of upstream and downstream enterprises in the medical industry, which can effectively improve the efficiency of operational decision-making. The framework has been proven to be effective by experts in related fields. Sang (2021) introduced the rough set and BP neural network into the supply chain performance evaluation model and set the target error level within 10%. After reducing the index system in the balanced scorecard, they brought them into the neural network for sample training. The case proved that the performance evaluation value was consistent with the result. The mean square error level meets the set value. Yan et al. (2019) used the frequency analysis method and analytic hierarchy process to screen out evaluation indicators and obtained reasonable and effective evaluation results through sample training and analysis. On this basis, they continued to optimize supply chain performance and developed practical suggestions for optimizing enterprise supply chain performance. Based on the aforementioned context, this study develops a data reduction technique based on adaptive statistics, develops a supply chain evaluation index system based on a review of relevant theories of supply chain performance evaluation, and employs an LMBP feedback neural network model to assess the index system to support the efficient operation of the enterprise supply chain.

Multimode market status data selection
The manufacturing sector includes a wide variety of businesses. The completely different product types and manufacturing process characteristics of various manufacturing enterprises in the upstream and downstream sectors (such as electronic component manufacturing enterprises, small household appliance manufacturing enterprises, and food manufacturing enterprises) will result in the production of large amounts of multimodal joint data (Wang et al., 2022). In recent years, China's auto industry has  et al., 2019) is used with industry characteristics to screen and validate upstream and downstream industries' multimode market status data indicators with a significant influence. Twenty-eight types of market status data are selected for analysis to study the impact of supply chain management on the company's operational decisions. The multimode market status data of upstream and downstream industries are shown in Table 1. The massive multidimensional data analysis of upstream and downstream industry market status can significantly improve the efficiency of enterprise operation decisionmaking and promote enterprises to gain more in-depth insight into the supply chain through multi-angle and multi-level data analysis. However, it is unwise to analyze big data directly in the age of big data. Not every kind of data can play a positive role for extensive data, and redundant and noisy data often play a negative role. Therefore, it is necessary to reduce the volume of data through data reduction technology to reduce the computational complexity and improve the efficiency of analysis and mining.

Data reduction method based on adaptive statistics
The extraction of key factors of supply chain performance is a process of dimensionality reduction and feature selection of high-dimensional data. Principal components analysis (PCA) is a method to simplify data sets. Using dimension-reduction ideas is a commonly used method to transform multiple attributes into a few extended attributes. However, the traditional PCA algorithm is based on the sample correlation coefficient or covariance matrix. The research of robust statistics shows that these two statistics lack tolerance for outliers. Therefore, the traditional PCA algorithm is significantly affected by outliers and unsuitable for direct analysis of performance data of complex supply chains. In addition, traditional PCA only examines the linear relationship between variables, which restricts its use in practical analysis. To improve the performance of traditional PCA, this study proposes an enterprise operational data reduction method based on Adaptive statistics. It selects statistics with good robustness and easy calculation as indicators. The robust estimation uses the following common statistics as scale parameters, as shown in Eqs. (1)-(5).
Mean absolute deviation.
Median absolute deviation Quartered layout Q n scale: In this study, an adaptive robust scaling parameter estimation method is proposed, and its iterative algorithm is as follows: Step 1. Initialize the number of iterations i = 0, and calculate the initial estimator of the scale parameterσ 0 .
Step 3. Estimate the scale score function: take g(·) as the basis function, and take the polynomial function: g j (u) = u j .
We compared the robustness of these types of statistics through numerical experiments to screen out projection indicators with better robustness. To investigate the robustness of statistics, we randomly generate a certain number of samples according to a specific distribution and then create a group of samples with another distribution type as the pollution distribution. By adjusting the ratio of original samples to noise, the pollution rate ε, examine the changes in statistics to determine whether the object under investigation has a certain tolerance.
It can be seen from Figs. 1 and 2 that adaptive statistics have good robustness in the case of high noise. Therefore, this study uses adaptive statistics as the projection index function to characterize the dispersion of high-dimensional data on a one-dimensional projection and obtain a robust data reduction and dimension reduction method.
We identified a feature subset with high criticality among 28 types of market status data using the data reduction technique of adaptive statistics, which can best reveal the market state and support the effectiveness of business operation decision-making based on supply chain management. Figure 3 shows the calculation results of the criticality of 28 types of data indicators. Variables 1,5,8,11,14,15,18,21,23 and 25 are significant. Therefore, the feature subset we set up is X 1 ,X 2 ,X 3 ,X 4 ,X 5 ,X 6 ,X 7 ,X 8 ,X 9 ,X 10 , which includes the above 10 data indicators, including raw material transportation and arrival data, component supply and consumption data, production equipment working status data, various planning data, material integrity data, demand forecast data, supply source data, customer order data, customer personalized demand data and short-term trend data. It represents the critical data representing the market status of upstream and downstream industries.

A COMPANY OPERATION DECISION MODEL BASED ON SUPPLY CHAIN MANAGEMENT DATA
To build a model suitable for evaluating cross-border e-commerce supply chain performance, this study designed the LMBP algorithm by combining the BP and LM algorithms. While maintaining the convergence characteristics of the gradient steepest descent method, it also has the convergence speed of the Gauss-Newton method. N represents the number of input nodes, M represents the number of output nodes, jth represents the hidden layer unit, i th represents the input layer unit, net pj represents the input value to the hidden layer, x pj represents the value of the input layer unit, w ji represents the weight, w jo represents the threshold, g(net pj ) represents the excitation function of the hidden layer, w ko is the threshold of layer k, y represents the target output,ŷ represents the network output,ŷ pk is the target output of the k th , and η is the learning rate under the Newton method. The sum of the mean square error of the LMBP network is (y pk −ŷ pk ).
The input value of the hidden layer and the excitation function of the hidden layer are, respectively: w jt x pi (9) g net pj = 1/(1 + e −net pj ).
The target output ofŷ pk is expressed by Eq. (11): net pk = N j=0 w kj g net pj + w k0 ,ŷ pk = g net pk .
The weight update Eq. is: Here, the I is the unit matrix, and the weight value is updated after batch training, which makes the neural network learning process more stable and rapid and can adapt to a large amount of calculation. In the modeling process, the more critical the data, the more accurate the modeling. The iteration process mainly includes four steps: (1) Take all inputs, initialize the weight value, and variable µ(generally µ=0.01).
(2) Calculate the actual output, mean square error sum function, and Jacobian matrix of the input value.
(3) Gauss-Newton method is used to update the weights to form new parameters.
Ding ( (4) Calculate the mean square error E(w) repeatedly. If the mean square error decreases, update it. If the mean square error does not drop, do not update and skip to step 3. The LMBP calculation process is shown in Fig. 4.
Supply chain management adopts the coordinated management mechanism, links and coordinates the functional organizations, realizes the transformation of internal functional management to process management, puts the enterprise itself in the chain process, and realizes the coordinated linkage of the front and back end of the organization. In this study, LMBP feedback neural network algorithm is used to build a performance evaluation model of a cross-border e-commerce supply chain based on the LMBP algorithm, as shown in Fig. 5.
The model is mainly divided into three stages to evaluate the performance of the cross-border e-commerce supply chain: (1) Collect and process data: at this stage, it is necessary to collect data, complete the sorting and calculation of 28 evaluation indicators of the six-dimensional balanced scorecard, how complete the processing of qualitative data, how to adjust variables according to historical data, and how to normalize the actual data.
(2) LMBP algorithm is used for data training and learning: as shown in Fig. 6, the number of hidden layers, nodes, and transfer functions are selected scientifically, which makes data prediction and indicator evaluation more accurate and efficient.
(3) Training result analysis and performance optimization plan; finally, we need to analyze and process the training results in the early stage. On the one hand, we need to record the process and status of weight updating and neural network establishment. On the other hand, we need to analyze the path and scheme to improve the performance level by comparing the poor performance data with the excellent performance data.
In this study, two layers of the neural network are selected: one layer of the hidden layer and one layer of the output layer. For the number of remote layer nodes and neurons, it is necessary to ensure that there are enough and avoid excessive numbers. A disproportionate number of nodes will also lead to excessive computation and overfitting problems. Refer to the empirical Eq. (13) for the number of hidden layer nodes and the Ockham razor principle-the number of hidden layer nodes in a neural network should be less than the number of training set samples. This study analyzed 28 market status data of automobile manufacturers in China's listed manufacturing industry in 2022. The evaluation results are 4-dimensional data; the number of nodes in the input layer R is 28, and the number of nodes in the output layer S is 4. The calculation shows that the number of hidden layer nodes is reasonable between 7-16. n = (R + S) 1/2 + a,1 ≤ a ≤ 10.
To determine the optimal number of hidden layer nodes, we used the mean square error of training prediction results as a reference to train the 12-month data in 2021, continuously increase the number of hidden layer nodes, and repeat the same data. It can be concluded from Fig. 6 that the optimal number of remote layer nodes is 12, and the MSE at this time is the minimum, which is 2.62e−12.
The traditional BP algorithm uses the gradient steepest descent to update the weight value. Although it has good convergence, it has slowed local convergence defects. The LMBP algorithm combines the Gauss-Newton method and the steepest descent method, which can maintain the convergence characteristics of the gradient-most vertical descent method and effectively improve the convergence speed of the Gauss-Newton method. By understanding the LMBP algorithm's characteristics, this research builds a company operation model based on the LMBP algorithm. This model is combined with an evaluation index system composed of multimodal joint data of upstream and downstream industry market status. This can make establishing index weights more efficient and provide a more objective and novel method for evaluating supply chain performance.

AN EMPIRICAL ANALYSIS OF THE OPTIMIZATION OF CORPORATE OPERATION DECISION
The feature subset of the multidimensional data analysis of upstream and downstream industrial market state with high criticality is collected for this study using the adaptive statistics-based data reduction methodology. The weight of the feature subset is obtained based on the LMBP feedback neural network algorithm. Then, according to the supply chain performance analysis model and indicator system, we provide the basis for the company's operational decisions.
Because the original data has different dimensions, it needs to be expected. Different input variables usually fall into different ranges, leading to significant variables suppressing the performance of other variables, thus affecting the ability of the model to discover laws through input and output values. Therefore, we first normalize the data to eliminate the impact of dimensions. Use Eq. (14) for processing.
x i represents the value of index data before normalization, y i is the value after normalization, x max and x min represent the maximum and minimum values of the indicator data value range.
To intuitively represent the level of key data of upstream and downstream industries' market status, we define different score ranges and divide them into four performance levels: poor, reasonable, sound, and excellent. Concerning the literature and the opinions of managers of automobile manufacturing enterprises, we obtain the performance score range shown in Table 2. At the same time, in the analysis process, the four-dimensional data of (1000), (0100), (0010), and (0001) represent poor, medium, sound, and excellent, respectively.
The performance level of each 12 months in 2021 is shown in Table 3, which is represented by data and expressed as a matrix.   The number of hidden layers in the network of the company's operation decision-making model based on the LMBP algorithm proposed in this study is 2, the number of nodes in the input layer is 28, the number of nodes in the hidden layer is 12, the number of nodes in the output layer is 4, the transfer function is the logsig + purelin function, and the training function is trainlr. We used the dT matrix to represent the 12-month supply chain performance evaluation results of automobile manufacturers' output by the model in 2021. It can be seen from the dT results that the maximum error level is lower than 0.4%, which is an acceptable level. Therefore, S enterprise can obtain more accurate evaluation results using the supply chain performance evaluation index system and LMBP algorithm. The model is accurate and efficient.
In the above analysis process, we can see that the months when the supply chain performance of the automobile manufacturing enterprise is evaluated as ''poor'' are February and March, respectively. The 4-dimensional data is expressed as (1 0 0 0), the months when the supply chain performance is evaluated as ''excellent'' are November and December, and the 4-dimensional data is expressed as (0 0 0 1). We selected February as the object of decision optimization to discuss improving the supply chain performance optimization path in February.
The optimization model provides strategic guidance to optimize supply chain management and improve supply chain performance. For example, if you try to improve the performance by 75%, the normalized value of the customer order X8 should be increased from 0.664 to 0.69. It can be calculated from the reverse direction; the customer order should be increased from 26,300 in February to 32,800, while the material completeness rate X5 should be increased from 95% to 98.23%. As a result, by increasing the customer order and material integrity rate, the automotive manufacturing company can strengthen supply chain management and boost supply chain performance in the subsequent work.

CONCLUSIONS
As enterprises face the challenge of increased product diversity, they are also encountering heightened competition and more diverse customer needs in upstream and downstream industrial markets. This raises the bar for the operational decision-making abilities of companies. In this study, we selected automobile manufacturers in China's listed manufacturing industry as research samples. By combining industry-specific characteristics with the Delphi method, we identified 28 types of market status data indicators with significant impact on upstream and downstream industries. We were able to analyze and mine the data more effectively using data reduction techniques based on adaptive statistics. We pinpointed a highly critical feature subset of 10 indicators that can effectively reveal the state of the market and support efficient operational decision-making based on supply chain management. Using the LMBP algorithm, we developed a novel method for supply chain performance evaluation, offering strategic guidance for optimizing supply chain management and improving supply chain performance. Our study indicates that automobile manufacturing enterprises can enhance their supply chain management capability by improving customer order and material readiness rates.
However, we acknowledge some limitations in this study and suggest areas for future research and improvement in management practices, including: (1) Considering the internal interaction mechanism in the significant data capability structure dimension research, there are still hierarchical distinctions within big data capabilities.
(2) Exploring the impact of industry differences on extensive data research, as different industries have different priorities in the application process of big data.