Next Article in Journal
Analysis of the Accuracy of Ten Algorithms for Orientation Estimation Using Inertial and Magnetic Sensing under Optimal Conditions: One Size Does Not Fit All
Next Article in Special Issue
Performance Study of Distance-Weighting Approach with Loopy Sum-Product Algorithm for Multi-Object Tracking in Clutter
Previous Article in Journal
Match Analysis of Soccer Refereeing Using Spatiotemporal Data: A Case Study
Previous Article in Special Issue
Unscented Particle Filter Algorithm Based on Divide-and-Conquer Sampling for Target Tracking
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Credibility Assessment Method of Sensor Data Based on Multi-Source Heterogeneous Information Fusion

School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(7), 2542; https://doi.org/10.3390/s21072542
Submission received: 19 February 2021 / Revised: 30 March 2021 / Accepted: 31 March 2021 / Published: 5 April 2021
(This article belongs to the Special Issue Multi-Sensor Fusion for Object Detection and Tracking)

Abstract

:
The credibility of sensor data is essential for security monitoring. High-credibility data are the precondition for utilizing data and data analysis, but the existing data credibility evaluation methods rarely consider the spatio-temporal relationship between data sources, which usually leads to low accuracy and low flexibility. In order to solve this problem, a new credibility evaluation method is proposed in this article, which includes two factors: the spatio-temporal relationship between data sources and the temporal correlation between time series data. First, the spatio-temporal relationship was used to obtain the credibility of data sources. Then, the combined credibility of data was calculated based on the autoregressive integrated moving average (ARIMA) model and back propagation (BP) neural network. Finally, the comprehensive data reliability for evaluating data quality can be acquired based on the credibility of data sources and combined data credibility. The experimental results show the effectiveness of the proposed method.

1. Introduction

Wireless sensor networks (WSN) have become popular in various areas [1,2,3,4]. For example, in agricultural production, WSN are used to monitor the growth environment and status of crops; in terms of industrial safety, WSN are applied to monitor the safety of dangerous working environments such as coal mines, oil drilling, and nuclear power plants to ensure the safety of workers. However, due to the interference of many external factors such as sensor aging, instrument failure, measurement methods, and human interference, the data collected by sensors may not be precise or reliable [5]. The existence of these unreliable data will lead to inefficient data utilization, waste of economic costs, and even serious decision-making errors. Thus, false alarms and missed alarms, which reduce the performance of the monitoring system greatly, will be caused by the unreliable monitoring system. If the credibility of the original data can be evaluated effectively, then the unreliable and low-quality data can be detected from the original data, contributing to the data processing, and the remaining high-reliability data will improve the early warning capabilities of the supervision and inspection system, and thereby the safety of people’s life and property can be guaranteed to the maximum extent. For real event data, if the data collected by data sources reflect the real environment, this method considers the data as credible, which could be used in further data analysis. When the data of multiple sensors are interfered, this method will not be able to make an effective evaluation of the data, so the event data and unreliable data cannot be handled.
In order to evaluate the credibility of data, scholars have conducted many methodological studies. For example, Marsh [6] first proposed a mathematical model for calculating credibility, and at the same time, explained the definition of data credibility. He considered that credibility is a way to understand a complex environment, a means to provide additional robustness to independent subjects, and an effective indicator to make judgments based on the experience of others.
According to the above definition, Marsh suggested several important directions for subsequent research. Galland [7] presented three fixed-point algorithms based on the complexity of the data source to evaluate the credibility of the data provided by the data source. Furthermore, based on the experimental results, he designed two schemes to enhance robustness of the method: establishing weights for different data sources and using prior knowledge. Aiming at the uncertainty and conflict of multi-source data, Guo [8] et al. used a reliability evaluation framework based on evidence theory, which combines prior information and context information, and the algorithm can be improved by correlating with more information. Jikui Wang et al. [9] suggested an evaluation method of data with Bayesian estimation theory, which relies too much on the accuracy of the prior data of the data source. Scholars such as Fernando [10] aimed at the problem of evaluating the credibility of sensors for autonomous driving systems by machine learning to build a framework for data credibility, and using Lidar to verify the system. However, models are directly established from the data in the above algorithms, without constructing the spatio-temporal relationship between the data sources. Thus, it is not possible to exploit the advantage that the correlation between data sources can handle the uncertainty of the data. For the literature [11], Xinjian Wan et al. designed a credibility measurement model, which considered the relationship between data sources, in light of the problem that the credibility of air quality monitoring data cannot be evaluated. Hence, this was our main comparison method in this article.
Multi-source data fusion is a processing means for multi-source data to derive estimates and judgments from the original data sources through knowledge-based reasoning and recognition to enhance the confidence of data, improve reliability, and decreased uncertainty, with the advantages of comprehensive information, high fault tolerance, reduced cost, and broader coverage. It is a complex estimation process. In order to make full use of the information in the image fusion process, Xue J et al. [12] developed a data fusion method based on Bayesian, and finally the fused image was obtained by maximum a posteriori probability (MAP). However, the Bayesian criterion is greatly limited in engineering applications due to the need to obtain the prior probability of each data source, and this requirement is difficult to know in advance. In [13], three multi-sensor data fusion algorithms based on Kalman filter, namely state vector fusion (SVF), measurement fusion (MF), and gain fusion (GF) are implemented in a tracking system. Results showed that the Kalman filter-based state vector fusion algorithm performed well comparatively for the system. Although the Kalman filter has good performance in target tracking applications, it also has the problem of poor real-time performance. L.Z. et al. [14] presented weather recognition methods based on the weighted fusion of weather-specific features for classifying five types of weather. It is simple to achieve the weighted average method, but the real-time performance is also poor. Moreover, some algorithms do not take the fluctuation of data collected by sensors in the time series into account, and do not reflect the accuracy of data fusion. The BP neural network has the advantages of simple design, fast training speed, strong self-adaptive ability, and it has good fault tolerance because it does not cause a great impact on the global training results after its local or partial neurons are damaged [15]. Here, this paper proposes a new comprehensive credibility of data measurement (CDM) method, which combines the correlation between data sources and the fusion algorithm. This method makes full use of the relationship between data sources to further enhance the robustness. It also utilizes the historical data of the target data source and the multi-source heterogeneous data to improve the accuracy of the evaluation. Based on the above characteristics of the BP neural network, the fusion algorithm was selected as a BP neural network. Moreover, the ARIMA model [16] has the characteristics of stable prediction, simple, and easy implementation. Therefore, the model based on ARIMA and BP neural network can provide stable, real-time, and accurate assessment of the credibility of data.
The rest of this article is organized as follows. The definition of credibility and the components of data credibility are introduced in Section 2, and the basic concepts of the ARIMA model and BP neural network are described in detail. The proposed method through examples is verified and the experimental results are presented in Section 3. Our conclusion and next work plan are presented in Section 4.

2. Materials and Methods

Currently, there is no clear and accepted definition of the credibility of data. Pinyol I [17] discussed the objective aspect of credibility and argued that credibility can be regarded as a decision in the process of interacting with another party. Sabater J [18] analyzed the user’s credibility evaluation mechanism of eBay, Amazon Auctions, and OnSale Exchange, and noted that the credibility models of these online markets all used single data that do not depend on contextual information. In [19], the authors considered that data credibility is determined by a combination of data integrity, consistency, and data accuracy. Consistency indicates whether the logical relationships between related data are correct, data integrity describes the severity of data loss in the dataset, and accuracy demonstrates how close the measured data are to the real data. For this article, the credibility of the data provided by the data source was considered to be the degree to which the measured object in the expected environment reflects the true state. That is, the discussion focused on the accuracy index as an important index to measure the performance of the data credibility measurement model.
First, there is an overview of the methodology. Furthermore, there is an introduction of the symbols, a definition of comprehensive data credibility, a meaning, and the metrics of the parameters in formulas are described in detail. Finally, we briefly review the related algorithms presented.

2.1. Overview

In order to clarify the relationship of the different factors, we present the framework of the method, and the specific process of the method is shown in Figure 1.
Figure 1a shows the framework of the method, which illustrates the relationship of the credibility of the data source and the combined credibility of the data, and shows the components of the different type of credibility. Figure 1b demonstrates the process of predicting data directly through the time correlation of data. Figure 1c shows the process of predicting the heterogeneous data through the spatio-temporal correlation.

2.2. Symbols

In this section, we describe the symbols used in this paper. The symbols and descriptions are shown in Table 1.

2.3. Definition of Comprehensive Data Credibility

The evaluation method of comprehensive data credibility combines multiple sensors and multiple types of sensors in both temporal and spatial dimensions. Comprehensive data credibility is composed of data source credibility and combined data credibility. The calculation of data credibility is as follows:
R t i = { S t i × D t i , S t i > Θ 0 e l s e
where R t i is the comprehensive data credibility of the data source i ; S t i represents the credibility of the data source i ; D t i is the combined data credibility of the data source i ; and Θ is the credibility threshold of the data source. When the data source is credible, the comprehensive data credibility is the result of multiplying the credibility of the data source and the combined credibility of the data. If the credibility of the data exceeds 0.5, the data is credible, otherwise the data is not credible. In this paper, the data credibility is represented by the continuous value between the interval [0,1]. Next, the credibility of the data source and the combined credibility of the data are elaborated.

2.4. Definition of the Credibility of the Data Source

Data source credibility will be calculated through trend correlation and mean aggregation, used to describe whether the data source is credible and whether it is working properly. If the data source is not credible, the credibility of the data is determined to be zero. The formula for the credibility of the data source is as follows.
S t i = K 11 C tr i + K 12 C a v e i
where S t i is the credibility of the data source; C t r i is the trend correlation of the data source i ; C a v e i is the mean aggregation of the data source i ; and K 11 and K 12 weighting factors of trend correlation and mean aggregation, respectively, where K 11 + K 12 = 1 . The weights are determined based on the importance of trend correlation and mean aggregation. Through multiple iterations, the weights are determined with the highest detection probability and the lowest false alarm rate. The definition of the detection probability and false alarm rate is introduced in the experimental section. The following parameters were the same.
(1)
Calculation of trend correlation
Setting a distance threshold φ , this is a Euclidean distance. When the distance between the target data source and a data source is less than the threshold, the data source is regarded as an adjacent data source. Trend correlation refers to the average degree of correlation between the historical data trend of the target data source and its adjacent data sources. If the data source is damaged or does not work correctly, C tr i will be smaller. The mathematical expression of the trend correlation is as follows.
C tr i = 1 N j = 1 N | r ( T r i ( d s ) , T r j ( d s ) ) |
r ( T r i ( d s ) , T r j ( d s ) ) = s = 1 S ( T r i ( d s ) T r i ( d ) ¯ ) ( T r j ( d s ) T r j ( d ) ¯ ) s = 1 S ( T r i ( d s ) T r i ( d ) ¯ ) 2 s = 1 S ( T r j ( d s ) T r j ( d ) ¯ ) 2
where T r i ( d s ) is the trend of a piece of the historical data; r is a correlation function; the data source j is adjacent to the data source i ; and N is the number of adjacent data sources of the target data source; the vertical line indicates the absolute value; S is the number of historical data; and T r ( d ) ¯ represents the mean value of the historical data.
(2)
Calculation of mean aggregation
Mean aggregation is the degree of aggregation between the average value of the target data source and its neighboring data sources of the same type over the same time. Similarly, it also describes the degree of dispersion of average data between data sources with spatial correlation. For example, when a data source is manually interfered, it is possible that the data trend of the target source is still consistent with that of the neighboring data sources, but the value will decrease, and the mean value will deviate from the mean value of other neighboring data sources, then C a v e i will be reduced. The calculation formula for the mean aggregation is as follows:
C ave i = 1 | 1 M p = 1 M | D a i D a p | d ¯ i p 1 M N p = 1 M q = 1 N | D a p D a q | d ¯ p q | + 1
d ¯ i p =   1 - d i p p = 1 M d i p
where D a i is the average value of the historical data; the data source p is the adjacent sensor of the data source i ; M is the number of adjacent sensors of the data source i ; the data source q is adjacent to the data source p ; and N is the number of adjacent sensors of the data source p ; and d ¯ is the normalized distance factor between data sources.

2.5. Definition of the Combined Credibility of Data

Combined credibility of the data consists of direct credibility of the data and heterogeneous credibility of the data. The formula for the combined credibility of the data is as follows:
D t i = ( K 21 F t i + K 22 M t i )
where D t i is the combined credibility of the data provided by data source i ; F t i is the direct credibility of the data provided by data source i ; M t i is the Heterogeneous credibility of data; and K 21 and K 22 are the weighting factors of direct credibility of the data and heterogeneous credibility of the data, respectively, where K 21 + K 22 = 1 .
(1)
Calculation of direct credibility of the data
Direct credibility of the data is the degree of membership of the measured value to the theoretical real value. Namely, the historical data of the data source is used for model fitting based on the ARIMA model, and then the theoretical real value at the current moment is predicted according to the time correlation based on the fitting function. The directly credibility is obtained based on the measured value and theoretical real value. The formula for the direct reliability of the data is as follows.
F t i = { 1 Δ f t i η   3 η Δ f t i 2 η η < Δ f t i < 3 η 0 Δ f t i > 3 η
Δ f t i = | R p t i S p t i |
where R p t i is the measure data of the data source i ; S p t i is the theoretical real value (direct prediction data) based on the historical data; and η   is the credibility threshold of the data.
(2)
Calculation of heterogeneous credibility of the data
Heterogeneous credibility of the data is the degree of membership of the measured value to the theoretical real value. Multi-source heterogeneous data, which are from the target data source and its multiple neighboring heterogeneous data sources within the same period of time, are used in data prediction currently, based on the BP neural network. Then, heterogeneous reliability is calculated according to theoretical real value and measured value. The formula for the heterogeneous reliability of the data is as follows.
M t i = { 1 Δ m t i η   3 η Δ m t i 2 η η < Δ m t i < 3 η 0 Δ m t i > 3 η
Δ m t i = | R p t i M p t i |
where M p t i is the theoretical real value (heterogeneous prediction data) of data source i at time t by fusing the data of the adjacent data sources.

2.5.1. Autoregressive Integrated Moving Average Model

  • Overview
In this part, we introduce the proposed time series forecasting model. The model can extract the historical data trend of the data source to predict the future data. First of all, the parameters used in this section are explained. The parameters and descriptions are shown in Table 2.
The full name of the A R I M A is called the autoregressive integrated moving average model, which is also recorded as A R I M A ( p , d , q ) . A R I M A is the most common statistical model used for time series forecasting [20,21], which is an extension of the A R M A ( p , q ) model [22,23,24]. The relationship between the two models is shown in Figure 2.
The A R M A model consists of an autoregressive model and a moving average model, where, p and q are the two parameters of the A R M A model, and the meanings of these parameters are shown in Table 2. The mathematical representation of the A R M A model is as follows:
y t = μ + i = 1 p φ i y t i + e t + j = 1 q θ j e t j
where y t is the prediction value at time t ; μ is a constant; φ is the coefficient of the A R model; θ is the coefficient of the M A model; and e is the error between the predicted value and the measured value.
Compared with the A R M A model, the A R I M A model adds a difference algorithm, the purpose of which is to convert a non-stationary sequence into a stationary sequence to meet the stationary demand of A R I M A . Among them, p , d , q , are the three parameters of the A R I M A model. The meaning of each parameter is shown in Table 2. The mathematical representation of the A R I M A model is as follows:
y t ( d ) = μ + i = 1 p φ i y t i ( d ) + e t + j = 1 q θ j e t j
where y t ( d ) is the predicted value of the d-order difference of the data at time t.
It is important to establish an appropriate model for the accuracy of data prediction. There are three steps to build a time series prediction model, namely, model ordering. The model flow chart is shown in Figure 3.
  • First, it is necessary to judge whether the time series data are stable, and if they are not stable, corresponding processing is required such as difference and logarithmic operations;
  • Second, if the data series is stable, the next step is to determine the order p , q , which can be determined artificially by calculating the autocorrelation function ( A C F ) and the partial autocorrelation function ( P A C F );
  • Third, since there is a certain subjective component in the process of determining the order in the previous step, it is necessary to verify whether the parameters are reasonable through the index function, and the index function is selected as the Akaike Information Criterion ( A I C )/Bayesian Information Criterion ( B I C ); and
  • Finally, after the model is established, the future data can be predicted through the model.
  • Parameter Selection
In the A R I M A model, the parameters of ( p , q ) can be determined artificially through the A C F and P A C F , and the visualizations are shown in Figure 4. Finally, ( p , q ) were determined as (3,3). Due to the strong subjective factors in order determination through A C F and P A C F , we can again determine the order according to the information criterion function method. In this experiment, the A I C index function was selected, and the parameters ( p , q ) were selected as (5,3). Then, the results of the data credibility were obtained by the model with two different parameters and the results are shown in Figure 5.
Observing Figure 5a,b, the difference between the results was not very great. Both sets of parameters can be used. In this paper, the parameters of ( p , q ) are finally confirmed as (5,3).

2.5.2. Back Propagation (BP) Neural Network

  • Overview
The B P neural network is a multi-layer feedforward neural network [25,26,27] including an input layer, an output layer, and one or more hidden layers. The main characteristics are signal forward propagation and error back propagation. Specifically, the training process of the neural network can be divided into two stages. The first stage is signal forward propagation, where the input of the network is from the input layer through the hidden layer before finally obtaining the result from the output layer; in the second stage, error back propagation, error and gradient go from the output layer to the hidden layer and finally back to the input floor. The basic simple structure is shown in Figure 6.
Suppose this network has n inputs, m outputs, and a hidden layer, among which there are s neurons in the hidden layer. Suppose the output of the hidden layer is a j , the bias is θ j , the weight is w i j , and the activation function is f 1 . The output of the output layer is y k , the bias is θ j , the weight is w j k , and the activation function is f 2 . The mathematical expression for calculating the output of the hidden layer is as follows:
a j = f 1 ( i = 1 n w i j x i + θ j ) ( j = 1 , 2 , , s )
The mathematical expression for calculating the output of the output layer is as follows:
y k = f 2 ( j = 1 s w j k a j + θ k ) ( k = 1 , 2 , , m )
The error function of the network is defined as follows:
e = k = 1 m ( y k t k ) 2
where t k is the true value (expected value) of the output.
The objective of the training process is to reduce the error as much as possible. The forward process is to calculate y k and the error e , and then back-propagate the error to modify the weight coefficient of each layer, repeat iteratively, and until the training is ended
  • Parameters Selection
This neural network also takes advantage of the spatio-temporal relationship be-tween data sources, and uses data from neighboring data sources as input for learning. During the learning process, the number of neurons in each layer has a greater impact on the training results as it not only affects the training speed, but also affects the training accuracy. Here, on the basis of other parameters unchanged, we set the number of hidden layers of the network to 2, and the influence of the number of different neurons in each layer on the training accuracy is shown in Figure 7.
In Figure 7, it can be seen that when the number of neurons in each hidden layer is six, the error was the smallest. Therefore, the number of neurons was six, and the other parameters were set as follows: Learning rate = 0.01, b a t c h s i z e = 12 , and e p o c h = 20 .

3. Results and Discussion

A simulation was conducted to verify the effectiveness of the proposed method, and whether the credibility measurement method has the ability to detect unreliable data from the original dataset.

3.1. Experiment Design

The selected data of the simulation experiment were from the alleyway smoke simulation dataset generated by P y r o s i m software, which was developed abroad for fire simulation. The distribution of the dataset is shown in Figure 8. In the process of data collection, two kinds of data, smoke density and temperature, were collected respectively.
The red circle area is the distribution of the data sources used in the experiment. In order to see the correlation distribution of the data sources more clearly, the top view of the region is shown in Figure 9.
The dataset contained 1000 sets of data provided by eight data sources. Table 3 shows the data samples.
Many parameters are mentioned in this paper, which include the weight of trend correlation, the weight of mean aggregation, the weight of data direct credibility, the weight of data heterogeneous credibility, the credibility threshold of the data, and the credibility threshold of the data source. In this experiment, these parameters were set based on subjective judgment and experience, and the parameters were specifically set as shown in Table 4.
Before the experiment, the data of smoke density and temperature were considered to be completely reliable after manual screening and verification. Based on this data, the artificial interference to the data was divided into normal interference and large abnormal interference, and the purpose of the experiment was to judge whether the presented method could identify abnormal interference and give the corresponding results of data credibility.
In this dataset, 100 sets of data in the time of T499–T500 were selected as the data to be evaluated in data samples, in which SD08 was taken as the target data source, and SD08 was interfered with between T499 and T500, according to the previously designed data interference type. Other data sources were all neighboring data sources of the target data source, where the neighboring data sources with the same type as the target data source are called neighboring data sources of the same type.
The flow chart of the entire experimental process is shown in Figure 10. First, the trend correlation of the data was calculated according to Equation (3). Before the calculation, the historical data interval needs to be selected. Through tests, when the data interval length was 10, it could best reflect the real situation of the environment in real time without reducing the accuracy. Moreover, Equation (5) was used to calculate the mean aggregation of the data. Based on the trend correlation and mean aggregation, the credibility of the data source was calculated by Equation (2). Second, using the ARIMA models, the direct credibility of data was acquired by Equations (8) and (9) using the BP neural network, and the heterogeneous data reliability was obtained according to Equations (10) and (11). Furthermore, based on the direct credibility of data and heterogeneous credibility of the data, the data combined credibility was calculated by Equation (7). Finally, the comprehensive data credibility was achieved according to Equation (1).
Next, the interfered target data and the data generated by its neighboring data sources are demonstrated in Figure 11.
The data in red circles in Figure 11a are the artificially disturbed data. The remaining data are normal data. Figure 11b is the data from the adjacent sources.

3.2. Analysis of Results

3.2.1. Experiment Result

In order to evaluate the performance of the algorithm, we defined the detection probability and false alarm rate of incredible data, which can be seen in Equations (17) and (18). Figure 12 shows the data trend correlation and mean aggregation. In Figure 12a, the trend correlation was mainly maintained between [0.6, 0.8], which showed a downward trend, but did not reach a low value during T440–T460. This is because the interference for sensors only makes the value smaller, but the fluctuation of the value is not completely changed. Hence, the trend correlation during this period was not too low. However, for the mean aggregation, the target data were at a lower level for a longer time than the data from the neighboring sources of the same type between time T440 and T460. The difference between the data became larger, leading to a smaller mean aggregation. Figure 12b, indeed, shows the corresponding change.
p d = d D
p f a = f F
In the above equations, p d is the detection probability; d is the number of incredible data detected correctly; D is the total number of incredible data added; p f a is the false alarm rate; and f is the number of incredible data detected incorrectly. F is the total number of credible data.
After determining the parameters of each algorithm, the parameters were also configured according to Table 4, and then the data credibility measurement model based on multi-source heterogeneous information fusion was built. During the experiment, three sets of single-point error data and 21 sets of error data in the time T440–T460 were artificially constructed to simulate the situation where the data source provided the error data, mainly to prove the model’s ability of detecting unreliable data. The evaluation results are presented in Figure 13.
In Figure 13, the credibility of the data remained stable between time T400 and T420. As the target data were not disturbed in this time, the data are credible. At the time of T420, T464, and T472, the credibility of the data dropped significantly, especially at the time of T420 and T472. Due to the great interference, the data credibility dropped below 0.5. During the time period T440–T460, the target data source was not credible, and the data credibility directly dropped to zero because the data were changed too large. The experimental results showed that when the original data were modified manually, the model could detect the wrong data and could dynamically adjust the credibility of the data in a larger and smaller range. When the measured data differs greatly from the real data, the credibility of the data will drop below 0.5 or even zero.

3.2.2. Performance Comparison

In order to verify the advantage of the spatio-temporal relationship in the BP neural network for predicting data, some neural network prediction models were used to compare the performance in the simulations. These models are explained below. This comparison used two datasets, one was the dataset generated by P y r o s i m software, and the other was the Intel indoor dataset [28], as a benchmark dataset to study the data prediction in the wireless sensor networks. This Intel indoor dataset consisted of 54 sensor nodes distributed in the laboratory, and the location distribution of 54 sensor nodes is shown in Figure 14. The Intel Berkeley Research lab used the Mica2Dot wireless sensor as the collection node to collect the four physical quantities of ambient temperature, humidity, light, and battery voltage.
Here, we used node 13 as an example and as the center, and then selected node 11, node 12, and node 14 as its neighboring sources. Considering the data correlations of signal-node and multi-node, which are shown in Table 5 and Table 6, the paper selected the temperature and humidity data of node 13, the temperature data of node 11, node 12, and node 14, and the humidity data of node 12 and node 14 as the input data of the BP neural network, which were used to predict the temperature data of node 13.
For the simulation dataset, according to the different input data, we can divide the network into the homogeneous network and heterogeneous network. The BP neural network is equivalent to thee heterogeneous network because of using different types of data, smoke data, and temperature data of different sources. Figure 15a shows the data prediction of the different network. For the benchmark dataset, the BP network in this paper selected the multi-data of multi-node, so it is called the multi-node multi-feature (MNMF) model here. In contrast to this model, a signal-node multi-feature (SNMF) model and a multi-node single-feature (MNSF) are presented. The SNMF model uses the data of a signal node, for example, the temperature and humidity data of node 13. The MNSF model uses the same type of data of a multi-node, for example, the temperature data of node 13, node 11, node 12, and node 14. Figure 15b shows the data prediction of the three prediction models. The mean-square error (MSE) is used as evaluation indicator to evaluate the prediction model. The formula for evaluation standard is shown as Equation (19).
M S E ( y , y * ) = 1 m i = 1 m ( y i y i * ) 2
In the above equation, y i is the true value; y i * is the predicted value; and m is the number of samples. The calculated evaluation indicators are shown in Table 7.
Figure 15a shows that the heterogeneous network had a better performance than the homogenous network, and Figure 15b shows that the MNMF model had better prediction than the other models. It can be seen from Figure 8 that the heterogeneous network and the MNMF model both had a low prediction error and a high prediction accuracy. Moreover, the heterogeneous network and the MNMF model used the spatio-temporal relationship between different data sources and fused heterogeneous information. This experiment proved that the BP neural network in this paper has a great performance for predicting data.
According to the analysis in Section 3.2.1, the feasibility and validity of the proposed model were verified. In this section, the methodology in this paper is compared with the Air Quality Credible Measurement Model (AQCM) in [11]. In order to further verify the effectiveness of the method, we again selected the Intel Lab data as the benchmark to study. The interfered target data and the data generated by its neighboring data sources of the benchmark dataset are demonstrated in Figure 16. The comparison result based on the two datasets is depicted in Figure 17.
In Figure 17a, the vertical axis represents the data credibility, and the horizontal axis expresses the time period. At T420, T464, and T472, the CDM model reflected the current state of the data more accurately than the AQCM model. For the AQCM model, only at T472 did the credibility of the data decrease significantly, but it was not less than 0.5. Therefore, the data at time T420, T464, and T472 were judged as credible, which is not consistent with the real situation. In the T440–T460 time period, the evaluation results of the two methods were less than 0.5, and the data were considered to be unreliable, whereas the evaluation result of the methodology in this paper was closer to zero. Figure 17b shows that the CDM model had a better evaluation performance than the AQDM model, especially at times of T50 and T100. For the temperature of node 13 in the benchmark dataset, the data at time T50 and T100 were unreliable. However, the AQDM model made the data reliable. In time period of T150–T180, the evaluation result of the AQDM was closer to zero. It can be seen from Figure 16 that the CDM model had an accurate evaluation and a flexible capability of assessment.

3.2.3. Extent Experiment

The purpose of this extended experiment was to verify the robustness of the model. The data source SD11 adjacent to the target source was also interfered in the same time period of T440–T446. The CDM model was applied to this situation, and the results are demonstrated in Figure 18.
Figure 18a shows the credibility of heterogeneous data when two data sources are interfered at the same time. During T440–T460, as the input data of the BP neural network suffer from the same interference, this will cause the network output to also be disturbed, therefore, the evaluation results of data credibility will also be disturbed. In Figure 18b, the comprehensive evaluation of data credibility in T440–T460 were not destroyed, precisely because the correlation between data sources was utilized. The mean aggregation between the target source and other adjacent sources was still close to 0, so that the data credibility in T440–T460 approached 0, which verified that the model also had good robustness.

4. Discussion

This paper used two kinds of data credibility—a direct and a heterogeneous credibility. The advantages relative to the individual use of one of them are shown below. The direct credibility utilized the time relationship through the historical data of the target source. Compared with the historical data of other sources, historical data and current data of the same target source had the greatest correlation in normal condition. Therefore, it is necessary to assess the credibility of the data in normal condition. Heterogeneous credibility utilizes the spatio-temporal relationship of the data sources, adding different dimensions to improve the evaluation accuracy. This is vital when the historical data of the target source is not reliable, if the source has been damaged. The combination of the direct and the heterogeneous credibility makes the method more accurate and more robust than the individual use of one of them.
Multi-sensor data fusion is an efficient tool for improving the performance because of the spatio-temporal relationship between data sources. However, there are some challenges and future work to be discussed.
(1)
Improvements of the application of the method
In this work, we proposed a method that could effectively utilize the spatio-temporal relationship of the data sources. In order to simplify the problem, we assumed that data integrity and time lag between data were not considered in our experimental set. However, in field measurement, missing data is one of the most common problems. When sources lose data, the performance of this method will be harmed. Due to the hugeness and redundancy of wireless sensor network data, we should apply proper data preprocessing methods to deal with data missing and other data problems. In general, data interpolation is an effective method including quadratic interpolation, cubic interpolation, linear interpolation, and so on.
(2)
Development of the fusion algorithm
Each fusion method has its own set of advantages and limitations. The combination of several different fusion schemes has been approved as a useful strategy that may achieve a better performance of results [29]. However, the selection and arrangement of fusion schemes do not have not determined criteria [30]. Further investigations are necessary to design the general framework to combine the fusion algorithm.
(3)
Enhancement of the robustness of the model
The proposed method uses many neighboring sources to improve the performance. This method also has limitations. When there are many data sources that are simultaneously interfered, the model may fail. We expect that future research will continue to optimize the model.

5. Conclusions

Aiming at the problem of the low accuracy of data credibility assessment, this paper proposed a method for evaluating the credibility of sensor data based on multi-source heterogeneous information fusion with the A R I M A model and the B P neural network. This method considers two factors: the spatio-temporal dependency between data sources and the temporal correlation between time series data. In this paper, the simulation dataset was used for the experiment and the Intel indoor dataset as a benchmark dataset was used to compare the performance. These experiments show that the proposed method, making use of the advantages of multiple sources, has a high evaluation accuracy and a flexible adjustment ability. At the same time, it is also robust to some extent.

Author Contributions

Conceptualization, Y.F. and J.H.; Methodology, Y.F. and J.H.; Validation, Y.F.; Formal analysis, R.D.; Data curation, Z.C.; Writing—original draft preparation, Y.F. and J.H.; Writing—review and editing, R.D. and Z.C.; Supervision, Z.C. and R.D.; Project administration, R.D. and Z.C.; Funding acquisition, R.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Project of China (grant number 2018YFC0808302).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Buratti, C.; Conti, A.; Dardari, D.; Verdone, R. An overview on wireless sensor networks technology and evolution. Sensors 2009, 9, 6869–6896. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Ramson, S.R.J.; Moni, D.J. Applications of wireless sensor networks—A survey. In Proceedings of the 2017 International Conference on Innovations in Electrical, Electronics, Instrumentation and Media Technology (ICEEIMT), Coimbatore, India, 3–4 February 2017; pp. 325–329. [Google Scholar]
  3. Jesus, G.; Casimiro, A.; Oliveira, A. A survey on data quality for dependable monitoring in wireless sensor networks. Sensors 2017, 17, 2010. [Google Scholar] [CrossRef] [PubMed]
  4. Kocakulak, M.; Butun, I. An overview of Wireless Sensor Networks towards internet of things. In Proceedings of the 2017 IEEE 7th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 9–11 January 2017; pp. 1–6. [Google Scholar]
  5. Abdelgawad, A.; Bayoumi, M. Data fusion in WSN. In Resource-Aware Data Fusion Algorithms for Wireless Sensor Networks; Springer: Boston, MA, USA, 2012; pp. 17–35. [Google Scholar]
  6. Marsh, S.P. Formalising Trust as a Computational Concept. Available online: https://www.nr.no/~abie/Papers/TR133.pdf (accessed on 30 April 1994).
  7. Galland, A.; Abiteboul, S.; Marian, A. Corroborating information from disagreeing views. In Proceedings of the Third ACM International Conference on Web Search and Data Mining, New York, NY, USA, 4–6 February 2010; pp. 131–140. [Google Scholar]
  8. Guo, H.; Shi, W.; Deng, Y. Evaluating Sensor Reliability in Classification Problems Based on Evidence Theory. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2006, 36, 970–981. [Google Scholar] [CrossRef] [PubMed]
  9. Wang, J.K. Bayesian conflicting Web data credibility algorithm. J. Zhejiang Univ. (Eng. Sci.) 2016, 50, 2380–2385. [Google Scholar]
  10. Castaño, F.; Strzelczak, S.; Villalonga, A.; Haber, R.E.; Kossakowska, J. Sensor Reliability in Cyber Physical Systems Using Internet-of-Things Data: A Review and Case Study. Remote Sens. 2019, 11, 2252. [Google Scholar] [CrossRef] [Green Version]
  11. Wang, X.; Li, X.; Zhao, Y.; Li, Y.; Zhang, B. Credibility Analysis of Air Quality Data Based on Improved Measurement Method. In Proceedings of the 2019 Chinese Control and Decision Conference (CCDC), Nanchang, China, 3–5 June 2019; pp. 4421–4425. [Google Scholar] [CrossRef]
  12. Xue, J.; Leung, Y.; Fung, T. A Bayesian data fusion approach to spatio-temporal fusion of remotely sensed images. Remote Sens. 2017, 9, 1310. [Google Scholar] [CrossRef] [Green Version]
  13. Anitha, R.; Renuka, S.; Abudhahir, A. Multi sensor data fusion algorithms for target tracking using multiple measurements. In Proceedings of the 2013 IEEE International Conference on Computational Intelligence and Computing Research, Enathi, India, 26–28 December 2013; pp. 1–4. [Google Scholar]
  14. Li, Z.; Li, Y.; Zhong, J.; Chen, Y. Multi-class weather classification based on multi-feature weighted fusion method. In Proceedings of the IOP Conference Series: Earth and Environmental Science, Dalian, China, 4–5 July 2020; Volume 558, p. 042038. [Google Scholar]
  15. Guo, Y.; Zhao, Z.; Huang, L. SoC estimation of Lithium battery based on improved BP neural network. Energy Procedia 2017, 105, 4153–4158. [Google Scholar] [CrossRef]
  16. Benvenuto, D.; Giovanetti, M.; Vassallo, L.; Angeletti, S.; Ciccozzi, M. Application of the ARIMA model on the COVID-2019 epidemic dataset. Data Brief 2020, 29, 105340. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, T.C.; Xie, Y.Z.; Yan, H. Research of multi sensor information fusion technology based on extension neural network. Math. Model. Eng. Probl. 2016, 3, 129–134. [Google Scholar] [CrossRef]
  18. Pinyol, I.; Sabater-Mir, J. Computational trust and reputation models for open multi-agent systems: A review. Artif. Intell. Rev. 2013, 40, 1–25. [Google Scholar] [CrossRef]
  19. Sabater, J.; Sierra, C. Review on Computational Trust and Reputation Models. Artif. Intell. Rev. 2005, 24, 33–60. [Google Scholar] [CrossRef]
  20. Cai, L.; Zhu, Y. The challenges of data quality and data quality assessment in the big data era. Data Sci. J. 2015, 14, 2. [Google Scholar] [CrossRef]
  21. Ariyo, A.A.; Adewumi, A.O.; Ayo, C.K. Stock price prediction using the ARIMA model. In Proceedings of the 2014 UKSim-AMSS 16th International Conference on Computer Modelling and Simulation, Cambridge, UK, 26–28 March 2014; pp. 106–112. [Google Scholar]
  22. Liu, C.; Hoi, S.C.; Zhao, P.; Sun, J. Online arima algorithms for time series prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  23. Jang, J.; Shin, S.; Lee, H.; Moon, I.-C. Forecasting the Concentration of Particulate Matter in the Seoul Metropolitan Area Using a Gaussian Process Model. Sensors 2020, 20, 3845. [Google Scholar] [CrossRef] [PubMed]
  24. Abonazel, M.R.; Abd-Elftah, A.I. Forecasting Egyptian GDP using ARIMA models. Rep. Econ. Financ. 2019, 5, 35–47. [Google Scholar]
  25. Yu, F.; Xu, X. A short-term load forecasting model of natural gas based on optimized genetic algorithm and improved BP neural network. Appl. Energy 2014, 134, 102–113. [Google Scholar] [CrossRef]
  26. Du, Y. Application and analysis of forecasting stock price index based on combination of ARIMA model and BP neural network. In Proceedings of the 2018 Chinese Control and Decision Conference (CCDC), Shenyang, China, 9–11 June 2018; pp. 2854–2857. [Google Scholar]
  27. Ji, S.; Yu, H.; Guo, Y.; Zhang, Z. Research on sales forecasting based on ARIMA and BP neural network combined model. In Proceedings of the 2016 International Conference on Intelligent Information Processing, Wuhan, China, 18–21 December 2016; pp. 1–6. [Google Scholar]
  28. Intel. Intel Lab Data. Available online: http://db.csail.mit.edu/labdata/labdata.html (accessed on 19 April 2019).
  29. Gang, W.A.; Xiangyang, L.I.; Guangen, W.U.; Yong, G.U.; Simin, M.A. Research on Data Fusion Method Based on Rough Set Theory and BP Neural Network. In Proceedings of the 2020 International Conference on Computer Engineering and Application (ICCEA), Guangzhou, China, 18–20 March 2020; pp. 269–272. [Google Scholar]
  30. Dong, J.; Zhuang, D.; Huang, Y.; Fu, J. Advances in multi-sensor data fusion: Algorithms and applications. Sensors 2009, 9, 7771–7784. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Overall flow chart. (a) The framework of the method. (b) The direct prediction of the data. (c) The heterogeneous prediction of the data.
Figure 1. Overall flow chart. (a) The framework of the method. (b) The direct prediction of the data. (c) The heterogeneous prediction of the data.
Sensors 21 02542 g001
Figure 2. The relationship between A R M A and A R I M A .
Figure 2. The relationship between A R M A and A R I M A .
Sensors 21 02542 g002
Figure 3. Program flow diagram of the model.
Figure 3. Program flow diagram of the model.
Sensors 21 02542 g003
Figure 4. The visualization of A C F and P A C F .
Figure 4. The visualization of A C F and P A C F .
Sensors 21 02542 g004
Figure 5. The comparison of the A R I M A model with two different parameters. (a) The parameters of ( p , q ) were (3,3); (b) the parameters of ( p , q ) were (5,3).
Figure 5. The comparison of the A R I M A model with two different parameters. (a) The parameters of ( p , q ) were (3,3); (b) the parameters of ( p , q ) were (5,3).
Sensors 21 02542 g005
Figure 6. Program flow diagram of the model.
Figure 6. Program flow diagram of the model.
Sensors 21 02542 g006
Figure 7. The influence of the different number of neurons on the result.
Figure 7. The influence of the different number of neurons on the result.
Sensors 21 02542 g007
Figure 8. Software model diagram.
Figure 8. Software model diagram.
Sensors 21 02542 g008
Figure 9. Top view of distribution.
Figure 9. Top view of distribution.
Sensors 21 02542 g009
Figure 10. The flow chart of the experimental process.
Figure 10. The flow chart of the experimental process.
Sensors 21 02542 g010
Figure 11. Data visualization. (a) Interfered data; (b) Data from neighboring sources.
Figure 11. Data visualization. (a) Interfered data; (b) Data from neighboring sources.
Sensors 21 02542 g011
Figure 12. The credibility of the data source. (a) The trend correlation; (b) The mean aggregation.
Figure 12. The credibility of the data source. (a) The trend correlation; (b) The mean aggregation.
Sensors 21 02542 g012
Figure 13. Assessment results of comprehensive data credibility.
Figure 13. Assessment results of comprehensive data credibility.
Sensors 21 02542 g013
Figure 14. Location distribution of the sensor nodes.
Figure 14. Location distribution of the sensor nodes.
Sensors 21 02542 g014
Figure 15. Prediction on two datasets. (a) Prediction on the simulation dataset. (b) Prediction on the benchmark dataset.
Figure 15. Prediction on two datasets. (a) Prediction on the simulation dataset. (b) Prediction on the benchmark dataset.
Sensors 21 02542 g015
Figure 16. Data visualization. (a) Interfered data. (b) Data from neighboring sources.
Figure 16. Data visualization. (a) Interfered data. (b) Data from neighboring sources.
Sensors 21 02542 g016
Figure 17. Comparison of credibility measurement models. (a) Simulation dataset. (b) Benchmark dataset.
Figure 17. Comparison of credibility measurement models. (a) Simulation dataset. (b) Benchmark dataset.
Sensors 21 02542 g017
Figure 18. The evaluation result of data credibility. (a) Results of the BP assessment of data credibility. (b) Results of the comprehensive data credibility.
Figure 18. The evaluation result of data credibility. (a) Results of the BP assessment of data credibility. (b) Results of the comprehensive data credibility.
Sensors 21 02542 g018
Table 1. Symbols and descriptions.
Table 1. Symbols and descriptions.
NotationsDescription
C tr i Data trend correlation of data sources
C ave i Data mean aggregation of data source
S i The credibility of the data source
T r i Data trend function of data source
F i Direct credibility of the data
M i Heterogeneous credibility of data
D i Combined credibility of the data
R p i Measured value of data source
S p i Direct prediction data of data source
M p i Heterogeneous prediction data of data source
R i ( d n ) Comprehensive data credibility of data sources
η   The credibility threshold of the data
Θ The credibility threshold of the data source
Table 2. List of parameters.
Table 2. List of parameters.
ParametersDescription
p called autoregressive term, also called A R term
q the number of moving average terms, also called M A terms
d the difference term, also called the integrated term
Table 3. Original data sample.
Table 3. Original data sample.
TimeSD06SD07SD09SD011THCP04THCP05THCP06SD08
000002020200
1000020.0014520.0001220.000120
2000020.0107120.0008220.0010
33.69 × 10−131.1 × 10−140020.0205120.0037620.002570
436.850660.2507259.38 × 10−16021.3580820.618120.006682.49 × 10−13
580.587952.666310.0004124.05 × 10−1328.289425.3147720.0146527.33613
99991.5918893.429593.4609792.4529849.4599947.5713546.1367494.31777
Table 4. Experimental parameter settings.
Table 4. Experimental parameter settings.
ParametersDefault ValueDescription
K 11 0.2Weight coefficient of trend correlation
K 12 0.8Weight coefficient of mean aggregation
K 21 0.4Weight coefficient of data direct credibility
K 22 0.6Weight coefficient of data heterogeneous credibility
η   10The credibility threshold of the data
Θ 0.5The credibility threshold of the data source
φ 10The distance between the target source and other sources.
Table 5. Correlation of the signal-node.
Table 5. Correlation of the signal-node.
CorrelationTemperatureHumidityLightVoltage
Temperature1.00−0.9430.6870.699
Humidity−0.9431.00−0.549−0.689
Light0.687−0.5491.000.41
Voltage0.699−0.6890.411.00
Table 6. Correlation of the multi-node.
Table 6. Correlation of the multi-node.
CorrelationTemperatureHumidityLightVoltage
No. 13, No. 110.9050.6550.8710.912
No. 13, No. 120.8850.8500.8550.897
No. 13, No. 140.9330.8820.9020.845
Table 7. Prediction loss of different models on the two datasets.
Table 7. Prediction loss of different models on the two datasets.
DatasetModelRSE
Simulation datasetHeterogeneous network1.2178
Homogeneous network1.7224
Benchmark datasetMNMF0.062
SNMF0.176
MSNF0.141
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Feng, Y.; Hu, J.; Duan, R.; Chen, Z. Credibility Assessment Method of Sensor Data Based on Multi-Source Heterogeneous Information Fusion. Sensors 2021, 21, 2542. https://doi.org/10.3390/s21072542

AMA Style

Feng Y, Hu J, Duan R, Chen Z. Credibility Assessment Method of Sensor Data Based on Multi-Source Heterogeneous Information Fusion. Sensors. 2021; 21(7):2542. https://doi.org/10.3390/s21072542

Chicago/Turabian Style

Feng, Yanling, Jixiong Hu, Rui Duan, and Zhuming Chen. 2021. "Credibility Assessment Method of Sensor Data Based on Multi-Source Heterogeneous Information Fusion" Sensors 21, no. 7: 2542. https://doi.org/10.3390/s21072542

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop