Abstract
The problem to find out the truth from inconsistent information is defined as Truth Discovery. The essence of truth discovery is to estimate source quality. Therefore the measuring mechanism of data source will immensely affect the result and process of truth discovery. However the state-of-the-art algorithms dont consider how source quality is affected when null is provided by source. We propose to use the Silent Rate, True Rate and False Rate to measure source quality in this paper. In addition, we utilize Probability Graphical Model to model truth and source quality which is measured through null and real data. Our model makes full use of all claims and null to improve the accuracy of truth discovery. Compared with prevalent approaches, the effectiveness of our approach is verified on three real datasets and the recall has improved significantly.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
With the development of information technology, the Internet has penetrated into all corner of human social life. The data on internet have accumulated sharply, and these data have been integrated into an information ocean. One of the important features of this information is diversity, so for any object, heterogeneous descriptions can be found on internet from multiple sources. The inconsistency or conflict of these diverse descriptions causes great confusion for us to identify true information from each other. Therefore, identifying the accurate and complete information from conflicting descriptions is the key factor for information integration. The problem is the Truth Discovery proposed in document [9].
In order solve the problem, this paper makes use of the Hub Authority method [2, 4] to solve the problem by the quality difference of source. And we will redesign the metrics of source quality, and measure the quality of sources with three indexes, such as silent rate, true rate and false rate, to improve the accuracy of truth discovery. The main work of this paper presents is as follows:
-
1.
Redesign metrics for source quality. The quality of source is measured by three indexes, such as silent rate, true rate and false rate. The silent rate in the new metrics can make full use of the null data provided by the source, and can describe the source quality more comprehensively.
-
2.
Fertilize the plate model of probabilistic graph to construct model. The relationship among the data source, the object and the truth is constructed, and a probabilistic graph model is established to deduce the truth. Combing the conditions and methods of influence propagation in probabilistic graphs, the truth probability of every claim is deduced.
2 Related Work
The problem of truth discovery is defined for the first time in document [9], and a TruthFinder algorithm is proposed for solving this kind of problem. The method is similar to the iterative mechanism of Hub Authority, which synchronously infer the truth of object and the quality of the source. Inspired by it, a series of similar methods have been developed to study various factors that affect truth discovery [2,3,4]. The logical criterion of this method is: the higher the source quality is, the more likely it provides truth, at the same time, the more truth it provides, the higher the source quality is. These kind of algorithm is called heuristic method.
Based on the above processing logic, recent research has transformed truth discovery into a framework for optimization. Each source is assigned a weight according to its credibility, and the optimization objective is to minimize the distance between the target value and the truth. Document [6] which is extended in [5] uses the weight to represent the distribution of the credibility of each source, and uses a variety of loss functions to deal with heterogeneous data objects, so that the truth discovery of heterogeneous data is integrated into an objective function. Document [11] builds an optimization framework based on min-max entropy to estimate the truth of objects with noise. In a word, this kind of method updates the truth and the weight of the data source iteratively until the weight converges, and obtains the truth and the weight of all the data sources.
The other method is probability method. These methods solve the problem of truth discovery in document [1, 7, 8, 10]. The core idea is that multi-source data is considered as a mixed distribution, and the credibility of the source is integrated into the probabilistic model in the form of random variables. The probabilistic model is proposed by [10], and the quality of data sources is modeled by using two types of errors – false positive and false negative. Document [7] extends the truth discovery to the field of social group awareness. In document [8], the authors propose to use stochastic Gauss models to represent sources. So the mean represents the truth and variance represents the credibility of source. Theoretically, both solutions are proven to be contractive to an \(\epsilon \)-ball around the maximum likelihood estimate.
3 Implementation
In this section, we first introduce some terms and annotations. Suppose that S is a collection of sources, and j is one of the sources; N is the set of observation objects, and n is one of the objects. If a source j can only provide a claim for an observation object, then the \(n's\) claim obtained from j is denoted as \(x^{j}_{n}\). The \(n's\) truth is denoted as \(t_n\). The \(n's\) claim set is \(X_n\) = \(\{x^{j}_{n}\}_{j\in S_n}\) and \(N's\) claim set is X = \(X_1\cup \cdots \cup X_n\). In view of the fact that it is impossible to obtain the claims of object n from some sources, so the claims are only obtained from S\(_n\) i.e. a subset of S. We denote the collection of non-repeated claims of n as \(V_n=\{v_{ni}\},i=1,\cdots ,N' \).
In order to simplify the solution model, two hypotheses are given for the source and the object.
Hypotheses1 : The data sources are independent of each other, and they provide claims independently.
Hypotheses2 : The objects are independent of each other. The alteration of one object’s claim do not affect others’.
We construct a Bayesian network model as shown in Fig. 1 to calculate the probability of truth. In the network, each node represents a random variable, the dark node represents the known variable, and the light colored node represents the unknown variable. The solid black point \(\delta \) is the prior probability of the truth, and the hollow points represent the weight of the source measurement indexes. \(\beta ,\gamma ,\varTheta \) is a group of super parameters. \(\beta \) represents the weight of the false rate when calculating the truth. The \(\gamma \) is the weight of the true rate, and the \(\varTheta \) is the weight of the silent rate, satisfying equation \(\beta +\gamma +\varTheta =1\). The directed edges that connect nodes represent dependencies among variables. \(\vert \) S \(\vert \) and \(\vert \) N \(\vert \) in the corner of the box represent the number of sources and objects.
3.1 Source Quality
The most important factor that affects the accuracy of truth discovery is the quality of sources, and the previous algorithm does not consider the null data how to affect source quality. In this paper, true rate, false rate to measure source quality.
FalseRate(FR): The rate of all claims is not truth, which is provided by source j, i.e. \(FR=\frac{Fp}{Empt+Fp+Tp}\).
TrueRate(TR): The rate of all claims is truth, which is provided by source j, i.e. \(TR=\frac{Tp(1-FR)}{Empt+Fp+Tp}\).
SilentRate(SR): The rate of all claims is null, which is provided by source j, i.e. \(SR=\frac{Empt(1-FR)}{Empt+Fp+Tp}\).
Empt is the number of null among all the claims that source j provides i.e. \(Empt=\sum ^N_{n=1}Signal(x^j_n=null)\). Fp is the expectation of false claims that the source j provides when the probability of truth is \(p(t_n=v_{ni})\). Then, \( Fp=p(t_n=v_{ni})\times \sum ^N_{n=1}Signal(x^j_n \ne v_{ni} \& \& x^j_n \ne null)\). Tp is the expectation of true claims that the source j provides when the probability of truth is \(p(t_n=v_{ni})\). Then, \(Tp=p(t_n=v_{ni})\times \sum ^N_{n=1}Signal(x^j_n ==v_{ni})\). The Signal(.) is an indicator function in the formula: \(Signal(t)=\{^{1,\quad t=true}_{0,\quad t=false}\).
The quality of source j is \(\phi _j=(\phi ^1_j,\phi ^2_j,\phi ^3_j)\) which is a tri-tuple. The value of them is \(\phi ^1_j=FR,\phi ^2_j=TR,\phi ^3_j=SR\).
3.2 Truth Inference
According to the hypothesis 2, all |N| objects are independent of each other. From the probability graph model of Fig. 1, the probability of the observed values of all observed objects is:
\(P(x^1_n,\cdots ,x^{|S|}_n|\phi _S)\) is a joint probability density function of Object n when |S| sources provide claims.
The \(V_n\) is the claim set of Object n whose truth, \(t_n\), has \(|V_n|\) possible values. Then:
According to the hypothesis 1, if the sources are independent of each other to provide claims, and there is no mutual replication of claims, the probability of obtaining all the claims of object n from |S| sources is a joint probability function:
According to the definition of the problem, the truth probability of the object n is actually the truth probability under the condition of the current claim and the source quality, i.e.
According to Bayes formula,
In formula (5), the denominator is exactly the same for all objects. Then, the truth probability \(p(t_n=v_{ni})\) is proportional to the molecular term. After substituted formula (3), formula (5) changed into formula (6).
Formula (6) is a continued product. The likelihood of the truth probability is obtained.
3.3 Iterative Computation
The model uses silent rate, truth rate and false rate to measure source quality. \(\beta \), \(\gamma \) and \(\theta \) is the weight to adjust impact on three index. Therefore, \(\beta \), \(\gamma \) and \(\theta \) is substituted into the formula (7). The recursive formula of the final truth probability can be obtained.
Give an assumption: \( f1=\sum ^{|X_n|}_{i=1}\sum ^{N}_{n=1}{r_{ni}}^{(k)}\times Signal(x^j_n \ne v_{ni} \& \& x^j_n\ne null)\), \(f2=\sum ^{N}_{n=1}Signal(x^j_n==null)\), \(f3=\sum ^{|X_n|}_{i=1}\sum ^{N}_{n=1}{r_{ni}}^{(k)}\times Signal(x^j_n==v_{ni})\)
We can get the formulas of \({\phi ^1_j}^{(k)},{\phi ^2_j}^{(k)}\) and \({\phi ^3_j}^{(k)}\).
![figure a](http://media.springernature.com/lw685/springer-static/image/chp%3A10.1007%2F978-3-319-93713-7_37/MediaObjects/469704_1_En_37_Figa_HTML.gif)
The whole procedure of the FTS is shown in Algorithm 1. First, the prior probabilities of the claims of each object are initialized by uniform distribution and source quality \(\phi \) is initialized by standard normal distribution. Then, according to the initial source quality, the whole iteration process is started. Finally, the (k+1)-th \(r_{ni}\) i.e. truth probability is calculated with k-th iteration of source quality i.e. \({\phi _j}^{k}\). The new source quality can be produced with new truth probability. The iterative algorithm is repeated until the source quality updates less than the threshold and the algorithm converges. The truth and source quality is calculated synchronously.
4 Experiment
In this section the truth discovery algorithm will run on three real data sets to verify the performance of the designed FTS model. As a comparison, the FTS and two state-of-art algorithms are tested under the same conditions to acquire respective recall.
4.1 Experimental Setup
The configuration of all experiments in this paper: CPU is Intel(R) Core(TM) i7-6700 3.40 GHz. Memory is 64G and OS is Windows 7. In this paper, three real datasets (http://lunadong.com/) such as WeatherDataset, WeatherDataset, and FlightDataset are used to verify the effect of the algorithm.
4.2 Experimental Result
In order to verify the accuracy of the algorithm, the truth discovery accuracy test was carried out on three real datasets, and the test results were listed in Table 1.
In the experiment, the set of ground truth is split into two parts randomly, one is used for verification, and the other is used for testing. \(\beta ,\gamma \) and \(\theta =1-\beta -\gamma \) is used to adjust the weight of silent rate, true rate and false rate. During the processes of verification, \(\beta ,\gamma \) is taken from the range (0,1) and the step is 0.02 and the optimal combination of, is obtained. With the optimal \(\beta ,\gamma ,\theta \), recall is calculated on testing set.
From the experimental data in Table 1, we can see that our FTS method is better than the classical truth discovery algorithm on recall. It means silent rate has positive impact on truth discovery. The recall on weather and flight datasets is significantly decreased. The main reason for this phenomenon is that the number of sources in these two datasets is small, which affects the effect of truth discovery.
In order to test the influence of silent data source on the recall, the Book-Author data set are processed, and randomly selected 20\(\%\), 40\(\%\), 60\(\%\), 80\(\%\) of the sources to keep silent i.e. claims is null. The experimental results are shown in Fig. 2. With the increase of silent source, the recalls of all algorithm decline steadily. If 100\(\%\) of the sources remain silent, no algorithm can predict the truth, so the recall is 0. Comparing with the three algorithms, the Vote algorithm drops the fastest, which shows that the accuracy of the algorithm heavily depends on the number of sources that provide claims. The TrueFinder algorithm is also affected by the number of silent sources, and its performance on recall is lower than FTS. With comprehensive comparison, FTS algorithm is affected little by the number of silent source. The reason is that the effect of silent rate has been considered in FTS model.
5 Conclusion
First of all, the metrics of source quality has been redesigned. In FTS model, the reliability of data source quality is measured by true rate, false rate and silent rate. Thus the situation that source provides null is comprehensively comprised. Secondly, a probabilistic graph model is established to construct the relationship among the source, the object and the truth. Then the truth is deduced. In this model, the relationship among the source metrics, the truth and the claims of the source is presented in the form of figures. Using the conditions and methods of influence propagation in probabilistic graphs, the probabilities of every claim as truth are deduced. The experimental results on three datasets show that the new algorithm significantly improves the recall of truth discovery compared with the traditional classical algorithm.
References
Blanco, L., Crescenzi, V., Merialdo, P., Papotti, P.: Probabilistic models to reconcile complex data from inaccurate data sources. In: International Conference on Advanced Information Systems Engineering, pp. 83–97 (2010)
Dian, Y., Hongzhao, H., Taylor, C., Ji, H., Chi, W., Shi, Z., Jiawei, H., Clare, V., Malik, M.I.: The wisdom of minority: Unsupervised slot filling validation based on multi-dimensional truth-finding. In: Proceedings of 2014 International Conference on Computational Linguistics, pp. 1567–1578 (2014)
Furong, L., Mong Li, L., Wynne, H.: Entity profiling with varying source reliabilities. In: ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1146–1155 (2014)
Galland, A., Abiteboul, S., Marian, A., Senellart, P.: Corroborating information from disagreeing views, pp. 131–140 (2010)
Li, Q., Li, Y., Gao, J., Su, L., Zhao, B., Demirbas, M., Fan, W., Han, J.: A confidence-aware approach for truth discovery on long-tail data. Very Large Data Bases 8(4), 425–436 (2014)
Li, Q., Li, Y., Gao, J., Zhao, B., Fan, W., Han, J.: Resolving conflicts in heterogeneous data by truth discovery and source reliability estimation, pp. 1187–1198 (2014)
Wang, D., Kaplan, L.M., Le, H.K., Abdelzaher, T.F.: On truth discovery in social sensing: a maximum likelihood estimation approach, pp. 233–244 (2012)
Xiao, H., Gao, J., Wang, Z., Wang, S., Su, L., Liu, H.: A truth discovery approach with theoretical guarantee. In: ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1925–1934 (2016)
Yin, X., Han, J., Yu, P.S.: Truth discovery with multiple conflicting information providers on the web. IEEE Trans. Knowl. Data Eng. 20(6), 796–808 (2008)
Zhao, B., Rubinstein, B.I.P., Gemmell, J., Han, J.: A Bayesian approach to discovering truth from conflicting sources for data integration. Very Large Data Bases 5(6), 550–561 (2012)
Zhou, D., Basu, S., Mao, Y., Platt, J.: Learning from the wisdom of crowds by minimax entropy, pp. 2195–2203 (2012)
Acknowledgement
This work was supported by National Natural Science Foundation of China (No. 61601458) and National Key Research and Development Program of China (No. 2016YFB0801004, 2016YFB0801305).
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Zhang, J., Wang, S., Wu, G., Zhang, L. (2018). A Effective Truth Discovery Algorithm with Multi-source Sparse Data. In: Shi, Y., et al. Computational Science – ICCS 2018. ICCS 2018. Lecture Notes in Computer Science(), vol 10862. Springer, Cham. https://doi.org/10.1007/978-3-319-93713-7_37
Download citation
DOI: https://doi.org/10.1007/978-3-319-93713-7_37
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-93712-0
Online ISBN: 978-3-319-93713-7
eBook Packages: Computer ScienceComputer Science (R0)