Skip to main content
Log in

A framework to restrict viewing of peer commentary on Web objects based on trust modeling

  • Original Article
  • Published:
Social Network Analysis and Mining Aims and scope Submit manuscript

Abstract

In this paper, we present a framework aimed at assisting users in coping with the deluge of information within social networks. We focus on the scenario where a user is trying to digest feedback provided on a Web document (or a video) by peers. In this context, it is ideal for the user to be presented with a restricted view of all the commentary, namely those messages that are most beneficial in increasing the user’s understanding of the document. Operating within the computer science subfield of artificial intelligence, the centerpiece of our approach is a modeling of the trustworthiness of the person leaving commentary (the annotator), determined on the basis of ratings provided by peers, adjusted by a modeling of the similarity of those peers to the current user. We compare three competing formulae for restricting what is shown to users which vary in the extent to which they integrate trust modeling, to emphasize the value of this component. By simulating the knowledge gains achieved by users (inspired by methods used in peer-based intelligent tutoring), we are able to validate the effectiveness of our algorithms. Overall, we offer a framework to make the Social Web a viable source of information, through effective modeling of the credibility of peers. When peers are misguided or deceptive, our approach is able to remove these messages from consideration, for the user.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Notes

  1. This is on par with what is known as the Hamann similiarty measure (Choi and Cha 2010).

  2. The annotator reputation serves only as a proxy for ratings when an annotation has not received votes (e.g., is relatively new). We begin with an assumption of 0.5 as the annotator reputation (on a scale of [0,1]).

  3. Having an explicit representation of unrated annotations allows differentiation between rated and not yet rated.

  4. The function that we used to determine the similarity of two users in their rating behavior examined annotations that both users had rated and scored the similarity based on how many ratings were the same (both thumbs-up or both thumbs-down). The overall similarity score ranged from −1 to 1.

  5. Our usage of these terms is not perfectly analogous to Zhang and Cohen (2008). We clarify the distinctions further in Sect. 5.

  6. A value of 0.2 was used for \(\gamma\) in our simulations.

  7. This is a novel approach that assists greatly in confirming the value of our proposed approach in environments where there may be a massive number of peers, something that would be otherwise challenging to confirm through user studies alone. We return to clarify these challenges in Sect. 5.

  8. Our collaborative learning algorithm to select the document is described in greater detail in Champaign and Cohen (2010). Once that document is identified, we reason about which annotations attached to it will be shown. In brief, this work details an approach which combines objects, annotations, and the specific user to model how experiencing that combination of object and annotations will modify that particular user’s knowledge.

  9. The negative impact was introduced to simulate the possibility of misinformation from a poor quality learning object or a learning object that does a good job teaching one concept, while undermining the understanding of another concept.

  10. While this might initially seem to be a small scale simulation, it should be noted that this is the equivalent to 400 students each studying for over 300 h—a total of 133,333 h of instruction for each simulation.

  11. This simulates the idea that an annotation left on a learning object adjusts how it is received by students. It can adjust the quality of instruction (raising or lowering the impact) or adjust the target audience (raising or lowering the target level of instruction) when combined with the learning object being examined.

  12. This perfect knowledge was obtained by running the simulated learning twice, once with the annotation and learning object, and once with just the learning object. A simulated user gave a positive rating if it learned more with the annotation and a negative rating if it learned more without.

References

  • Breese JS, Heckerman D, Kadie C (1998) Empirical analysis of predictive algorithms for collaborative filtering. Morgan Kaufmann Publishers, Burlington

    Google Scholar 

  • Champaign J, Cohen R (2010) A distillation approach to refining learning objects. In: Proceedings of 2010 educational data mining conference, pp 283–284

  • Champaign J, Zhang J, Cohen R (2011) Coping with poor advice from peers in peer-based intelligent tutoring: the case of avoiding bad annotations of learning objects. In: Proceedings of user modeling, adaptation and personalization (UMAP), pp 38–49

  • Choi S, Cha S, Tappert CC (2010) A Survey of binary similarity and distance measures, J Syst Cybern Inf 8(1):43–48

  • Gorner J, Zhang J, Cohen R (2011) Improving the use of advisor networks for multi-agent trust modelling. In: 2011 9th Annual international conference on privacy, security and trust (PST), pp 71–78

  • Hang CW, Singh MP (2012) Generalized framework for personalized recommendations in agent networks. Auton Agents Multi-Agent Syst 25(3):475−498

  • Hang CW, Zhang Z, Singh MP (2012) Generalized trust propagation with limited evidence. IEEE Comput 46(3):78–85

  • Herlocker JL, Konstan JA, Terveen LG, Riedl JT (2004) Evaluating collaborative filtering recommender systems. ACM Trans Inf Syst Secur (TISSEC) 22(1):5–53

    Article  Google Scholar 

  • Jsang A, Ismail R (2002) The beta reputation system. In: Proceedings of the 15th Bled conference on electronic commerce

  • Matsuda N, Cohen WW, Sewall J, Lacerda G, Koedinger KR (2007) Evaluating a simulated student using real students data for training and testing. In: User modeling, pp 107–116

  • McCalla G (2004) The ecological approach to the design of e-learning environments: purpose-based capture and use of information about learners. J Interact Media Educ 7:1–23

    Google Scholar 

  • Nepal S, Bista SK, Paris C (2012a) An association based approach to propagate social trust in social networks. In: Workshop and poster proceedings of the 20th conference on user modeling, adaptation, and personalization (TRUM 2012)

  • Nepal S, Bista SK, Paris C (2012b) SRec: a social behaviour based recommender for online communities. In: Workshop and poster proceedings of the 20th conference on user modeling, adaptation, and personalization (SRS 2012)

  • Nepal S, Sherchan W, Paris C (2011) Building trust communities using social trust. In: UMAP 2011 workshops, pp 243–255

  • Natural Sciences and Engineering Research Council of Canada (NSERC). NSERC Strategic Network: Healthcare support through information technology enhancements. PI: Dr. D. Plant

  • Sardana N (2013) Recommending messages to users in participatory media environments: a bayesian credibility approach. Master’s thesis, University of Waterloo

  • Seth A, Zhang J, Cohen R (2010) Bayesian credibility modeling for personalized recommendation in participatory media. In: De Bra P, Kobsa A, Chin D (eds) User modeling, adaptation, and personalization, vol 6075. Lecture notes in computer science. Springer, Berlin, pp 279–290

    Google Scholar 

  • Teacy WTL, Patel J, Jennings NR, Luck M (2006) Travos: trust and reputation in the context of inaccurate information sources. Auton Agents Multi-Agent Syst 12(2):183–198

    Article  Google Scholar 

  • VanLehn K, Ohlsson S, Nason R (1996) Applications of simulated students: an exploration. J Artif Intell Educ 5:135–175

    Google Scholar 

  • Vassileva J, McCalla G, Greer J (2003) Multi-agent multi-user modeling in I-Help. User Model User-Adapt Interact 13:179–210

    Article  Google Scholar 

  • Zhang J, Cohen R (2007) A comprehensive approach for sharing semantic web trust ratings. Comput Intell 23(3):302–319

    Article  MathSciNet  Google Scholar 

  • Zhang J, Cohen R (2008) Evaluating the trustworthiness of advice about seller agents in e-marketplaces: a personalized approach. Electron Commer Res Appl 7(3):330–340

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to J. Champaign.

Appendix: Algorithmic complexity

Appendix: Algorithmic complexity

Although our results demonstrate a clear ability to avoid displaying low-quality annotations, the system does require some computational effort to decide which comments will be displayed, the rate at which this computational effort scales will be quite important when considering the deployment of the system, especially in very large environments like social networks. Ideally, a deployed scheme should scale linearly in the resource usage otherwise required by the site (i.e., at worst a constant multiple of the product of the total number of users with the average number of web objects loaded by each user). This would be consistent with the usage rates of other resources required in a social network setting (storage, bandwidth, etc.). We will show that, although a naive implementation of our system would not scale linearly under all trust measures (the Cauchy measure in particular), a trivial change to the implementation would allow for linear scaling and thus widespread deployment, by using approximation algorithms when necessary.

We consider three possible trust metrics in this paper: the Tally, Cauchy, and trust-based measures. Computing the Tally measure for an individual user viewing a particular Web object means computing the sum of all votes cast for the object. However, this only needs to be done once, since the Tally measure of an object’s value is the same for every user (i.e., it does not take into account the similarity between users). Further, the Tally measure for an object can be updated in constant time: A new vote simply changes the total score of the object. Consequently, the Tally measure requires a constant amount of storage space per object, no additional storage space per user, and both computation and bandwidth scale linearly in the total number of votes cast (and thus, in the worst case, the total number of users times the average number of Web objects loaded per user). Further, it is trivial to store annotations sorted by their Tally scores, and thus any of the techniques used to select an annotation to display (apart from the Greedy God) could choose an annotation in constant time as well. Thus, when using the Tally measure, our system is guaranteed to be efficient and could be readily deployed to a large scale environment.

The Cauchy measure initially appears to be rather more computationally expensive, but we will show that with a simple sampling algorithm, we can achieve a linear runtime similar to the Tally measure (though with a larger constant factor). When computing the Cauchy measure to determine whether a particular user viewing a particular Web object should be shown a particular annotation, we first must compute the \(T_q\) term in Eq. 5. Although this requires \(O(|A_q| \times \hbox{mean}(|R^{ai}|))\) computational effort, where \(\hbox{mean}(|R^{ai}|)\) is the mean number of votes over the annotator \(q\)’s annotations, and \(|A_q|\) is the number of such annotations, this term can be computed just once, and updated incrementally, much like the Tally measure, for a given annotation. It can also be retrieved in constant time when required. Consequently, the computation of this term scales linearly with the usage of other required resources. The other term in Equation 5, however, is a weighted average of votes cast, based on the similarity of the viewer to each of the voters in question. A naive implementation would then require us to store a record of every vote cast by every user (which is still linear in terms of other resource usage, though potentially large), and also to perform \(O(|U|*|C|)\) work to compute the final value, where \(|U|\) is the total number of users in the system, and \(|C|\) is the total number of annotations across all such objects. Since this amount of computational effort is required every time any user accesses any annotation, it scales roughly quadratically with the usage of other required resources and would clearly be infeasible for use in a large scale setting like a commercial social network. However, in the event that the number of users who have interacted with a given object is large (>100), statistical sampling techniques can be used to obtain an estimate of \((vF^A - vA^a)\) in constant time. To accomplish this, we could draw a small constant number (n) of users from those who cast votes on a particular annotation. If the users are drawn without replacement, than a value of \((vF^A - vA^a)\) computed using only the drawn votes is an unbiased estimator for the true value, with error proportionate to \(\frac{1}{\sqrt{(}n)}\). Similarly, when computing the similarity between any two users, we could draw samples from the set of objects both users have cast ballots over and thus provide a constant bound on both the computation time and error in our estimations. The resulting Cauchy approximation would scale linearly with the usage of other required resources and so could also be deployed in a social network, or other large-scale application. However, the constant coefficients would be somewhat larger than for the other two methods.

The trust-based measure, a combination of the Tally measure with the \(T_q\) term used in the Cauchy measure, scales linearly with other required resource usage by virtue of both its components scaling linearly (as shown above). Like the Tally measure, annotations can be sorted according to their trust-based scores and then selected for display efficiently. Consequently, the trust-based measure can also be applied directly to large scale applications.

Given the performance gains provided by the Cauchy measure in situations where authors generate undesirable annotations most of the time, its slightly larger (though still just a linear multiple of other resource usage in the worst case) computational and storage costs could be justified. However, in applications where most content is desirable to show, the Tally and trust-based measures may be preferred for their lower resource usage, and strong absolute performance.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Champaign, J., Cohen, R., Sardana, N. et al. A framework to restrict viewing of peer commentary on Web objects based on trust modeling. Soc. Netw. Anal. Min. 4, 203 (2014). https://doi.org/10.1007/s13278-014-0203-7

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s13278-014-0203-7

Keywords

Navigation