Distributed ACO Based Reputation Management in Crowdsourcing

  • Safina Showkat AraEmail author
  • Subhasis Thakur
  • John G. Breslin
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11874)


Crowdsourcing is an economical and efficient tool that hires human labour to execute tasks which are difficult to solve otherwise. Verification of the quality of the workers is a major problem in Crowd sourcing. We need to judge the performance of the workers based on their history of service and it is difficult to do so without hiring other workers. In this paper, we propose an Ant Colony Optimization (ACO) based reputation management system that can differentiate between good and bad workers. Using experimental evaluation, we show that, the algorithm works fine on the real scenario and efficiently differentiate workers with higher reputations.


Crowdsourcing Reputation Ant Colony Optimization Decentralised social network 



This publication has emanated from research supported in part by a research grant from Science Foundation Ireland (SFI) under Grant Number SFI/12/RC/ 2289-P2(Insight) and by a research grant from SFI and the Department of Agriculture, Food and the Marine on behalf of the Government of Ireland under Grant Number SFI/12/RC/3835(VistaMilk), co-funded by the European Regional Development Fund.


  1. 1.
  2. 2.
  3. 3.
    von Ahn, L., Dabbish, L.: Labeling images with a computer game. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (2004)Google Scholar
  4. 4.
    Ara, S.S., Thakur, S., Breslin, J.G.: Expertise discovery in decentralised online social networks. In: ASONAM 2017 (2017)Google Scholar
  5. 5.
    Awwad, T.: Context-aware worker selection for efficient quality control in crowdsourcing (2018)Google Scholar
  6. 6.
    Dalvi, N., Dasgupta, A., Kumar, R., Rastogi, V.: Aggregating crowdsourced binary ratings. In: WWW 2013 (2013)Google Scholar
  7. 7.
    Dorigo, S.: Ant Colonies and the Mesh-Partitioning Problem, pp. 203–208. MIT Press, Cambridge (2004)Google Scholar
  8. 8.
    Dwarakanath, A., Shrikanth, N.C., Abhinav, K., Kass, A.: Trustworthiness in enterprise crowdsourcing: a taxonomy and evidence from data. In: ICSE 2016 (2016)Google Scholar
  9. 9.
    Anta, A.F., Georgiou, C., Mosteiro, M.A., Pareja, D.: Algorithmic mechanisms for reliable crowdsourcing computation under collusion. PLoS One 10, e0116520 (2015) CrossRefGoogle Scholar
  10. 10.
    Ho, C.J., Slivkins, A., Vaughan, J.W.: Adaptive Contract Design for Crowdsourcing Markets: Bandit Algorithms for Repeated Principal-Agent Problems (2014)Google Scholar
  11. 11.
    Jagabathula, S., Subramanian, L., Venkataraman, A.: Reputation-based worker filtering in crowdsourcing. In: Advances in Neural Information Processing Systems 27 (2014)Google Scholar
  12. 12.
    Kamar, E., Horvitz, E.: Incentives for truthful reporting in crowdsourcing. In: AAMAS 2012 (2012)Google Scholar
  13. 13.
    Liu, N., Yang, H., Hu, X.: Adversarial detection with model interpretation. In: KDD 2018 (2018)Google Scholar
  14. 14.
    Liu, S., Miao, C., Liu, Y., Yu, H., Zhang, J., Leung, C.: An incentive mechanism to elicit truthful opinions for crowdsourced multiple choice consensus tasks. In: WI-IAT (2015)Google Scholar
  15. 15.
    Lu, J., Pan, L., Liu, S., Liu, X.: Distributed ACO based on a crowdsourcing model for multiobjective problem. In: CSCWD (2017)Google Scholar
  16. 16.
    Movshovitz-Attias, D., Movshovitz-Attias, Y., Steenkiste, P., Faloutsos, C.: Analysis of the reputation system and user contributions on a question answering website: stackoverflow. In: ASONAM 2013 (2013)Google Scholar
  17. 17.
    Tan, C.H., Agichtein, E., Ipeirotis, P., Gabrilovich, E.: Trust, but verify: predicting contribution quality for knowledge base construction and curation. In: WSDM 2014 (2014)Google Scholar
  18. 18.
    Viswanath, B., Post, A., Gummadi, K.P., Mislove, A.: An analysis of social network-based sybil defenses. SIGCOMM Comput. Commun. Rev. 41, 363–374 (2011)Google Scholar
  19. 19.
    Ye, B., Wang, Y., Liu, L.: Crowd trust: a context-aware trust model for worker selection in crowdsourcing environments. In: ICWS (2015)Google Scholar
  20. 20.
    Bachrach, Y., Graepel, T., Minka, T., Guiver, J.: How to grade a test without knowing the answers - a Bayesian graphical model for adaptive crowdsourcing and aptitude testing. In: ICML 2012 (2012)Google Scholar
  21. 21.
    Yu, H., Shen, Z., Leung, C.: Bringing reputation-awareness into crowdsourcing. In: ICICS 2013 (2013)Google Scholar

Copyright information

© Springer Nature Switzerland AG 2019

Authors and Affiliations

  • Safina Showkat Ara
    • 1
    Email author
  • Subhasis Thakur
    • 1
  • John G. Breslin
    • 1
  1. 1.Insight Centre for Data AnalyticsNUI GalwayGalwayIreland

Personalised recommendations