Skip to main content

Towards an Impact-Driven Quality Control Model for Imbalanced Crowdsourcing Tasks

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 10041))

Abstract

Crowdsourcing have been gaining increasing popularity as a highly distributed digital solution that surpasses both borders and time-zones. Moreover, it extends economic opportunities to developing countries, thus answering the call of impact sourcing in alleviating the welfare of poor labor in need. Nevertheless, it is constantly criticized for the associated quality problems and risks. Attempting to mitigate these risks, a rich body of research has been dedicated to design countermeasures against free riders and spammers, who compromise the overall quality of the results, and whose undetected presence ruins the financial prospects for other honest workers. Such quality risks materialize even more severely with imbalanced crowdsourcing tasks. In fact, while surveying this literature, a common rule of thumb can be indeed derived: the easier it is to cheat the system and go undetected, the more restrictive and across-the-board discriminating countermeasures are taken. Hence, also honest yet low-skilled workers will be placed on par with spammers, and consequently exposed and deprived of much-needed earnings. Therefore in this paper, we argue for an impact-driven quality control model, which fulfills the impact-sourcing vision, thus materializing the social responsibility aspect of crowdsourcing, while ensuring high quality results.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    http://clickers.micromappers.org/.

  2. 2.

    http://www.tomnod.com/.

  3. 3.

    http://www.zentralblatt-math.org/zbmath/.

References

  1. Sorokin, A., Forsyth, D.: Utility data annotation with Amazon Mechanical Turk. In: Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Anchorage, AK. IEEE (2008)

    Google Scholar 

  2. Lofi, C., Selke, J., Balke, W.T.: Information extraction meets crowdsourcing: a promising couple. Datenbank-Spektrum 12(2), 109–120 (2012)

    Article  Google Scholar 

  3. Kouloumpis, E., Wilson, T., Moore, J.: Twitter sentiment analysis: the good the bad and the OMG! In: International Conference on Weblogs & Social Media, Barcelona, Spain (2011)

    Google Scholar 

  4. Selke, J., Lofi, C., Balke, W.-T.: Pushing the boundaries of crowd-enabled databases with query-driven schema expansion. In: International Conference on Very Large Data Bases (VLDB), Istanbul, Turkey (2012)

    Google Scholar 

  5. Ipeirotis, P.G., Provost, F., Wang, J.: Quality management on Amazon mechanical turk. In: ACM SIGKDD on Human Computation Workshop (HCOMP), New York, USA (2010)

    Google Scholar 

  6. El Maarry, K., Güntzer, U., Balke, W.-T.: Realizing impact sourcing by adaptive gold questions: a socially responsible measure for workers’ trustworthiness. In: International Conference on Web-Age Information Management (WAIM), Qingdao, Shandong, China (2015)

    Google Scholar 

  7. El Maarry, K., Balke, W.-T.: Retaining rough diamonds: towards a fairer elimination of low-skilled workers. In: International Conference on Database Systems for Advanced Applications (DASFAA), Hanoi, Vietnam (2015)

    Google Scholar 

  8. El Maarry, K., Güntzer, U., Balke, W.-T.: A majority of wrongs doesn’t make it right. In: Conference on Web Information Systems Engineering (WISE), Miami, USA (2015)

    Google Scholar 

  9. Wang, J., Ipeirotis, P.G., Provost, F.: Managing crowdsourced workers. In: winter Conference on Business Intelligence, Salt Lake City, Utah, USA (2011)

    Google Scholar 

  10. Dawid, P., Skene, A.M.: Maximum likelihood estimation of observer error-rates using the EM algorithm. J. Royal Stat. Soc. 28(1) (1979)

    Google Scholar 

  11. Raykar, V.C., Yu, S., Zhao, L.H., Valadez, G.H., Florin, C., Bogoni, L., Moy, L.P: Learning from crowds. J. Mach. Learn. Res. 11 (2010)

    Google Scholar 

  12. Whitehill, J., Ruvolo, P., Wu, T., Bergsma, J., Movellan, J.: Whose vote should count more: optimal integration of labels from labelers of unknown expertise. In: Advanced Neural Information Processing Systems (NIPS), Vancouver, Canada (2009)

    Google Scholar 

  13. Kuncheva, L.I., Whitaker, C.J., Shipp, C.A., Duin, R.P.W.: Limits on the majority vote accuracy in classifier fusion. J. Pattern Anal. Appl. 6(1), 22–31 (2003)

    Article  MathSciNet  MATH  Google Scholar 

  14. El Maarry, K., Balke, W.-T., Cho, H., Hwang, S., Baba, Y.: Skill ontology-based model for Quality Assurance in Crowdsourcing. In: International Conference on Database Systems for Advanced Applications (DASFAA), Uncrowd Workshop, Bali, Indonesia (2014)

    Google Scholar 

  15. Noorian, Z., Ulieru, M.: The state of the art in trust and reputation systems: a framework for comparison. Journal of theoretical and applied electronic commerce research 5(2), 97–117 (2010)

    Article  Google Scholar 

  16. Ignjatovic, A., Foo, N., Lee, C.T.: An analytic approach to reputation ranking of participants in online transactions. In: IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology, Sydney, Australia (2008)

    Google Scholar 

  17. Liu, X., Lu, M., Ooi, B.C., Shen, Y., Wu, S., Zhang, M.: CDAS: a crowdsourcing data analytics system. VLDB Endowment 5(10), 1040–1051 (2012)

    Article  Google Scholar 

  18. Yu, B., Singh, M.P.: Detecting deception in reputation management. In: International Joint Conference on Autonomous Agents and Multiagent Systems, Melbourne, VIC, Australia (2003)

    Google Scholar 

  19. Daltayanni, M., de Alfaro, L., Papadimitriou, P.: WorkerRank: Using employer implicit judgements to infer worker reputation. In: ACM International Conference on Web Search and Data Mining (WSDM), Shanghai, China (2015)

    Google Scholar 

  20. Hossain, M.: Users’ motivation to participate in online crowdsourcing platforms. In: Conference on Innovation, Management and Technology Research, Malacca, Malaysia (2012)

    Google Scholar 

  21. Kazai, G.: In search of quality in crowdsourcing for search engine evaluation. In: European Conference on Advances in Information Retrieval, Dublin, Ireland (2011)

    Google Scholar 

  22. Boim, R., Greenshpan, O., Milo, T., Novgorodov, S., Polyzotis, N., Tan, W.C.: Asking the right questions in crowd data sourcing. In: International Conference on Data Engineering, Washington, DC, USA (2012)

    Google Scholar 

  23. Altman, G., Bland, J.M.: Diagnostic tests. 1: sensitivity and specificity. British Med. J. 308(6943), 1552 (1994). (Clinical research edition)

    Article  Google Scholar 

  24. Altman, G., Bland, J.M.: Diagnostic tests 2: predictive values. British Med. J. 309(6947), 102 (1994). (Clinical research edition)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kinda El Maarry .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing AG

About this paper

Cite this paper

Maarry, K.E., Balke, WT. (2016). Towards an Impact-Driven Quality Control Model for Imbalanced Crowdsourcing Tasks. In: Cellary, W., Mokbel, M., Wang, J., Wang, H., Zhou, R., Zhang, Y. (eds) Web Information Systems Engineering – WISE 2016. WISE 2016. Lecture Notes in Computer Science(), vol 10041. Springer, Cham. https://doi.org/10.1007/978-3-319-48740-3_9

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-48740-3_9

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-48739-7

  • Online ISBN: 978-3-319-48740-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics