Encyclopedia of Database Systems

2018 Edition
| Editors: Ling Liu, M. Tamer Özsu

Crowd Mining and Analysis

  • Yael Amsterdamer
  • Tova Milo
Reference work entry
DOI: https://doi.org/10.1007/978-1-4614-8265-9_80657

Synonyms

Crowd data mining, Crowd mining

Definition

Crowd mining is the process of identifying data patterns in human knowledge, with the assistance of a crowd of web users. The focus is on domain areas in which data is partially or entirely undocumented and where humans are the main source of knowledge, such as data that involves people’s habits, experiences, and opinions. A key challenge in mining such data is that the human knowledge forms an open world and it is thus difficult to know what kind of information one should be looking for.

In classic databases, a similar problem is addressed by data mining techniques that identify interesting patterns in recorded data such as relational databases or textual documents. These techniques, however, are not suitable for the crowd. This is mainly due to properties of the human memory, such as the tendency to remember simple trends and summaries rather than exact details, which should be taken into consideration when gathering and analyzing...

This is a preview of subscription content, log in to check access.

Recommended Reading

  1. 1.
    Doan A, Franklin M, Kossmann D, Kraska T. Crowdsourcing applications and platforms: a data management perspective. Proc VLDB Endowment. 2011;4(12):1508–9.Google Scholar
  2. 2.
    Davidson SB, Khanna S, Milo T, Roy S. Using the crowd for top-k and group-by queries. In: Proceedings of the 16th International Conference on Database Theory; 2013. p. 225–36.Google Scholar
  3. 3.
    Franklin MJ, Kossmann D, Kraska T, Ramesh S, Xin R. CrowdDB: answering queries with crowdsourcing. In: Proceedings of the ACM SIGMOD International Conference on Management of Data; 2011. p. 61–72.Google Scholar
  4. 4.
    Marcus A, Wu E, Karger DR, Madden S, Miller RC. Human-powered Sorts and Joins. Proc VLDB Endowment. 2011;5(1):13–24.CrossRefGoogle Scholar
  5. 5.
    Parameswaran AG, Park H, Garcia-Molina H, Polyzotis N, Widom J. Deco: declarative crowdsourcing. In: Proceedings of the 21st ACM International Conference on Information and Knowledge Management; 2012. p. 1203–12.Google Scholar
  6. 6.
    Trushkowsky B, Kraska T, Franklin MJ, Sarkar P. Crowdsourced enumeration queries. In: Proceedings of the 29th International Conference on Data Engineering; 2013. p. 673–84.Google Scholar
  7. 7.
    Venetis P, Garcia-Molina H, Huang K, Polyzotis N. Max algorithms in crowdsourcing environments. In: Proceedings of the 21st International World Wide Web Conference; 2012. p. 989–98.Google Scholar
  8. 8.
    Amsterdamer Y, Grossman Y, Milo T, Senellart P. Crowd mining. In: Proceedings of the ACM SIGMOD International Conference on Management of Data; 2013. p. 241–52.Google Scholar
  9. 9.
    Amarilli A, Amsterdamer Y, Milo T. On the complexity of mining itemsets from the crowd using taxonomies. In: Proceedings of the 17th International Conference on Database Theory; 2014. p. 15–25.Google Scholar
  10. 10.
    Amsterdamer Y, Davidson SB, Milo T, Novgorodov S, Somech A. OASSIS: query driven crowd mining yael. In: Proceedings of the ACM SIGMOD International Conference on Management of Data; 2014. p. 1–12.Google Scholar
  11. 11.
    Bradburn NM, Rips LJ, Shevell SK. Answering autobiographical questions: the impact of memory and inference on surveys. Science. 1987;236(4798): 158–61.CrossRefGoogle Scholar
  12. 12.
    Srikant R, Agrawal R. Mining generalized association rules. In: Proceedings of the 21st International Conference on Very Large Data Bases; 1995. p. 407–19.Google Scholar

Copyright information

© Springer Science+Business Media, LLC, part of Springer Nature 2018

Authors and Affiliations

  1. 1.Department of Computer ScienceBar Ilan UniversityRamat GanIsrael
  2. 2.School of Computer ScienceTel Aviv UniversityTel AvivIsrael