skip to main content
10.1145/2675133.2675214acmconferencesArticle/Chapter ViewAbstractPublication PagescscwConference Proceedingsconference-collections
research-article

Flock: Hybrid Crowd-Machine Learning Classifiers

Published:28 February 2015Publication History

ABSTRACT

We present hybrid crowd-machine learning classifiers: classification models that start with a written description of a learning goal, use the crowd to suggest predictive features and label data, and then weigh these features using machine learning to produce models that are accurate and use human-understandable features. These hybrid classifiers enable fast prototyping of machine learning models that can improve on both algorithm performance and human judgment, and accomplish tasks where automated feature extraction is not yet feasible. Flock, an interactive machine learning platform, instantiates this approach. To generate informative features, Flock asks the crowd to compare paired examples, an approach inspired by analogical encoding. The crowd's efforts can be focused on specific subsets of the input space where machine-extracted features are not predictive, or instead used to partition the input space and improve algorithm performance in subregions of the space. An evaluation on six prediction tasks, ranging from detecting deception to differentiating impressionist artists, demonstrated that aggregating crowd features improves upon both asking the crowd for a direct prediction and off-the-shelf machine learning features by over 10%. Further, hybrid systems that use both crowd-nominated and machine-extracted features can outperform those that use either in isolation.

References

  1. Crowdflower. http://www.crowdflower.com.Google ScholarGoogle Scholar
  2. Anderson, A., Huttenlocher, D., Kleinberg, J., and Leskovec, J. Discovering value from community activity on focused question answering sites: a case study of stack overflow. In Proc. KDD (2012). Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. André, P., Kittur, A., and Dow, S. P. Crowd synthesis: Extracting categories and clusters from complex data. In Proc. CSCW (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Attenberg, J., Ipeirotis, P. G., and Provost, F. J. Beat the machine: Challenging workers to find the unknown unknowns. In Proc. AAAIW (2011).Google ScholarGoogle Scholar
  5. Blumenstock, J. E. Automatically assessing the quality of wikipedia articles. UCB School of Information (2008).Google ScholarGoogle Scholar
  6. Branson, S., Wah, C., Schroff, F., Babenko, B., Welinder, P., Perona, P., and Belongie, S. Visual recognition with humans in the loop. In Proc. ECCV. 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Chilton, L. B., Little, G., Edge, D., Weld, D. S., and Landay, J. A. Cascade: Crowdsourcing taxonomy creation. In Proc. CHI (2013). Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Dai, P., Weld, D. S., et al. Decision-theoretic control of crowd-sourced workflows. In Proc. AAAI (2010).Google ScholarGoogle Scholar
  9. Deng, J., Krause, J., and Fei-Fei, L. Fine-grained crowdsourcing for fine-grained recognition. In Proc. CVPR (2013). Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Domingos, P. A few useful things to know about machine learning. Communications of the ACM (2012). Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Fails, J. A., and Olsen Jr, D. R. Interactive machine learning. In Proc. IUI (2003). Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Fogarty, J., Tan, D., Kapoor, A., and Winder, S. Cueflik: interactive concept learning in image search. In Proc. CHI (2008). Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Gazzaniga, M. S., LeDoux, J. E., Gazzaniga, M., and LeDoux, J. The integrated mind. 1978.Google ScholarGoogle ScholarCross RefCross Ref
  14. Gentner, D., Loewenstein, J., and Thompson, L. Learning and transfer: A general role for analogical encoding. Journal of Educational Psychology (2003).Google ScholarGoogle Scholar
  15. Gentner, D., and Markman, A. B. Structure mapping in analogy and similarity. American Psychologist (1997).Google ScholarGoogle Scholar
  16. Gomes, R. G., Welinder, P., Krause, A., and Perona, P. Crowdclustering. In Advances in Neural Information Processing Systems (2011), 558--566.Google ScholarGoogle Scholar
  17. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., and Witten, I. H. The weka data mining software: an update. ACM SIGKDD Explorations (2009). Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Hammond, K. R., Hursch, C. J., and Todd, F. J. Analyzing the components of clinical inference. Psychological review (1964).Google ScholarGoogle Scholar
  19. Hartmann, B., Abdulla, L., Mittal, M., and Klemmer, S. R. Authoring sensor-based interactions by demonstration with direct manipulation and pattern recognition. In Proc. CHI (2007). Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Kahneman, D. A perspective on judgment and choice: mapping bounded rationality. American psychologist (2003).Google ScholarGoogle Scholar
  21. Kamar, E., Kapoor, A., and Horvitz, E. Lifelong learning for acquiring the wisdom of the crowd. In Proc. JCAI (2013). Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Kim, J., Cheng, J., and Bernstein, M. S. Ensemble: Exploring complementary strengths of leaders and crowds in creative collaboration. In Proc. CSCW (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Kittur, A., Peters, A. M., Diriye, A., and Bove, M. Standing on the schemas of giants: socially augmented information foraging. In Proc. CSCW (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Lasecki, W. S., Song, Y. C., Kautz, H., and Bigham, J. P. Real-time crowd labeling for deployable activity recognition. In Proc. CSCW (2013). Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Law, E. Attribute Learning using Joint Human and Machine Computation. PhD thesis, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Law, E., Settles, B., Snook, A., Surana, H., Von Ahn, L., and Mitchell, T. Human computation for attribute and attribute value acquisition. In FPGV (2011).Google ScholarGoogle Scholar
  27. Law, E., and Von Ahn, L. Input-agreement: a new mechanism for collecting data using human computation games. In Proc. CHI, ACM (2009). Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Le, J., Edmonds, A., Hester, V., and Biewald, L. Ensuring quality in crowdsourced search relevance evaluation: The effects of training question distribution. In Proc. SIGIR (2010).Google ScholarGoogle Scholar
  29. Li, H., Zhao, B., and Fuxman, A. The wisdom of minority: Discovering and targeting the right group of workers for crowdsourcing. In Proc. WWW (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Lowe, D. G. Object recognition from local scale-invariant features. In ICCV (1999). Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Ott, M., Choi, Y., Cardie, C., and Hancock, J. T. Finding deceptive opinion spam by any stretch of the imagination. In Proc. ACL (2011). Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Parameswaran, A., Sarma, A. D., Garcia-Molina, H., Polyzotis, N., and Widom, J. Human-assisted graph search: it's okay to ask questions. Proc. VLDB (2011). Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. Parikh, D., and Grauman, K. Interactively building a discriminative vocabulary of nameable attributes. In CVPR (2011). Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Patel, K., Bancroft, N., Drucker, S. M., Fogarty, J., Ko, A. J., and Landay, J. Gestalt: integrated support for implementation and analysis in machine learning. In Proc. UIST (2010). Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. Patel, K., Fogarty, J., Landay, J. A., and Harrison, B. Investigating statistical machine learning as a tool for software development. In Proc. CHI (2008). Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. Raykar, V. C., Yu, S., Zhao, L. H., Valadez, G. H., Florin, C., Bogoni, L., and Moy, L. Learning from crowds. JMLR (2010). Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Rzeszotarski, J., and Kittur, A. Crowdscape: interactively visualizing user behavior and output. In Proc. UIST (2012). Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. Rzeszotarski, J. M., and Kittur, A. Instrumenting the crowd: using implicit behavioral measures to predict task performance. In Proc. UIST (2011). Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. Settles, B. Closing the loop: Fast, interactive semi-supervised annotation with queries on features and instances. In Proc. EMNLP (2011). Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Sheng, V. S., Provost, F., and Ipeirotis, P. G. Get another label- improving data quality and data mining using multiple, noisy labelers. In Proc. KDD (2008). Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. Skinner, B. F. Reinforcement today. American Psychologist (1958).Google ScholarGoogle Scholar
  42. Snow, R., O'Connor, B., Jurafsky, D., and Ng, A. Y. Cheap and fast - but is it good?: evaluating non-expert annotations for natural language tasks. In Proc. EMNLP (2008). Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. Talbot, J., Lee, B., Kapoor, A., and Tan, D. S. Ensemblematrix: interactive visualization to support machine learning with multiple classifiers. In Proc. CHI (2009). Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. Vijayanarasimhan, S., and Grauman, K. Large-scale live active learning: Training object detectors with crawled data and crowds. In CVPR (2011). Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. Wais, P., Lingamneni, S., Cook, D., Fennell, J., Goldenberg, B., Lubarov, D., Marin, D., and Simons, H. Towards building a high-quality workforce with mechanical turk. In Proc. NIPSCSS (2010).Google ScholarGoogle Scholar
  46. Whitehill, J., Ruvolo, P., Wu, T., Bergsma, J., and Movellan, J. R. Whose vote should count more: Optimal integration of labels from labelers of unknown expertise. In Proc. NIPS (2009).Google ScholarGoogle Scholar
  47. Yu, L., Kittur, A., and Kraut, R. E. Distributed analogical idea generation: Inventing with crowds. In Proc. CHI (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. Yu, L., and Nickerson, J. V. Cooks or cobblers': crowd creativity through combination. In Proc. CHI (2011). Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. Parkash, A., and Parikh, D. Attributes for classifier feedback. In Proc. ECCV. 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Flock: Hybrid Crowd-Machine Learning Classifiers

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      CSCW '15: Proceedings of the 18th ACM Conference on Computer Supported Cooperative Work & Social Computing
      February 2015
      1956 pages
      ISBN:9781450329224
      DOI:10.1145/2675133

      Copyright © 2015 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 28 February 2015

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      CSCW '15 Paper Acceptance Rate161of575submissions,28%Overall Acceptance Rate2,235of8,521submissions,26%

      Upcoming Conference

      CSCW '24

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader