[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/2505515.2505755acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

Optimizing plurality for human intelligence tasks

Published: 27 October 2013 Publication History

Abstract

In a crowdsourcing system, Human Intelligence Tasks (HITs) (e.g., translating sentences, matching photos, tagging videos with keywords) can be conveniently specified. HITs are made available to a large pool of workers, who are paid upon completing the HITs they have selected. Since workers may have different capabilities, some difficult HITs may not be satisfactorily performed by a single worker. If more workers are employed to perform a HIT, the quality of the HIT's answer could be statistically improved. Given a set of HITs and a fixed "budget", we address the important problem of determining the number of workers (or plurality) of each HIT so that the overall answer quality is optimized. We propose a dynamic programming (DP) algorithm for solving the plurality assignment problem (PAP). We identify two interesting properties, namely, monotonicity and diminishing return, which are satisfied by a HIT if the quality of the HIT's answer increases monotonically at a decreasing rate with its plurality. We show for HITs that satisfy the two properties (e.g., multiple-choice-question HITs), the PAP is approximable. We propose an efficient greedy algorithm for such case. We conduct extensive experiments on synthetic and real datasets to evaluate our algorithms. Our experiments show that our greedy algorithm provides close-to-optimal solutions in practice.

References

[1]
Amazon Web Services LLC. Amazon Mechanical Turk: Best Practices Guide. http://mturkpublic.s3.amazonaws.com/docs/MTURK_BP.pdf.
[2]
D. W. Barowy et al. Automan: A platform for integrating human-based and digital computation. OOPSLA, 2012.
[3]
C. Cao and J. She et al. Whom to ask? Jury selection for decision making tasks on micro-blog services. VLDB, 2012.
[4]
E. Dale and J. S. Chall. A formula for predicting readability. Educational research bulletin, 1948.
[5]
S. Dowdy. Statistics for Research. Wiley, 1983.
[6]
M. J. Franklin et al. CrowdDB: answering queries with crowdsourcing. In SIGMOD, 2011.
[7]
S. Guo and A. Parameswaran et al. So who won? dynamic max discovery with the crowd. SIGMOD, 2012.
[8]
X. Liu, M. Lu, B. Ooi, Y. Shen, S. Wu, and M. Zhang. CDAS: A crowdsourcing data analytics system. VLDB, 2012.
[9]
A. Marcus, E. Wu, D. Karger, S. Madden, and R. Miller. Human-powered sorts and joins. VLDB, 2011.
[10]
A. Marcus, E. Wu, S. Madden, and R. Miller. Crowdsourced databases: Query processing with people. CIDR, 2011.
[11]
A. Marcus and E. Wu et al. Demonstration of qurk: A query processor for human operators. SIGMOD, 2011.
[12]
A. Marshall. 1920. Principles of economics, 8, 1890.
[13]
S. Martello and P. Toth. Knapsack problems: algorithms and computer implementations. John Wiley & Sons, Inc., 1990.
[14]
L. Mo, R. Cheng, B. Kao, and X. Yang et al. Optimizing plurality for human intelligence tasks. Technical Report. http://www.cs.hku.hk/research/techreps/document/TR-2013-01.pdf.
[15]
B. Mozafari et al. Active learning for crowd-sourced databases. arXiv preprint arXiv:1209.3686, 2012.
[16]
A. Parameswaran and H. Garcia-Molina et al. Crowdscreen: Algorithms for filtering data with humans. SIGMOD, 2012.
[17]
A. Parameswaran and A. Sarma et al. Human-assisted graph search: it's okay to ask questions. VLDB, 2011.
[18]
T. Saaty. Mathematical methods of operations research. Dover Publications, 2004.
[19]
B. Trushkowsky, T. Kraska, M. J. Franklin, and P. Sarkar. Crowdsourced enumeration queries. ICDE, 2013.
[20]
J. Wang, T. Kraska, M. Franklin, and J. Feng. Crowder: crowdsourcing entity resolution. VLDB, 2012.
[21]
S. Whang et al. Question selection for crowd entity resolution. Technical report, Stanford InfoLab, 2012.
[22]
X. Yang et al. On incentive-based tagging. ICDE, 2013.

Cited By

View all
  • (2022)A Survey on Task Assignment in CrowdsourcingACM Computing Surveys10.1145/349452255:3(1-35)Online publication date: 3-Feb-2022
  • (2022)Multi-task versus consecutive task allocation with tasks clustering for Mobile Crowd Sensing SystemsProcedia Computer Science10.1016/j.procs.2021.12.212198:C(67-76)Online publication date: 1-Jan-2022
  • (2020)A Crowd-Sensing Framework for Allocation of Time-Constrained and Location-Based TasksIEEE Transactions on Services Computing10.1109/TSC.2017.272583513:5(769-785)Online publication date: 1-Sep-2020
  • Show More Cited By

Index Terms

  1. Optimizing plurality for human intelligence tasks

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    CIKM '13: Proceedings of the 22nd ACM international conference on Information & Knowledge Management
    October 2013
    2612 pages
    ISBN:9781450322638
    DOI:10.1145/2505515
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 27 October 2013

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. crowdsourcing
    2. data quality

    Qualifiers

    • Research-article

    Conference

    CIKM'13
    Sponsor:
    CIKM'13: 22nd ACM International Conference on Information and Knowledge Management
    October 27 - November 1, 2013
    California, San Francisco, USA

    Acceptance Rates

    CIKM '13 Paper Acceptance Rate 143 of 848 submissions, 17%;
    Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

    Upcoming Conference

    CIKM '25

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)5
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 13 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2022)A Survey on Task Assignment in CrowdsourcingACM Computing Surveys10.1145/349452255:3(1-35)Online publication date: 3-Feb-2022
    • (2022)Multi-task versus consecutive task allocation with tasks clustering for Mobile Crowd Sensing SystemsProcedia Computer Science10.1016/j.procs.2021.12.212198:C(67-76)Online publication date: 1-Jan-2022
    • (2020)A Crowd-Sensing Framework for Allocation of Time-Constrained and Location-Based TasksIEEE Transactions on Services Computing10.1109/TSC.2017.272583513:5(769-785)Online publication date: 1-Sep-2020
    • (2019)FROG: A Fast and Reliable Crowdsourcing FrameworkIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2018.284939431:5(894-908)Online publication date: 1-May-2019
    • (2018)Understanding Crowdsourcing Systems from a Multiagent Perspective and ApproachACM Transactions on Autonomous and Adaptive Systems10.1145/322602813:2(1-32)Online publication date: 31-Jul-2018
    • (2018)Online Sequencing of Non-Decomposable Macrotasks in Expert CrowdsourcingACM Transactions on Social Computing10.1145/31404591:1(1-33)Online publication date: 10-Jan-2018
    • (2017)VoxPLProceedings of the 2017 CHI Conference on Human Factors in Computing Systems10.1145/3025453.3026025(2347-2358)Online publication date: 2-May-2017
    • (2016)Using Hierarchical Skills for Optimized Task Assignment in Knowledge-Intensive CrowdsourcingProceedings of the 25th International Conference on World Wide Web10.1145/2872427.2883070(843-853)Online publication date: 11-Apr-2016
    • (2016)Crowdsourced Data ManagementIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2016.253524228:9(2296-2319)Online publication date: 1-Sep-2016
    • (2015)QASCAProceedings of the 2015 ACM SIGMOD International Conference on Management of Data10.1145/2723372.2749430(1031-1046)Online publication date: 27-May-2015
    • Show More Cited By

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media