arxivst stuff from arxiv that you should probably bookmark

Understanding Task Design Trade-offs in Crowdsourced Paraphrase Collection

Abstract · Apr 19, 2017 14:41 ·

offs diversity generation correctness workers paraphrases worker prompt paraphrase cs-cl cs-hc

Arxiv Abstract

  • Youxuan Jiang
  • Jonathan K. Kummerfeld
  • Walter S. Laseck

Linguistically diverse datasets are critical for training and evaluating robust machine learning systems, but data collection is a costly process that often requires experts. Crowdsourcing the process of paraphrase generation is an effective means of expanding natural language datasets, but there has been limited analysis of the trade-offs that arise when designing tasks. In this paper, we present the first systematic study of the key factors in crowdsourcing paraphrase collection. We consider variations in instructions, incentives, data domains, and workflows. We manually analyzed paraphrases for correctness, grammaticality, and linguistic diversity. Our observations provide new insight into the trade-offs between accuracy and diversity in crowd responses that arise as a result of task design, providing guidance for future paraphrase generation procedures.

Read the paper (pdf) »