This document discusses active learning from crowdsourced data. It describes challenges with crowdsourcing including varying annotator expertise, noisy labels, and limited budgets. It then summarizes several papers that address these challenges through techniques such as probabilistic multi-labeler models, optimal selection of training points and annotators, and considering annotator accuracy and sample difficulty. The document provides details on models, algorithms, and extensions to other tasks such as multi-label learning and deep learning from crowds.
Related topics: