Information Sciences 375 (2017) 138–154
Contents lists available at ScienceDirect
Information Sciences
journal homepage: www.elsevier.com/locate/ins
Judgment analysis of crowdsourced opinions using
biclustering
Sujoy Chatterjee
a
, Malay Bhattacharyya
b,∗
a
Department of Computer Science and Engineering, University of Kalyani, Nadia – 741235, India
b
Department of Information Technology, Indian Institute of Engineering Science and Technology, Shibpur, Howrah – 711103, India
a r t i c l e i n f o
Article history:
Received 13 October 2015
Revised 1 July 2016
Accepted 14 September 2016
Available online 28 September 2016
Keywords:
Judgment analysis
Opinion ensemble
Majority voting
Biclustering
a b s t r a c t
Annotation by the crowd workers serving online is gaining focus in recent years in diverse
fields due to its distributed power of problem solving. Distributing the labeling task among
a large set of workers (may be experts or non-experts) and obtaining the final consensus is
a popular way of performing large-scale annotation in a limited time. Collection of multi-
ple annotations can be effective for annotation of large-scale datasets for applications like
natural language processing, image processing, etc. However, as the crowd workers are not
necessarily experts, their opinions might not be accurate enough. This causes problem in
deriving the final aggregated judgment. Again, majority voting (MV) is not suitable for such
problems because the number of annotators is limited and they have multiple options to
choose. This might cause too much conflicts among the opinions provided. Additionally,
there might exist annotators who randomly try to annotate (provide spam opinions for)
too many questions to maximize their payment. This can incorporate noise while deriving
the final judgment. In this paper, we address the problem of crowd judgment analysis in
an unsupervised way and a biclustering-based approach is proposed to obtain the judg-
ments appropriately. The effectiveness of this approach is demonstrated on four publicly
available small-scale Amazon Mechanical Turk datasets, along with a large-scale Crowd-
Flower dataset. We also compare the algorithm with MV and some other existing algo-
rithms. In most of the cases the proposed approach is competitively better than others.
But most importantly, it does not use the entire dataset for deriving the judgment.
© 2016 Elsevier Inc. All rights reserved.
1. Introduction
Crowdsourcing is one of the emerging fields that has been shown to have wide applications in the areas of data mining,
machine learning, bioinformatics, etc. [17,18,25]. It provides the new opportunities to tackle diverse real-life problems using
the united power of independent crowd workers. The recent popularity of online crowdsourcing services has also become
useful for managing large-scale labeling tasks. The crowdsourcing model was formally introduced by Howe in 2006 [4,9].
However, Amazon Mechanical Turk (AMT) was the first crowd-powered system that appeared in 2005 to successfully solve
diverse problems using the online crowd workers. Other than AMT, various crowdsourcing platforms like WikiProjects, Kick-
start, Crowdfyerged, etc. are the most popular tools. These crowdsourcing platforms can be categorized into two main types
∗
Corresponding author.
E-mail addresses: sujoy@klyuniv.ac.in (S. Chatterjee), malaybhattacharyya@it.iiests.ac.in, malaybhattacharyya@gmail.com (M. Bhattacharyya).
http://dx.doi.org/10.1016/j.ins.2016.09.036
0020-0255/© 2016 Elsevier Inc. All rights reserved.