Datasets on Crowdsourcing
Dataset Name (with link) Size: #questions, #answers, ratio (#answers/#questions) #Questions with Ground Truth Application Quuestion Content: Image or Text Question Type: Single Choice (N choose 1), or Rating Operator(s) Description
Part 1: Datasets with ground truth and workers' answers
Fact Evaluation Judgment Dataset 42624, 216725, 5.08 576 Fact Evaluation Text 3 choose 1  Selection, Join The task is to identify whether a fact (e.g., "Stephen Hawking graduated from Oxford") is correct, wrong, or ambiguous.
Fashion 10000: An Enriched Social Image Dataset for Fashion and Clothing 32398, 97194, 3 32398 Image Retrieval Image with Metadata (e.g., Tags) 2 choose 1  Selection, Collection, Join The task identifies whether or not an image is fashion-related (Note: the downloaded dataset size is 9.8G).
Sentiment Popularity 500, 10000, 20 500 Sentiment Analysis Text 2 choose 1  Selection, Join The task  aims at classifying movie reviews as either positive or negative.
Weather Sentiment 300, 6000, 20 300 Sentiment Analysis Text 5 choose 1 Selection, Join The task is to judge the sentiment of a tweet discussing the weather ("negative", "neutral", "positive", "unrelated to weather" or "cannot tell").
Face Sentiment 584, 5256, 9 584 Face Sentiment Identification Image 4 choose 1 Selection, Categorize, Join The task is to identify the sentiment (on whether it is neutral, happy, sad or angry) for a given face image.
Relevance Finding 20232, 98453, 4.87 3276 Relevance Finding Text Rating (5 choices) Selection, Categorize, Sort, Top-K, Join The task is to identify the relevance of a given topic and a given document in a 5-level rating, i,e, 2: highly relevant, 1: relevant, 0: non-relevant, -1: unknown, -2: broken link.
Dunchenne Smile Identification 2162, 30319, 14.02 160 Dunchenne Smile Identification Image   2 choose 1  Selection, Categorize, Join The task is to judge whether or not a smile (on an image) is a Dunchenne Simile.
HITSpam-Crowdflower 5380, 42762, 7.95 101 Spam Detection Text 2 choose 1  Selection, Join The task is to judge whether or not a HIT is a spam task.
HITSpam-Mturk 5840, 28354, 4.86 101 Spam Detection Text 2 choose 1  Selection, Join The task is to judge whether or not a HIT is a spam task.
Query Document Relevance 2165, 17395, 8.03 2165 Relevance Finding Text 2 choose 1  Selection, Join The task is to identify wehther a given query and a given document is relevant or not.
AdultContent 11040, 92721, 8.40 1517 Classification Text Rating (5 choices) Selection, Categorize, Sort, Top-K, Join The task is to identify the adult level of a website (G, P, R, X, B).
Emotion 700, 7000, 10 700 Emotion Rating Text Rating (choose a value from -100 to 100) Selection, Sort, Top-K, Join The task is to rate the emotion of a given text. There are 7 emotions (anger, disgust, fear, joy, sadness, surprise, valence), and a user gives a value from -100 to 100 for each emotion about the text.
Word pair similarity 30, 300, 10 30 Word Similarity Finding Text Rating (choose a numerical score from 0 to 10) Selection, Sort, Top-K, Join The task is to assign a numerical similarity score between 0 and 10 to a given text.
Recognizing Textual Entailment 800, 8000, 10 800 Textual Understanding Text  2 choose 1  Selection, Join The task is to identify whether a given Hypothesis sentence is implied by the information in the given text.
Temporal Ordering 462, 4620, 10 462 Event Ordering Text  2 choose 1  Selection, Sort, Top-K, Join The task is to identify whether or not one event happens before another event in a given context.
Word Sense Disambiguation 177,1770, 10 177 Word Sense Disambiguation Text  K choose 1  Selection, Join The task is to choose the most appropriate sense of a word (out of several given senses) in the given context.
Web search data 2665, 15567, 5.84 2652 Relevance Finding Text Rating (5 choices) Selection, Sort, Top-K, Join The task is to judge the relevance of query-URL pairs with a 5-level rating scale (from 1 to 5).
Duck 108, 4212, 39 108 Duck Identification Image  2 choose 1  Selection, Join The task is to identify whether the image contains a duck or not.
Dog 807, 8070, 10 807 Dog Breed Identification Image  4 choose 1  Selection, Categorize, Join The task is to recognize a breed (out of Norfolk Terrier, Norwich Terrier, Irish Wolfhound, and Scottish Deerhound) for a given dog.
Airline Twitter sentiment 55000,16000,3.4375 16,000 Sentiment Analysis Text K choose 1 Selection, Join A sentiment analysis job about the problems of each major U.S. airline. Twitter data was scraped from February of 2015 and contributors were asked to first classify positive, negative, and neutral tweets, followed by categorizing negative reasons (such as "late flight" or "rude service").
Company categorizations (with URLs) 28675,7152,4 7,152 Categorization Text 6 choose 1 Categorization A data set where business names were matched with URLs/homepages for the named businesses.
               
Part 2: Datasets with only ground truth ( no workers' answers )
Clothing pattern identification 14,750 14,750 Classification Image  K choose 1  Selection, Categorize A large dataset where contributors viewed pictures of dresses and classified their patterns. Sixteen popular pattern types were provided and each dress was judged by three contributors. Dataset is provided with their aggregated judgments and URLs for each dress. 
Sound detection and classification 8,000 8,000 Classification Audio  K choose 1  Selection, Categorize Contributors listened to short audio clips and identified white noise events like coughing, dropped keys, and barking dogs. They also tried to identify the scene, such as office, cafe, or supermarket and ranked the difficulty of each individual row. Audio clips range from about five to ten seconds. 
Relevancy of terms to a disaster relief topic 7,566 7,566 Relevance Finding Text 5 choose 1 Selection Contributors viewed a topic and a term and rated the relevancy of the latter to the former on a five point scale (1 being very irrelevant, 5 being very relevant). The topics all center around humanitarian aid or disaster relief and each topic was defined for contributors. They were also asked if the term was a specific person or place and whether it was misspelled.
Hate speech identification 14,442 14,442 Hate Sentiment Identification Text 3 choose 1 Selection, Categorize, Join Contributors viewed short text and identified if it a) contained hate speech, b) was offensive but without hate speech, or c) was not offensive at all. Contains nearly 15K rows with three contributor judgments per text string.
Economic News Article Tone and Relevance 8,000 8,000 Relevance Finding Text K choose 1 Selection, Categorize, Join Contributors read snippets of news articles. They then noted if the article was relevant to the US economy and, if so, what the tone of the article was. Tone was judged on a 9 point scale (from 1 to 9, with 1 representing the most negativity). Dataset contains these judgments as well as the dates, source titles, and text. Dates range from 1951 to 2014.
Football Strategy 3,731 3,731 Football Strategy Text 6 choose 1 Selection Contributors were presented a football scenario and asked to note what the best coaching decision would be. An scenario: "It is third down and 3. The ball is on your opponent's 20 yard line. There are five seconds left. You are down by 4." The decisions presented were punt, pass, run, kick a field goal, kneel down, or don't know. There are thousands of such scenarios in this job.
Numerical Transcription from Images 7,665 7,665 Numerical Transcription from Images Image Numerical Transcription Selection Contributors looked at a series of pictures from a footrace and transcribed bib numbers of the competitors. Some images contain multiple bib numbers or incomplete bib numbers.
Identifying key phrases in text 8,262 8,262 Textual Understanding Text K choose 1 Selection, Join Contributors looked at question/answer pairs (like "When did Bob Marley die? 1981") and a series of sentences surrounding that event (such as phrases from a Bob Marley obituary). They marked which of those sentences spoke directly to the question (such as "Robert Nesta "Bob" Marley, OM (6 February 1945 – 11 May 1981) was a Jamaican reggae singer, song writer, musician") for a wide variety of topics.  
Gender classifier data 20,000 20,000 Gender Identification Text/Image 3 choose 1 Selection, Categorize, Join This data set was used to train a CrowdFlower AI gender predictor. Contributors were asked to simply view a Twitter profile and judge whether the user was a male, a female, or a brand (non-individual). The dataset contains 20,000 rows, each with a user name, a random tweet, account profile and image, location, and even link and sidebar color.
Sentiment analyses of single words or short phrases 3,523 3,523 Sentiment Analysis Text K choose 2 Selection, Join Contributors looked at four words or bigrams (bigrams are just word pairs) and ranked the most positive and most negative ones in each set. For example, they saw quartets like "nasty, failure, honored, females" and chose which word was the most positive and most negative. Interestingly, each set was graded by eight contributors instead of the usual three. Dataset contains all 3,523 rows, but has 28K judgments.
Disasters on social media 10,877 10,877 Fact Evaluation Text 2 choose 1 Selection, Join Contributors looked at over 10,000 tweets culled with a variety of searches like "ablaze", "quarantine", and "pandemonium", then noted whether the tweet referred to a disaster event (as opposed to a joke with the word or a movie review or something non-disastrous).
Do these chemicals contribute to a disease? 5,160 5,160 Fact Evaluation Text 2 choose 1 Selection, Join Contributors read sentences in which both a chemical (like Aspirin) and a disease (or side-effect) were present. They then determined if the chemical directly contributed to the disease or caused it. Dataset includes chemical names, disease name, and aggregated judgments of five (as opposed to the usual three) contributors.
First GOP debate sentiment analysis 14,000 14,000 Sentiment Analysis Text K choose 1 Selection, Join We looked through tens of thousands of tweets about the early August GOP debate in Ohio and asked contributors to do both sentiment analysis and data categorization. Contributors were asked if the tweet was relevant, which candidate was mentioned, what subject was mentioned, and then what the sentiment was for a given tweet. We've removed the non-relevant messages from the uploaded dataset.
URL categorization 31,085 31,085 URL categorization Text K choose 1 Selection, Categorize, Join To create this large, enriched dataset of categorized websites, contributors clicked provided links and selected a main and sub-category for URLs. 
Classification of political social media 5,000 5,000 Classification Text K choose 1 Selection, Categorize, Join Contributors looked at thousands of social media messages from US Senators and other American politicians to classify their content. Messages were broken down into audience (national or the tweeter's constituency), bias (neutral/bipartisan, or biased/partisan), and finally tagged as the actual substance of the message itself (options ranged from informational, announcement of a media appearance, an attack on another candidate, etc.)
eCommerce search relevance 32,000 32,000 Relevance Finding Text Rating Selection, Categorize We used this dataset to launch our Kaggle competition, but the set posted here contains far more information than what served as the foundation for that contest. This set contains image URLs, rank on page, description for each product, search query that lead to each result, and more, each from five major English-language ecommerce sites.
Housing and wheelchair accessibility 10,00 10,00 Fact Evaluation Image Label Selection, Join Here, contributors viewed 10,000 Google maps images and marked whether they were residential areas. If they were, they noted which homes were most prevalent in the area (apartments or houses) and whether the area had proper sidewalks that are wheelchair friendly.
Primary emotions of statements 2,400 2,400 Sentiment Analysis Text 18 choose 1 Selection, Join Contributors looked at a single sentence and rated it's emotional content based on Plutchik's wheel of emotions. 18 emotional choices were presented to contributors for grading.
U.S. economic performance based on news articles 5,000 5,000 Relevance Finding Text Rating(Rating the indication on a scale of 1-9, with 1 being negative and 9 being positive.) Selection, Categorize Contributors viewed a new article headline and a short, bolded excerpt of a sentence or two from the attendant article. Next, they decided if the sentence in question provided an indication of the U.S. economy's health, then rated the indication on a scale of 1-9, with 1 being negative and 9 being positive. 
Police-involved fatalities since May 2013 2,355 2,355 Classification Text/Image K choose 1 Selection, Categorize A data categorization job where contributors compiled a database of police-involved shootings over a two-year span. Information contained includes: race, gender, city, state, whether the victim was armed, photos of the deceased, attending news stories, and more.
Comparing pictures of people 59,476 59,476 Image Retrieval Image Rating (5 choices) Selection, Collection, Join In this job, contributors viewed two pictures of people walking through the same room and were then asked to compare the person on the left to the person on the right.  Questions center on observable traits (like skin color, hair length, muscularity, etc.).
Twitter sentiment analysis: Self-driving cars 7,015 7,015 Sentiment Analysis Text Rating( very positive, slightly positive, neutral, slightly negative, or very negative.) Selection, Join A simple twitter sentiment analysis job where contributors read tweets and classified them as very positive, slightly positive, neutral, slightly negative, or very negative. They were also prompted asked to mark if the tweet was not relevant to self-driving cars.
Blockbuster database 410 410 Categorization Text Survey Selection, Categorize A data categorization job where we asked the crowd to find out information about the ten most popular movies, each year, for the past 40 years (1975-2015).
Government official database 5,000 5,000 Categorization Text K choose 1 Categorization A simple data categorization job wherein contributors viewed a cabinet member, minister, ambassador, etc., and separated their names from their titles. Data set contains names, positions, and years served.
Wikipedia image categorization 976 976 Image Categorization Image K choose 1 Selection, Collection, Join This data set contains hundreds of Wikipedia images which contributors categorized in the following ways:No person present/One person present/Several people present, but one dominant/Several people present, but none are dominant/Unsure. If the images were of one or several people, contributors further classified images by gender.
Image attribute tagging 3,235 3,235 Image Categorization Image K choose 1 Selection, Collection, Join Contributors viewed thousands of images and categorized each based on a given list of attributes. These attributes ranged from objective and specific (like "child" or "motorbike") to more subjective ones (like "afraid" or "beautiful"). Data set includes URLs for all images, multiple tags for each, and contributor agreement scores.
Mobile search relevance 647 647 Relevance Finding Text K choose 1 Selection, Categorize, Join Contributors viewed a variety of searches for mobile apps and determined if the intent of those searches was matched. One was a short query like "music player"; the other, a much longer one like "I would like to download an app that plays the music on the phone from multiple sources like Spotify and Pandora and my library." 
Progressive issues sentiment analysis 1,159 1,159 Sentiment Analysis Text K choose 1 Selection, Join Contributors viewed tweets regarding a variety of left-leaning issues like legalization of abortion, feminism, Hillary Clinton, etc. They then classified if the tweets in question were for, against, or neutral on the issue (with an option for none of the above). After this, they further classified each statement as to whether they expressed a subjective opinion or gave facts. 
Indian terrorism deaths database 27,233 27,233 Fact Evaluation Text Survey Selection, Join Contributors read sentences from the South Asia Terrorism Portal and quantified them. Contributors counted the deaths mentioned in a sentence and whether they were terrorists, civilians, or security forces. Database contains original sentences, state and district in which the deaths occurred, dates of the deaths, and more. (Test questions have been removed from the database for ease of visualization.)
Drug relation database 2,020 2,020 Fact Evaluation Text K choose 1 Selection, Join Contributors read color coded sentences and determined what the relationship of a drug was to certain symptoms or diseases. 
Blurry image comparison 511 511 Relevance Finding Image K choose 1 Selection, Categorize, Join Contributors viewed a pair of purposely blurry or saturated images. They were then asked which image more closely matched a particular word. Data set contains URLs for all images and image pairs, aggregated agreement scores, and variance amounts. Notably, a high number of contributors were polled for each image pairing (20 in total for each, giving this data set upwards of 10,000 judgements).
Objective truths of sentences/concept pairs 8,227 8,227 Fact Evaluation Text Rating (5 choices) Selection, Join Contributors read a sentence with two concepts. For example "a dog is a kind of animal" or "captain can have the same meaning as master." They were then asked if the sentence could be true and ranked it on a 1-5 scale. On the low end was "strongly disagree" and on the upper, "strongly agree."
Image sentiment polarity classification 15,613 15,613 Image Sentiment Analysis Image 5 choose 1 Selection, Join This data set contains over fifteen thousand sentiment-scored images. Contributors were shown a variety of pictures (everything from portraits of celebrities to landscapes to stock photography) and asked to score the images on typical positive/negative sentiment. Data set contains URL of images, sentiment scores of highly positive, positive, neutral, negative, and highly negative, and contributor agreement. 
Smart phone & tablet names database 1,600 1,600 Fact Evaluation Text 2 choose 1 Selection, Join Contributors viewed a particular model code (like C6730 or LGMS323), then searched for the name of the device itself (Kyocera C6730 Hydro or LG Optimus L70), then noted whether the device was a phone or tablet. 
Free text object descriptions 1,225 1,225 Image Retrieval Text Description Selection, Collection, Join Contributors viewed a pair of items and were asked to write sentences that describe and differentiated the two objects. In other words, if viewing an apple and a orange, they could not write "this is a piece of fruit" twice, but needed to note how they were different. Image pairings varied so that the same image would appear in different couples and the second image was always smaller. Data set contains URLs of images and three sentences written per item, per image.
News article / Wikipedia page pairings 3,000 3,000 Relevance Finding Text 2 choose 1 Selection, Categorize, Join Contributors read a short article and were asked which of two Wikipedia articles it matched most closely. For example, a brief biography of Mel Gibson could be paired with Gibson's general Wikipedia page or Lethal Weapon; likewise, Iran election results could be paired with a Wikipedia page on Iran in general or the 2009 protests. Data set contains URLs for both Wiki pages, the full text contributors read, and their judgements on each row. 
Is-A linguistic relationships 3,297 3,297 Fact Evaluation Text 2 choose 1 Selection, Join Contributors were provided a pair of concepts in a constant sentence structure. Namely: [Noun 1] is a [noun 2]. They were then asked to simply note if this sentence was then true or false. Data set contains all nouns and aggregated T/F judgements.
Weather sentiment 1000 1000 Sentiment Analysis Text K choose 1 Selection, Join Here, contributors were asked to grade the sentiment of a particular tweet relating to the weather. The catch is that 20 contributors graded each tweet. We then ran an additional job (the one below) where we asked 10 contributors to grade the original sentiment evaluation.
Weather sentiment evaluated 1000 1000 Sentiment Analysis Text K choose 1 Selection, Join Here, contributors were asked if the crowd graded the sentiment of a particular tweet relating to the weather correctly. The original job (above this one, called simply "Weather sentiment") involved 20 contributors noting the sentiment of weather-related tweets. In this job, we asked 10 contributors to check that original sentiment evaluation for accuracy.
Image classification: People and food 587 587 Image classification Image K choose 1 Selection, Categorize, Join A collection of images of people eating fruits and cakes and other foodstuffs. Contributors classified the images by male/female, then by age (adult or child/teenager). 
"All oranges are lemons," a.k.a. Semantic relationships between two concepts 3,536 3,536 Fact Evaluation Text/Image 2 choose 1 Selection, Join An interesting language data set about the relationship of broad concepts.All questions were phrased in the following way: "All [x] are [y]." For example, a contributor would see something like "All Toyotas are vehicles" and were then asked to say whether this claim was true or false. Contributors were also provided images, in case they were unclear as to what either concept is. This data set includes links to both images provided, the names given for [x] and [y], and whether the statement that "All [x] are [y]" was true or false.
The colors of #TheDress 1,000 1,000 Image Retrieval Image 2 choose 1 Selection, Collection, Join On February 27th, 2015, the internet was briefly obsessed with the color of a dress known simply as #TheDress. We ran a survey job to 1000 contributors and asked them what colors the dress was, as well as looked into a hypothesis that Night Owls and Morning People saw the dress differently.
McDonald's review sentiment 1,500 1,500 Sentiment Analysis Text 8 choose 1 Selection, Join A sentiment analysis of negative McDonald's reviews. Contributors were given reviews culled from low-rated McDonald's from random metro areas and asked to classify why the locations received low reviews.
Gender breakdown of Time Magazine covers <100 <100 Gender Identification Image 2 choose 1 Selection, Categorize, Join Contributors were shown images of Time Magazine covers since the late 1920s and asked to classify if the person was male or female. Data is broken down overall and on an annual basis.
Agreement between long and short sentences 2,000 2,000 Relevance Finding Text 3 choose 1 Selection, Categorize, Join Contributors were asked to read two sentences (the first was an image caption and the second was a shorter version) and judge whether the short sentence adequately describes the event in the first sentence (image caption).
Biomedical image modality 10,652 10,652 Image classification Image K choose 1 Selection, Categorize, Join A large data set of labeled biomedical images, ranging from x-ray and ultrasound to charts, graphs, and even hand-drawn sketches.
Academy Awards demographics 416 416 Fact Retrieval Text/Image K choose 1 Selection, Categorize, Sort, Top-K, Join A data set concerning the race, religion, age, and other demographic details of all Oscars winners since 1928 in the following categories:Best Actor/Best Actress/Best Supporting Actor/Best Supporting Actress/Best Director.
Corporate messaging 3,118 3,118 Categorization Text 3 choose 1 Categorization A data categorization job concerning what corporations actually talk about on social media. Contributors were asked to classify statements as information (objective statements about the company or it's activities), dialog (replies to users, etc.), or action (messages that ask for votes or ask users to click on links, etc.). 
Body part relationships 1,892 1,892 Fact Evaluation Text 2 choose 1 Selection, Join A data set where contributors classified if certain body parts were part of other parts. Questions were phrased like so: "[Part 1] is a part of [part 2]," or, by way of example, "Nose is a part of spine" or "Ear is a part of head."
Wearable technology database 582 582 Fact Evaluation Text K choose 1 Selection, Categorize, Sort, Top-K, Join A data set containing information on hundreds of wearables. Contains data on prices, company name and location, URLs for all wearables, as well as the location of the body on which the wearable is worn. 
Image descriptions 225,000 225,000 Image description Image 2 choose 1 Selection, Categorize, Join Contributors were shown a large variety of images and asked whether a given word described the image shown. For example, they might see a picture of Mickey Mouse and the word Disneyland, where they'd mark "yes." Conversely, if Mickey Mouse's pair word was "oatmeal," they would mark no.
Sentence plausibility 400 400 Fact Evaluation Text Rating (5 choices) Selection, Categorize, Join Contributors read strange sentences and ranked them on a scale of "implausible" (1) to "plausible" (5). Sentences were phrased in the following manner: "This is not an [x], it is a [y]."
Coachella 2015 Twitter sentiment 3,847 3,847 Sentiment Analysis Text K choose 1 Selection, Join A sentiment analysis job about the lineup of Coachella 2015.
Apple Computers Twitter sentiment 3,969 3,969 Sentiment Analysis Text K choose 1 Selection, Join Contributors were given a tweet and asked whether the user was positive, negative, or neutral about Apple. (They were also allowed to mark "the tweet is not about the company Apple, Inc.)
How beautiful is this image? (Part 1: People) 3,500 3,500 Image classification Image Rating(They were given a five-point scale, from "unacceptable" (blurry, red-eyed images) to "exceptional" (hi-res, professional-quality portraiture).) Selection, Categorize, Join Here, contributors were asked to rate image quality (as opposed to how pretty the people in the images actually are). They were given a five-point scale, from "unacceptable" (blurry, red-eyed images) to "exceptional" (hi-res, professional-quality portraiture) and ranked a series of images based on that criteria.
How beautiful is this image? (Part 2: Buildings and Architecture) 3,500 3,500 Image classification Image Rating( They were given a five-point scale, from "unacceptable" (out-of-focus cityscapes) to "exceptional" (hi-res photos that might appear in a city guide book) .) Selection, Categorize, Join Here, contributors were asked to rate image quality (as opposed to how gorgeous the buildings in the images actually are). They were given a five-point scale, from "unacceptable" (out-of-focus cityscapes) to "exceptional" (hi-res photos that might appear in a city guide book) and ranked a series of images based on that criteria.
How beautiful is this image? (Part 3: Animals) 3,500 3,500 Image classification Image Rating( They were given a five-point scale, from "unacceptable" (blurry photos of pets) to "exceptional" (hi-res photos that might appear in text books or magazines).) Selection, Categorize, Join Here, contributors were asked to rate image quality (as opposed to how adorable the animals in the images actually are). They were given a five-point scale, from "unacceptable" (blurry photos of pets) to "exceptional" (hi-res photos that might appear in text books or magazines) and ranked a series of images based on that criteria.
Language: Certainty of Events 13,386 13,386 Sentiment Analysis Text K choose 1 Selection, Join A linguistical data set concerning the certainly an author has about a certain word. For example, in the following sentence: "The dog ran out the door," if the word "ran" was asked about, the certainty that the event did or will happen would be high.
New England Patriots Deflategate sentiment 11,814 11,814 Sentiment Analysis Text K choose 1 Selection, Join Before the 2015 Super Bowl, there was a great deal of chatter around deflated footballs and whether the Patriots cheated. This data set looks at Twitter sentiment on important days during the scandal to gauge public sentiment about the whole ordeal. 
Sports Illustrated covers 32,000 32,000 Image Retrieval Image K choose 1 Selection, Collection, Join A data set listing the sports that have been on the cover of Sports Illustrated since 1955. 
The data behind data scientists 974 974 Fact Retrieval Text Survey Collection A look into what skills data scientists need and what programs they use. A part of our 2015 data scientist report which you can download. 
Judge the relatedness of familiar words and made-up ones 300 300 Relevance Finding Text Rating(1-5, from completely unrelated to very strongly related, respectively.) Selection, Categorize, Join Contributors were given a nonce word and a real word, for example, "leebaf" and "iguana." They were given a sentence with the nonce word in it and asked to note how related the nonce word and real word were. Here's a sample question: "Large numbers of leebaf skins are exported to Latin America to be made into handbags, shoes and watch straps."  Contributors then ranked the relation of "leebaf" to "iguana" on a scale of 1-5, from completely unrelated to very strongly related, respectively.
2015 New Year's resolutions 5,011 5,011 Sentiment Analysis Text K choose 1 Selection, Join A Twitter sentiment analysis of users' 2015 New Year's resolutions. Contains demographic and geographical data of users and resolution categorizations.
Smart phone app functionality 1,898 1,898 Fact Retrieval Text K choose 1 Collection Contributors read an app description, then selected the app's functionality from a pre-chosen list. Functionalities ranged from SMS to flashlight to weather to whether or not they used a phone's contacts. Contributors were allowed to select as many functionalities as applied for each app.
Naturalness of computer generated images 600 600 Image classification Image K choose 1 Selection, Categorize, Join Contributors viewed two rather bizarre looking images and were asked which was more "natural." Images were all computer generated faces of people in various states of oddness.
National Park locations 323 323 Fact Retrieval Text Label Collection A large data set containing the official URLs of United States national and state parks.
Colors in 9 Languages 4,000 4,000 Fact Retrieval Text K choose 1 Collection Dataset of 4000 crowd-named colors in 9 languages.  Includes the RGB color, the native language color, and the translated color.
Judge emotions about nuclear energy from Twitter 190 190 Sentiment Analysis Text 5 choose 1 Selection, Join This dataset is a collection of tweets related to nuclear energy along with the crowd's evaluation of the tweet's sentiment. The possible sentiment categories are: "Positive", "Negative", "Neutral / author is just sharing information", "Tweet NOT related to nuclear energy", and "I can't tell". We also provide an estimation of the crowds' confidence that each category is correct which can be used to identify tweets whose sentiment may be unclear.
Decide whether two English sentences are related 555 555 Relevance Finding Text Rating (5 choices) Selection, Categorize, Join This dataset is a collection of English sentence pairs. The crowd was asked about the truth value of the second sentence if the first sentence were true and to what extent the sentences are related on a scale of 1 to 5. The variance of this score over the crowd's judgments is included as well.
Similarity judgement of word combinations 6,274 6,274 Relevance Finding Text 7 choose 1 Selection, Categorize, Join Contributors were asked to evaluate how similar are two sets of words on a seven point scale with 1 being "completely different" and 7 being "exactly the same".
Sentiment Analysis - Global Warming/Climate Change 6,090 6,090 Sentiment Analysis Text 3 choose 1 Selection, Join Contributors evaluated tweets for belief in the existence of global warming or climate change. The possible answers were "Yes" if the tweet suggests global warming is occuring, "No" if the tweet suggests global warming is not occuring, and "I can't tell" if the wtweet is ambiguous or unrelated to global warming. We also provide a confidence score for the classification of each tweet.
Judge Emotion About Brands & Products 9,093 9,093 Sentiment Analysis Text 3 choose 1 Selection, Join Contributors evaluated tweets about multiple brands and products. The crowd was asked if the tweet expressed positive, negative or no emotion towards a brand and/or product. If some emotion was expressed they were also asked to say which brand or product was the target of that emotion.
Claritin Twitter 4,900 4,900 Relevance Finding Text K choose 1 Selection, Categorize This dataset has all  tweets that mention Claritin for October, 2012.  The tweets are tagged with sentiment, the author's gender, and whether or not they mention any of the top 10 adverse events reported to the FDA.