wisdom of the crowd
play

Wisdom of the Crowd you rely heavily on? (e.g., imdb ratings, CS - PowerPoint PPT Presentation

Reply in Zoom chat: Which peer signals do Wisdom of the Crowd you rely heavily on? (e.g., imdb ratings, CS 278 | Stanford University | Michael Bernstein Carta, online product reviews) Where we are, and where were going Week 1-2: Basic


  1. Reply in Zoom chat: Which peer signals do Wisdom of the Crowd you rely heavily on? (e.g., imdb ratings, CS 278 | Stanford University | Michael Bernstein Carta, online product reviews)

  2. Where we are, and where we’re going Week 1-2: Basic ingredients — motivation, norms, and strategies for managing growth Week 3: Groups — strong/weak ties, collaboration Week 4: Massive collaborations The Wisdom of the Crowd Crowdsourcing and Peer Production 2

  3. http://hci.st/wise Grab your phone, fill it out!

  4. How much do you weigh? My cerebral cortex is insufficiently developed for language 4

  5. Whoa, the mean guess is within 1% of the true value 5

  6. Innovation competitions in industry Innovation competitions for science 6

  7. Prediction markets AI data annotation at scale 7

  8. Today What is the wisdom of the crowd? What is crowdsourcing? Why do they work? When do they work? 8

  9. Wisdom of the crowd

  10. Crowds are surprisingly accurate at estimation tasks Who will win the election? How many jelly beans are in the jar? What will the weather be? Is this website a scam? Individually, we all have errors and biases. However, in aggregate, we exhibit surprising amounts of collective intelligence. 10

  11. “Guess the number of minutes it takes to fly from Stanford, CA to Seattle, WA.” 70 90 110 130 150 170 190 If our errors are distributed at random around the true value, we can recover it by asking enough people and aggregating. 11

  12. What problems can be solved this way? Jeff Howe [2009] theorized that that it required: Diversity of opinion Decentralization Aggregation function So — any question that has a binary (yes/no), categorical (e.g., win/ lose/tie), or interval (e.g., score spread on a football game) outcome 12

  13. What problems cannot be solved this way? Flip the bits! People all think the same thing People can communicate No way to combine the opinions For example, writing a short story (is much harder!) 13

  14. General algorithm 1. Ask a large number of people to answer the question Answers must be independent of each other — no talking! People must have a reasonable level of expertise regarding the phenomenon in question. 2. Average their responses 14

  15. Why does this work? [Simoiu et al. 2020] Independent guesses minimize the effects of social influence Showing consensus cues such as the most popular guess lowers accuracy If initial guesses are inaccurate and public, then the crowd never recovers Crowds are more consistent guessers then experts In an experiment, crowds are only at the 67th percentile on average per question…but at the 90th percentile averaged across questions! Think of this as the Tortoise and the Hare, except the Tortoise (crowd) is even faster — at the 67th percentile instead of the worst percentile 15

  16. Mechanism: ask many independent contributors to take a whack at the problem, and reward the top contributor 16

  17. Mechanism: ask paid data Mechanism: use a market to annotators to label the same image aggregate opinions and look for agreement in labels (much more on the implications of paid crowd work in the Future of Work lecture)

  18. Let’s check our http://hci.st/wise results

  19. Aggregation approaches

  20. Early crowdsourcing [Grier 2007] Two distributed workers work independently, and a third verifier adjudicates their responses 1760 British Nautical Almanac Nevil Maskelyne 20

  21. Work distributed via mail Work distributed via mail 21

  22. Charles Babbage = Two people doing the same task in the same way will make the same errors. 22

  23. I did it in 1906. And I have cool sideburns. You reinvented the same idea, but it was stickier this time because statistics had matured. Unfortunately, you also held some pretty problematic opinions about eugenics. 23

  24. Mathematical Tables Project WPA project, begun 1938 Calculated tables of mathematical functions Employed 450 human computers The origin of the term computer 24

  25. 20th Century Fox

  26. Enter computer science Computation allows us to execute these kinds of goals at even larger scale and with even more complexity. We can design systems that gather evidence, combine estimates, and guide behavior. 26

  27. Forms of crowdsourcing

  28. Definition Crowdsourcing term coined by Jeff Howe [2006] in Wired “Taking [...] a function once performed by employees and outsourcing it to an undefined (and generally large) network of people in the form of an open call.” 28

  29. Volunteer crowdsourcing Tap into intrinsic motivation to recruit volunteers Collaborative math proofs Kasparov vs. the world NASA Clickworkers Search for a missing person Wikipedia Ushahidi crisis mapping 29

  30. Automated sharing Opt in to sharing and aggregation Purple Air air quality sensors Waze traffic sharing (also includes manual) 30

  31. What if the task were embedded in another goal? Just like I get exercise on my commute to Stanford When I could still commute to Stanford *quiet sob* 31

  32. Games with a purpose [von Ahn and Dabbish ’08] Make the data labeling goal enjoyable. You are paired up with another person on the internet, but can’t talk to them. You see the same image. Try to guess the same word to describe it. 32

  33. Games with a purpose [von Ahn and Dabbish ’08] Let’s try it. Volunteers? Taboo words: Burger Food Fries 33

  34. Games with a purpose [von Ahn and Dabbish ’08] Let’s try it. Volunteers? Taboo words: Stanford Graduation Wacky walk Appendix 34

  35. reCAPTCHA “Oh, I see you’d like to make an account here. Sure would be a shame if you couldn’t get into my website. Maybe you should help me train my AI system and I’ll see if I can do something about letting you in.” 35

  36. Handling collusion and manipulation

  37. Not the name that the British were Stephen Colbert fans raid NASA’s expecting to see vote to name the new ISS wing 37

  38. A small number of malicious individuals can tear apart a collective effort. 38

  39. [Example via Mako Hill] 39

  40. [Example via Mako Hill] 40

  41. [Example via Mako Hill] 41

  42. [Example via Mako Hill] 42

  43. [Example via Mako Hill] 43

  44. Can we survive vandalism? Michael’s take: it’s a calculation of the cost of vandalism vs. the cost of cleaning it up. How much effort does it take to vandalize Wikipedia? How much effort does it take an admin to revert it? If effort to vandalize >>> effort to revert, then the system can survive. How do you design your crowdsourcing system to create this balance? 44

  45. Who do we trust? [Sheng, Provost, Ipeirotis, ’08] We need to answer two questions simultaneously: (1) What is the correct answer to each question? and (2) Which participants’ answers are most likely to be correct? Think of it another way: if people are disagreeing, is there someone who is generally right? An algorithm called Get Another Label solves this problem by answering the two questions simultaneously 45

  46. Get Another Label [Sheng, Provost, Ipeirotis, ’08] Inspired by Expectation Maximization Given current contributor (EM) algorithm from AI. estimates, estimate the probability of each answer Use the workers’ guesses to estimate the most likely answer for each question. Loop until Use those answers to estimate worker convergence quality. Use those estimates of quality to re-weight the guesses and re-compute Given current answer probabilities, estimate answers. Loop. contributor accuracy 46

  47. Bayesian Truth Serum [Prelec, Seung, and McCoy ’04] Inspiration: people with accurate meta-knowledge (knowledge of how much other people know) are often more accurate So, when asking for the estimate, also ask for each person’s predicted empirical distribution of answers Then, pick the answer that is more popular than people predict 47

  48. Bayesian Truth Serum [Prelec, Seung, and McCoy ’04] “When will HBO have its next hit show?” 1 year / 5 years / 10 years “What percentage of people do you think will answer each option?” 1 year / 5 years / 10 years An answer that 10% of people give but is predicted to be only 5% receives a high score 48

  49. Bayesian Truth Serum [Prelec, Seung, and McCoy Nature ’04] Calculate the population endorsement frequencies ¯ x k for each option k and the geometric average of the predicted frequencies ¯ y k Evaluate each answer according to its information score: log ¯ x k y k ¯ And reward people with accurate prediction frequency reports

  50. Judging quality explicitly Gold standard judgments [Le et al. ’10] Include questions with known answers Performance on these “gold standard” questions is used to filter submissions Gated instruction [Liu et al. 2016] Create a training phase where you know all the answers already, and give feedback on every right or wrong answer during training At the end of training, only let people go on if they have a high enough accuracy 50

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend