Uncategorized


Everyone has their own stories, no? The time that Wikipedia made you understand something a textbook couldn’t,

Stack Overflow is a programming help site. The site is heavily controlled by its community; as users gain points and achievements for contributing to the site, they gain more administrative power.

I’ll admit, I went through a period where I watched the questions page daily, looking for something I could answer. However, I found  that every single time, somebody gave a better answer much quicker. I’m sure it is eventually discouraging, but in my experience it made me competitive. If you don’t get the first answer, at least get a higher rated response. Programming is an endless process of puzzles and problem-solving, so it’s good to know that there’s a place where one in need can get help within minutes. As Jeff Atwood notes in the tweet above, it’s quality help too.

Clay Shirky has a wonderful talk about this phenomenon, wherein he argues that passion (or ‘love’) is a resource that is unfairly underestimated. He uses the example of support for the open-source programming language perl: in an absence of commercial support, there are more than enough passionate perl users online to fill the need. Watch it below.

In recent months, there have been two large-scale crowdsourcing project launched with the involvement of well-known internet personalities. I’m referring to Hunch—heading by Caterina Fake of Flickr fame— and Kickstarter—advised by Andy Baio of Upcoming.org and Waxy.org fame. While these projects are very different, it seems only appropriate for them to share a post.

Kickstarter is a site that lets projects collect funding pledges from users. One can add a project with a funding goal and set tiered rewards for patrons. You’re not expected to pay your pledge unless the project goal is met. Hunch is a crowdsourced decision tree site. Users creating content add decision-making questions to a decision, ones that affect the outcome of the final recommendation.

Had either of these been released elsewhere, the project would not provide any particularly extraordinary impact. Indeed, the paradox of crowdsourcing is that the idea alone will not sell the project, because the difficult barrier of critical mass. Critical mass, however, is itself dependent on having sold the idea already. In other words, new crowdsourcing project can have a great idea in principle, but potential users are aware that it will be difficult to achieve in practice. Such doubt is self-propogating, because knowing that you have doubts means admitting that others may too, further removing confidence in the objective.

A site like Fake’s former creation Flickr was able to grow a community on the back of a valuable individual experience. The community was not vital to the experience of Flickr. Kickstarter and Hunch, however, have no individual experience. If there was only one user, that user would have little to do (admitting that Hunch’s employee-created content can only go so for).

With strong talent behind both of the sites, there is no shortage of cleverness in their mechanics. However, the key to their first few months lies in the trusted celebrity behind them. Like doubt, confidence is be self-propagated. Baio and Fake have an audience already in place, and seeing that audience can swing cynics from “great idea but wouldn’t work” to thinking “just maybe.”

“Scepticism about Wikipedia’s basic viability made some sense back in 2001; there was no way to predict, even with the first rush of articles, that the rate of creation and the average quality would both remain high, but today those objections have taken on the flavor of the apocryphal farmer beholding his first giraffe and exclaiming, ‘Ain’t no such animal!’ Wikipedia’s utility for millions of users has been settled; the interesting questions are elsewhere.”
– Clay Shirky, Here Comes Everybody, p.117

In my work on crowdsourcing, my advisors warn me to be careful of how I speak about Wikipedia around academics, because scholars are still divided on it. Clay Shirky’s quote perfectly encapsulates the situation: if it is clear that it works and that it works well, the question shouldn’t be “does it work?” Rather, we should be asking why it works. Kevin Kelly suggests that Wikipedia is “impossible in theory, but possible in practice“: shouldn’t we be tweaking our theories then? Perhaps then, the issue is that if an expert were to praise Wikipedia as reliable, they undermine society’s need for experts. Larry Sanger, creator/co-founder or Wikipedia, says no, but it’s certainly food for thought.

I’ve recently been mulling over the question, “Why does Genius suck so much?” and the implications that it has.

Genius is the playlist generation tool in Apple’s iTunes music software. You choose a song that you’re in the mood for, and it creates an entire playlist of similar songs. Essentially, its a recommender system; if you like x you’ll like y. The problem is that you get a very narrow point of view, with very little genre skipping. and no pleasantly clever surprises.

What sets Genius apart from other song recommender systems is that its essentially powered by the crowds. Apple has the luxury of a rich data set of habits and rating, and it appears to factor heavily into the recommendations. Indeed, algorithmic playlist generators were creating better results years before Genius came on the scene. So, what does this mean for the crowd?

The fact that computers can be better than humans in understanding art is off-putting. I’m still working through this problem, but here are some thoughts toward untangling it.

Ratings data is emotionless. When you rate a song 1 or 5, you’re giving it a universal ‘like’/’dislike’. This data doesn’t factor the mood of the song or the emotion of the listener. This is all very removed for circumstance. As I suggested to Bill Turkel, perhaps such simple crowd-based recommendations are better for high-level suggestions, like artists you may like, but useless at the micro-level (unless that data crowds are contributing is more specific to the topic of recommendations). In contrast, technology can quite effective interpreting the types and patterns of sound which represent an emotion. Certainly it can’t easily understand whether a song is good, but if you want a slow, jazzy rock song, that’s fairly achievable. This is something in which music recommendation is fairly unique, as it is easy to interpret than it would be to interpret thousands of movie plots or millions of book themes.

Despite this, perhaps the most-cited example of a good music recommender is Pandora, which is an internet radio based on the Music Genome Project (MGP). The MGP does use humans to categorize songs, having professionals tag each song with over 400 tags and using an algorithm to weigh the values. Pandora’s success shows that humans are indeed effective at understanding music, given that they’re looking at it in the right way.

There’s also the effect of popular media that makes human-based recommendations unbalanced. If a lot of people like Coldplay, the range of music that it will be recommended for will be broad. This additionally creates an echo loop where popular music simply grows in popularity. Inversely, it is very difficult for new music to enter the loop. If everybody that likes The Strokes like Yeah Yeah Yeahs, the recommender will reinforce this, brushing aside any similar new bands.

However, such problems are limited to the balance of the algorithm. Last.fm, which tracks all of its users’ listened music, is fairly effective in recommending similar music. Also, because of their detailed information on what a user has listened to, they can suggest less listened to songs. Though they don’t offer playlist generation, I wouldn’t put this beyond their abilities.

So where do crowds factor in here? If anything, Pandora suggests that this is best left to professionals. Certainly, you can’t get that sort of exhaustivity with crowds. The answer may lie in reliability. Large groups would be able to make much simpler connections, but on a larger and more verified scale. When I make a playlist with Lou Reed’s Take a Walk on the Wild Side, I always follow it with Urge Overkill’s Girl, You’ll Be A Women Soon.  The songs are linked very little, but there’s something in me that recognizes the similarly cool feeling that I feel. If you could somehow capture millions of these sorts of links, that could lead somewhere.