Skip navigation

Tag Archives: microblogging


While there has been much breathless speculation about the threat that Twitter poses to Google’s iron grip on the search market, I haven’t seen many specific ideas about how that would work. By now we all know that searching real-time content is an increasingly valuable resource for information, but its utility seems to be limited to certain kinds of information like breaking news events and consumer product reviews.  It certainly doesn’t appear to be as flexible a mechanism for locating the wide range of content that traditional search has been since the inception of the world wide web.

The most intriguing speculation that I’ve read about the future of search has been this post about the notion of ‘PageRank for People’. In essence, the idea states that the current algorithms that govern the ranking of search results are inadequate because they rely too heavily on the location of the content being listed. Since Google relies on the volume of inbound links to judge the value of content, it favors content posted in popular locations. The thing is, a piece of good content is just as valuable if it is posted to ‘Bob’s Blog’ or the ‘New York Times’. The solution that was proposed, was a system that factors the ‘reputation’ and ‘authority’ of the content’s author when ranking search results. Just how to calculate these numbers, though, is the tricky part. One answer might be Twitter.

There are two aspects of Twitter that need to be changed. Both the ‘Retweeting’ and ‘Hashtag’ behaviors need to be provided as official features of the service. That means the mechanism for these actions needs to be separated from the 140 character text string of each tweet. That is to say, we should lose the ‘RT’ syntax and make the identity of the original poster some form of metadata that exists outside the post itself. Likewise, tags like those currently labeled with hash-signs ‘#’ should be saved as metadata separate from the actual tweet. With those two changes, searching the web could become a lot more useful. Here’s how it would work:

When I publish a post to Twitter, I should have the ability to tag that post with semantically accessible identifying labels. Each post associated with a given subject (as described by the label) potentially contributes to my ‘authority’ about that subject. Now, when someone retweets my post, they are in essence endorsing what I have said and contributing to their own ‘authority’ about that subject. If there was a scoring mechanism that assigned, say, one point for each endorsement, then you’ve created a system that establishes quantitative values for ‘authority’. Imagine it this way:


For each retweet, additional points are cascaded up the tree. That way the original poster is always given the most credit for contributing the idea, but those that help propagate it are given credit as well.  Authority is defined by the community, not the individual. It should be pointed out that for each subsequent retweet, the poster will have the opportunity to revise the tagged metadata, either adding more detail, or removing labels they believe are not appropriate. That way, the system guards against abuse by so-called trend-squatting.

Now, once we start getting values assigned for Twitter users’ authority on specific topics, search engines can start factoring this in to their rankings. So, content authored by an individual with higher authority for the subject of that content are favored over others. Content authored by organizations might be scored, in part, by the collective authority of that organization’s members. It would create a tremendous upward pressure to contribute value to the community. For some industries, I imagine that one’s scores in this respect would become a factor in employment decisions or compensation levels.

Now,  what I’ve described would need to be only one part of the search ranking algorithm. As described by the original post, there are many other factors that should be considered. Additionally, the scoring mechanism described above is probably far too simplistic and vulnerable to abuse. For example, one complication that could make the system more reliable would be to consider the reputation of the endorsing party when assigning a value to the score that their retweet provides the original poster. That way, it is more valuable to get retweeted by individuals with more authority. Note that ‘authority’ as I’ve described it is an entirely separate metric from ‘popularity’, which is defined by the number of followers that a user has.

There would be many additional side-benefits of such a system. For example, much of what is posted on Twitter are links to content elsewhere on the web. A robust labeling function would turn Twitter into a tagging system for the entire semantic web.

Here is a related discussion.



One of the lines I like to use when discussing how media is changing these days is, “The internet is punishing inefficiency.” I’ve forgotten who I lifted that line from, but I now add the following chaser: “The economic downturn is accelerating that process.” There is no doubt that the ways in which people are consuming information are changing rapidly. Newspapers are reeling, the recording industry has succumbed to the inevitable, and TV providers are getting nervous. Likewise, the industries that rely on gaming consumer attention, namely Advertising and PR, are scrambling to understand what their jobs are anymore. Part of the problem they face is that they are chasing a moving target, and no one is sure where it is headed. 

I tweeted a few weeks ago, that I thought the challenge of deriving value form Twitter was largely a problem of managing the ‘Signal to Noise Ratio’. There’s a lot of useful content out there, but if it isn’t parsed from the firehose for you, it’s of little value. This of course is why we have a media in the first place. The role of writers and broadcasters has always been to parse the wide world of ideas and bring us the nuggets that we value for their news or entertainment value.

In the beginning, ideas were spread sideways. That is to say we communicated almost entirely by word-of-mouth, and memes reproduced via a cascade of parallel paths. The trouble with this method was that it was slow and prone to errror.

Next, technology made a top down model practical. As it became economically feasible for small groups to communicate with large groups using things like print and broadcasts. This was the birth of ‘media’ and it dealt with the speed and error problems and was tremendously useful. The problem with this model, however was that the masses were all largely subjected to the same set of information, regardless of individual need. This system also made public attention vulnerable to manipulation, a loophole that the marketing industries exploited for centuries. 

Then another technological advancement made it practical for anyone to communicate with everyone. This made it much easier customize sets of data and package them for small audiences. The problem became that firehose I mentioned a bit ago. Suddenly we were back to the beginning of this cycle again. At first there was a sideways model online whereby users relied on tools like email to move the stuff around. Then a top down model arose as publishers and broadcasters attempted to apply their skillset to the web.

Then, for the first time, the web started to try something new. Taking advantage of the technology, models for parsing information were attempted that allowed the masses to self-regulate the firehose. For the first time we tried allowing our peers to do the parsing, leaning on our collective wisdom, services like Digg and Reddit emerged to help tailor content to individual need. The shortcomings of these so-called ‘democratized content forums’ was that they created a disadvantage for individual’s whose content needs stray far from the average. 

About two years ago I speculated about what the next model would be. At the time I became enamored with the idea that so-called ‘prediction markets‘ might just be that next model (see also ‘Infotopia‘ for a good discussion of this). It turns out (current economic crises notwithstanding) that stock market mechanisms are extremely efficient tools for extracting wisdom from groups. This fascination of mine resulted in an experiment called which was my attempt to apply the model to news and information. I now beleive I miss-read the trends. The actual next model is that which is being popularized by Twitter.

The innovation that this micro-blogging model provides is that content is parsed by the collective wisdom of the group, much like the voting forums, but it is a group of my choosing. As a Twitter user, I manage my own mini-mob, and that mini-mob manages my content for me. 

Now, managing the mob is not an easy thing to do. At the moment the tools available to accomplish this on Twitter are crude and cumbersome. But, these will improve as Twitter improves itself, or someone comes along to build a better micro-blogging platorm. Much of the attention of Twitter appliction developers so far has been focussed on creating new tools to parse out content directly from Twitter. There are an ever-growing library of live-streaming, filtering, and searching interfaces for Twitter. These are welcome and valuable additions to the service, but I think they miss the mark a bit in terms of taking full advantage of this new form of media. By foccussing on the content of the tweets themselves, most do not directly factor in the parsing ability of individual people. I’m waiting for more sophisticated tools that will allow me to find new tweeple, based on expertise, interests and quality. I want these tools to also help me prune the herd of tweeps that aren’t holding their own. In essence, I want tools that will help me make my mob better than your mob. 

There are some tools like this now. Mr Tweet comes to mind, as does Kevin Rose’s new pet WeFollow. These are good starts but leave a lot to be desired. For example, WeFollow is very comprehensive, but relies solely on the number of followers a Twitter user has to rank the list in each category. I think that this metric is a terrible measure of the quality of users for two reasons. Firstly, it’s too easy to game the system and run up that number without regard for merit. Secondly, it confuses quality with popularity skewing attention toward the center. That’s the same problem that Digg has, as I’ve already described. What is needed, is a more useful metric to measure, compare and match users up with each other. What does that look like? Well, I have at least one idea and will be posting much more about that in the future. 

So where does this all leave marketers? In many respects they are having the rug pulled out from beneath their feet. The top-down loopholes are becoming less and less effective. In many ways, they are being treated like the uninvited intruder in the new media party, trying to butt into the conversation and unable to do anything but annoy the consumer partygoers. Is all hope lost? Hardly. I think that with creative thinking, a firm understanding about what is actually happening, and a healthy respect for what consumers actually want, that industry will improve right along with the media. Speaking as a consumer, I don’t mind if marketers participate in the conversation I’m crafting for myself online as long as their actions provide value to me. If they don’t my mob will take care of it.