Algorithms and the news agenda

Well worth a read on the Feguson riots, and how different social media sites (notably Twitter vs Facebook) served up news about them:

“Now, we expect documentation, live-feeds, streaming video, real time Tweets… [Ferguson] unfolded in real time on my social media feed which was pretty soon taken over by the topic…

And then I switched to non net-neutral Internet to see what was up. I mostly have a similar a composition of friends on Facebook as I do on Twitter.

Nada, zip, nada.

This morning, though, my Facebook feed is also very heavily dominated by discussion of Ferguson. Many of those posts seem to have been written last night, but I didn’t see them then. Overnight, “edgerank” –or whatever Facebook’s filtering algorithm is called now?—?seems to have bubbled them up, probably as people engaged them more.

But I wonder: what if Ferguson had started to bubble, but there was no Twitter to catch on nationally? Would it ever make it through the algorithmic filtering on Facebook? Maybe, but with no transparency to the decisions, I cannot be sure.

Would Ferguson be buried in algorithmic censorship?

Would we even have a chance to see her?

This isn’t about Facebook per se—maybe it will do a good job, maybe not—but the fact that algorithmic filtering, as a layer, controls what you see on the Internet. Net neutrality (or lack thereof) will be yet another layer determining this. This will come on top of existing inequalities in attention, coverage and control.”

It’s a continual worry – how to ensure we see what’s important? Though, of course, the concept is nothing new – the algorithm is just an editor or an editorial policy in a different form. It’s something I’ve written about before when it relates to the EU, focusing on a BBC editorial policy that fails to cover EU affairs in mainstream news most of the time, and then serves up extremes.

This kind of human editorial determination of the appropriate news agenda based on perceived audience interests is arguably no massive degree different from a Facebook algorithm determining what is important based on how it interprets user interests. If anything, there’s a strong argument to be made that Facebook knows its audience better than any editor on any publication or TV show ever, due to the sheer quantities of data it possesses on its userbase.

But then what of *importance* – who determines this? Who overrides the algorithmic or standard editorial policy assumption? Is there a chance that an important story will get buried because a bit of code doesn’t see it as significant? Yes. But the same is true of any number of important news stories that human editors don’t pick up on, or choose to bury on page 23 because they don’t think their readers will be that interested.

As so often, the web may be a bit different, but there’s nothing that *new* here.

Understanding how people interact with your content: the code

Upworthy have released the code they use to track user engagement, with a nice bit of methodology explaining what they’re tracking and why they care:

“In the age of ever-present social media, our collective attentions have never been spread thinner. According to Facebook, each user has the potential to be served 1,500 stories in their newsfeed each time. For a heavy user, that number could be as much as 15,000. In this climate, how do you get people to pay attention? And, more importantly, how do you know they’re actually engaged?

“Clicks and pageviews, long the industry standards, are drastically ill-equipped for the job. Even the share isn’t a surefire measure that the user has spent any time engaging with the content itself. It’s in everyone’s interest — from publishers to readers to advertisers — to move to a metric that more fully measures attention spent consuming content. In other words, the best way to answer the question is to measure what happens between the click and the share. Enter Attention Minutes.”

Let the social media backlash backlash begin!

Some great charts here – the overhype of social has been increasingly grating in recent years, a repeat of the 2003-5 excitement over blogging as the future of everything journalism, or the great SEO craze of approximately the same period, where the right combination of metadata and keywords were seen as some kind of magic bullet that could take any site to the top of the first page of Google.

Thankfully, everyone’s woken up to the limitations of both blogging and SEO. We’re now hopefully now coming to the same realisation with social, with more and more myths about clicks, engagement, sales, and all sorts being shattered left, right and centre.

But what we really need is the backlash to the backlash will hopefully follow soon after. Because although neither blogging nor SEO were quite the massive game changers they were made out to be, both have had (and continue to have) a huge positive impact on both the online world and the media as a whole. We simply now have a better understanding of their limitations as well as their strengths – which puts us in a much stronger position. Add the same rational approach to social (an approach that anyone with a Twitter addiction as bad as mine could have told you about years ago), and we should end up stronger yet again.

The power of Google

Another Google algo update and, as ever, original, interesting, useful content is key to SEO success.

The hit eBay’s taken is interesting, though… An 80% drop in Google traffic coukd be a business-killer for anyone less big. And their content surely *is* original and relevant, what with the products changing all the time?

Possibly another impact from the authorship/Google+ changes the Google guys have introduced? After all, eBay product page writers are hardly likely to be verified Google+ authors. Is this why eBay are starting to invest in creating narrative content around their auctions?

Update: See also the ever-excellent Matthew Ingram on this, who points out the extremely worrying hit the long-running and much-loved Metafilter has taken:

“Reliant on Google not only for the bulk of its traffic but also the bulk of its advertising revenue, Metafilter has had to lay off almost half of its staff.”

The lesson?

Google can kill a site on a whim, and even the experts can’t tell us how or why, because Google’s algorithms are even more secret than the Colonel’s delicious blend of herbs and spices. Any site dependent on search for the bulk of its traffic is playing a very, very dangerous game.

Update 2: More detail on the Metafilter revenue/traffic decline, complete with stats.

The related power of Facebook to stifle updates from sources it has deemed to be suspect for whatever reason simply – and even the New York Times’ recently-leaked innovation report’s charts In the decline of its homepage – makes an obvious cliché all the more true even it comes to Web traffic: don’t put all your eggs in one basket.

If more than 25% of your traffic/revenue comes from one source, you’re in danger. More than 50%, you have a potential death sentence. All it takes is one thing to change, and you’re screwed.