Don’t shoot the Blackberry Messenger

BBM does things no web social network can do… it mirrors the flexibility of real life RIM’s fortunes have taken a catastrophic, Nokia-style nosedive in the past year – but it has a chance of pulling up. Admittedly, the odds are long, but this week the Canadian company began its fightback. It’s certainly right up … Read more

The BBC struggles with the concept of ‘tech bubble’

The BBC has a real problem with social media. It’s delighted when something new appears. It slips into the patrician role that comes naturally to broadcasters – and especially the BBC. It can express childlike wonderment – Wow! – at something new and amazing. Getting beyond that though, is where the trouble starts.

Perhaps the BBC is haunted by the idea that people simply get on and use new communication tools without “Auntie’s” assistance. The viewers typically also have much more realistic expectations of the technology than, say, pundits. So we keep hearing wonderment, and advice on how get online, a bit like a slightly mad primary school teacher.

The gears really grind when something more critical is required. This week the corporation’s news flagship Newsnight – one of the last remaining TV programmes for grown-ups – asked if there was a “tech bubble”. Investment is pouring into social media startups. Would it all end in tears?

Yet having the posed the question, the report and discussion that followed were designed to dispel understanding and analysis. Before long it had turned into a gathering of the Unicorn Preservation Society. We were even told that only people who might want to describe the web investments a “bubble” were self-serving opportunists.

Bad people, in other words, thinking bad thoughts.

Read more

The Autistic Network

Andrew’s Mailbag Isn’t it time for a War on Stupid Generalisations? These usually need a Czar, and I’d gladly volunteer.

Novelist Zadie Smith has written about the movie The Social Network and wonders if Mark Zuckerberg’s apparent extreme autism doesn’t manifest itself in both the reductive view of humanity that Facebook (and Web 2.0) software demands. That’s fine. It’s when she steers into observations about autism – largely borrowed, because she’s not an original thinker – from Jaron Lanier’s You Are Not A Gadget – that the trouble starts.

Read more

Nathan Barleys to fill Olympic chasm – PM

Prime Minister David Cameron has cast his gaze east across to Essex – and dreams of a landscape filled with social media marketing consultants and SEO boutiques as far as the eye can see.

In the aftermath of the Olympics, Cameron wants to put the land and property on the Lea Valley to private sector use, and his Big Idea is to “nudge” the Shoreditch and Hoxton crowd eastward.

“Our ambition is to bring together the creativity and energy of Shoreditch and the incredible possibilities of the Olympic park to help make east London one of the world’s great technology centres,” said Cameron today.

That would be a sight: a mass migration of tiny designer tricycles as the Nathan Barleys pedal across the Hackney Marshes to Essex.

Read more

Greatest Living Briton gets £30m for ‘web science’

As an alliance of the desperate, this one takes some beating. The Greatest Living Briton (Sir Timothy Berners Lee) has been thrown £30m of taxpayers’ money for a new institute to research “web science”.

Meanwhile the Prime Minister waxed lyrical today about the semantic web – how “data” would replace files, with machine speaking unto machine in a cybernetic paradise.

It’s really a confluence of two groups of people with a shared interest in bureaucracy.

Computer Science is no longer about creating graduates who can solve engineering challenges, but about generating work for the academics themselves. The core expertise of a CompSci department today is writing funding applications. And the Holy Grail for these paper chasers is a blank cheque for work which can be conducted without scrutiny for years to come. With its endless committees defining standards (eg, “ontologies”, “folksonomies”) that no one will ever use, the “Semantic Web” fits the bill perfectly.

Of course, most web data is personal communication that happens to have been recorded. Most of the rest is spam, generated by robots, or cut-and-paste material ‘curated’ by the unemployed or poor graduates – another form of spam, really. The enterprise is doomed. But nobody’s told the political class.

Read more

Panorama on the Digital Economy Bill

BBC1’s flagship current affairs program was devoted to file sharing last night, and contained something to piss off a range of lobbyists.

Usually when this happens, BBC producers often conclude “they’re doing something right”, and pour themselves a large, congratulatory drink. They shouldn’t, because while the program succeeded in trying to be “fair”, it failed in its larger mission to present the issue properly – something we already understand.

Read more

Web 2.0 and feedback loops: a conversation with James Harkin

Weiner

Don’t judge a book by the title. Especially if the title is something like Cyburbia. James Harkin, who worked with Adam Curtis on The Trap, has produced the first proper full-length critique of Web 2.0 – tracing the daftness back to the cybernetics pioneers of the 1940s.

It’s odd that something with so much hype as Web 2.0 has received so little intelligent criticism. Half of Nick Carr’s The Big Switch, looked at the social and psychological implications, and he’s following up at length in The Shallows.

But Cyburbia takes a different approach. By looking at the mania for feedback in a historical context, Harkin finds a common thread in subjects as diverse as military strategy, TV shows like Lost, as well as the interwebs.

Q. We’re used to cyber-everything but can you define cybernetics for us?

Harkin: There are a lot of definitions but the simple idea I use is this idea that what distinguishes human beings, or what’ smost important about humans, is that they exist on a continuous information loop defined by a constant stream of messages we’re sending or receiving.

Now you can interpret the world in that way – me picking up a glass, say – but it is just a metaphor. The story of my book is how this metaphor, created by Norbert Wiener, because of its beauty, became the inspiration for a new medium and influencing how we live. It’s given rise to all this incredible technology, but the idea of fitting ourselves into that mould will mean we’re the losers.

The central image of the book is Cyburbia, this strange alternate world where we watch each other and the minutiae of each others’ lives.

You might have stared out of your window in suburbia in the 1950s and seen a few people across the street, but now you can stare at millions of other people. The danger is that when you spend all your time deciphering what other people are up to, you never get around to doing something original on your own, because you’re so swamped by opportunities to go onto other people’s lives on blogs, social networks and Twitter.

Read more

Twitter’s Jam Festival


Writing about Twitter is the journalistic equivalent of eating the fluff from your navel. The posh papers love it. Menopausal middle-aged hacks love it. The BBC is obsessed with it. Instead of telling us something we didn’t know before, Twitter makes churnalism so easy, it practically automates the entire job.

The rest of the world, however, completely ignores it. But with the journalists’ attention fixed firmly on each others’ navels, they don’t seem to realise what a fringe activity Twitter is. Now we can quantify this a little.

Read more

Google cranks up the Consensus Engine

Image from Google's 2006 analyst presentation

Google this week admitted that its staff will pick and choose what appears in its search results. It’s a historic statement – and nobody has yet grasped its significance.

Not so very long ago, Google disclaimed responsibility for its search results by explaining that these were chosen by a computer algorithm. The disclaimer lives on at Google News, where we are assured that:

The selection and placement of stories on this page were determined automatically by a computer program.

A few years ago, Google’s apparently unimpeachable objectivity got some people very excited, and technology utopians began to herald Google as the conduit for a new form of democracy. Google was only too pleased to encourage this view. It explained that its algorithm “relies on the uniquely democratic nature of the web by using its vast link structure as an indicator of an individual page’s value. ”

That Google was impartial was one of the articles of faith. For if Google was ever to be found to be applying subjective human judgment directly on the process, it would be akin to the voting machines being rigged.

For these soothsayers of the Hive Mind, the years ahead looked prosperous. As blog-aware marketing and media consultants, they saw a lucrative future in explaining the New Emergent World Order to the uninitiated. (That part has come true – Web 2.0 “gurus” now advise large media companies).

It wasn’t surprising, then, that when five years ago I described how a small, self-selected number of people could rig Google’s search results, the reaction from the people doing the rigging was violently antagonistic. Who lifted that rock? they cried.

But what was once Googlewashing by a select few now has Google’s active participation. This week Marissa Meyer explained that editorial judgments will play a key role in Google searches.

Read more