Authors: Nick Harkaway
So, what to do? How do you preserve a text stored in a
mutable form? Obviously, you duplicate it and store the copies all over the place. Wikileaks, again, understood that the only way to be certain their trove of cables could not be sequestered or destroyed was to distribute it widely. The same logic applies to the global digital library; a real Gutenberg moment for the digital age is not the creation of a single database with a single gatekeeper, but the birth of any number of digital libraries accessible all over the world. For greater security of the contents, we’d also vary the format in which they were stored, making it harder for a virus to wreak havoc on them all. (You might argue that that brings its own problems of access, and it’s true, but it is also true that it can be seen as insurance against the obsolescence of a given format. It’s an interesting sidebar.)
The main debate comes back, in the end, to
Marshall McLuhan: what is the message of a given technological choice? What does it mean to have a single digital library controlled from within the United States by a tech giant based in California, as opposed to having a vast number of libraries around the world, each of them containing the same media and each widely accessible? Which proposes democracy, and which leaves the door open to less amenable forms of governance? When we’re structuring our technologies – and the systems that support them and grow from them – we have to choose, over and over again, the path that emphasizes the society we want.
It isn’t possible – at least, it’s not humanly possible, and not technologically possible
yet –
to map perfectly the consequences of an attempt to incentivize or disincentivize a given behaviour. Governments attempt it constantly using tax regulation, and the results are rarely if ever what was intended. Part of the understanding regarding the bank bailouts in the UK was that the banks would start lending to businesses again to encourage growth, but that isn’t happening either. Incentives that seem clear-cut can be more complex when you put them in context. For example, returning to
Freakonomics
for a moment: capital
punishment is supposed to be a deterrent to serious crime. However, according to Levitt and Dubner, the annual execution rate among death row detainees in New York – and let’s not forget that a percentage of these will be innocent of the crime of which they are accused, so there’s a measurable chance that if you commit a crime, someone else will die for it – is 2 per cent. That compares favourably with the 7 per cent annual chance of dying just because a person is a member of the Black Gangster Disciple Nation crack gang. (And that’s before you factor in the point I made earlier, that a system of social controls relies on positives as well as negatives, and many of those who commit serious crimes may well have crossed that Rubicon already, feeling that society has no path for them.)
Getting incentives flowing in the right direction is hard. Levitt and Dubner also cite a study that showed mechanics allowing cars that failed a clean air standard to pass because lenient mechanics got repeat business: great for the car business, good for drivers whose need or desire for their car outweighs their understanding of pollution and its negative consequences, but bad for the economy because those negatives are probably more expensive and bad for public health (which obviously is also an economic factor).
That said, the other thing you can learn from reading
Freakonomics
is that looking hard at the information in an intelligent way will help you track down situations where the incentives are all wrong. And a digital era attitude to how systems work, one that says that no model is perfect first time, but that a model can be made better and better by refinement, might be able to make it work. It would probably require that we as human beings learn to function intelligently as a group rather than in our own individual short-term favour, something else that becomes easier as digital technology allows you to appreciate your place in a larger tapestry, and why it’s better for you to accept a short-term loss in quest for a longer term gain.
Fundamentally, in reaching for a better world, we have to code for it: we have to create the technologies that will foster the culture and institutions that will produce it, and legislate for the climate that will allow it. It is not enough to cross our fingers and hope, now that we no longer have to.
Choice is a new experience for us in the context of social change. We have until now been part of a process of change prompted by climatic shifts and migrations, famines, warlords and technological advances: not an inexorable progress towards a ‘higher form’ of social life, but a sort of ragged wandering. Furthermore, in case it isn’t obvious, the traits that are
evolutionarily successful are not always those that we would desire for our societies. In biological evolution, a shining example of the cruelty of the selfish gene is the chicken. All but flightless, also tasty and nutritious, the chicken is the ideal domesticated bird. In consequence, the reported global chicken population in 2003 was 24 billion, making it the most successful avian in biological history. Any given chicken, of course, can expect to be eaten at any time, and a vast number of them live in appalling conditions in factory farms. The DNA doesn’t care as long as they breed.
Similarly, if there is a process akin to evolution that applies to societies, it does not necessarily select for mass happiness. If the population’s unhappiness makes them more likely to survive hard conditions or destroy a competitor, that will do fine. Nor, indeed, is such a process a guarantee of our own society’s survivability. We could be a dead-end. Nonetheless, accounts of history have tended – sometimes unconsciously, sometimes with reference to a teleological theory like Marx’s historical materialism in which all roads eventually lead to perfect socialism – to suggest a kind of upward progress, and that notion is lodged in the popular consciousness.
Francis Fukuyama proposed the end of history in 1992, suggesting that henceforth all change would
be technological: the advent of liberal democracy meant the end of sociopolitical evolution. But it’s not clear why that should be the case. The neatness of teleological ideas of history and the future makes me suspect them; and then, too, they’re so self-congratulatory: we’re better in every way than those who came before us.
It seems more likely to me that
human social change to this point has been a drift, a series of actions and reactions, mostly without any sense of what the outcome of a given choice would be. Our grasp of what’s going on is still fragmentary, but we can increasingly spot correlations and trends, and, importantly, we can spot them while they’re happening rather than long after. Not all of them are immensely sophisticated; some are almost embarrassingly biological. For example, a recent study by researchers at Cornell University analysed half a billion postings from 2.5 million Twitter users, and established a common emotional shape to their days, varying by day of the week. In other words, there appears to be a predictable pattern of positivity at certain times of day, negativity at others, which extends across the eighty-four countries represented in the data.
4
It’s simple enough, but consider for a moment what it may mean to someone who is trying to sell you something. A study of different kinds of people will reveal whether they buy when they’re happy because they feel confident or when they’re feeling tired and cranky because it cheers them up. Advertising that reaches the right person at the right moment will be more effective, especially in a connected world where purchase can happen at almost the moment of desire. Cross-referenced with a closer understanding of the kind of person someone is gleaned from their social media information, and a database of like and dislike matches, perhaps a credit score, and you have a slightly alarming persuasion engine which knows when and where we are most vulnerable to an ‘impulse purchase’, or can guess how to get a more stolid soul to choose ‘yes’ in the longer term.
Leaving aside the obnoxiousness of a system that can get you to fill your house with consumer junk you might have decided you didn’t need if left alone, we’re presently painfully well acquainted with the macro-effects of irrational and over-aspirational decisions taken under the tutelage of corporate institutions such as credit card companies and banks who were perhaps looking to their own bottom line rather more than to the best possible advice for the individual. The wrong kind of successful persuasion could leave us with even more ghastly financial wreckage.
At this moment, there is a kind of arms race taking place as we learn more about ourselves and what influences us – and we, as individuals, are definitely not in the lead. The world’s big companies, and most particularly those that control or have access to Internet data, presently know more about how we buy than we do, because, as in Google’s case, or Facebook’s or Amazon’s, the business of their day-to-day function entails the gathering of huge amounts of raw data which they have the know-how to collate. That’s to say that they know more about what influences us to buy, how we make the final ‘yes’ decision, and conversely what’s likely to prevent us from doing so. They know that people do, ridiculously, buy more items priced at 99p than they would items priced at £1 (in behavioural economics, the phenomenon is called ‘
left-digit bias’). They know that we take undue notice of anchor prices and indeed of random numbers that look similar to prices in assessing whether something gives good value. They formulate ways of saving such as
Amazon Prime, which for a single annual payment gives you ‘free’ one-day delivery – a bargain that results in people buying more and more from Amazon to offset the fee, so that what appears to be to the company’s disadvantage ultimately works out in Amazon’s favour.
Like casinos, large corporate entities have studied the numbers and the ways in which people respond to them. These are not
con tricks – they’re not even necessarily against our direct interests, although sometimes they can be – but they are hacks for the human mind, ways of manipulating us into particular decisions we otherwise might not make. They are also, in a way, deliberate underminings of the core principle of the free market, which derives its legitimacy from the idea that informed self-interest on aggregate sets appropriate prices for items. The key word is ‘informed’; the point of behavioural economics – or, rather, of its somewhat buccaneering corporate applications – is to skew our perception of the purchase to the advantage of the company. The overall consequence of that is to tilt the construction of our society away from what it should be if we were making the rational decisions classical economics imagines we would, and towards something else.
Governments, too, are looking to the technologies of ‘nudging’ to push us towards certain behaviours and away from others, trying, perhaps, to push back the ghost of the classical Greek orator
Isocrates: ‘
Democracy destroys itself because it abuses its right to freedom and equality. Because it teaches its citizens to consider audacity as a right, lawlessness as a freedom, abrasive speech as equality, and anarchy as progress.’ But the idea of nudging really means subconsciously influencing the voting population to things they might otherwise not choose, rather than arguing the case directly. There’s a very fine line here: the
feedback system of road signs that tell you your speed is, obviously, a hint to slow down, but it is overt and leaves you with a clear choice. The classic example of ‘nudging’ is making a decision ‘opt out’ rather than ‘opt in’ – organ donation, for instance, or indeed the original
Google Book Settlement. In a governmental context it’s an Orwellian style of dictatorship that, by presenting heavily skewed choices or options that involve irritating effort versus others that do not and avoiding genuinely difficult balanced choices between two possible courses of action, effectively makes the population practise docility until it becomes
nicely habitual. Like cross-subsidization, ‘libertarian paternalism’ is the quiet death of genuine choice, the creation of a society in which a class of supposed experts rules over a common herd. It also throws away any hope of using the power of the crowd to make good decisions, because it supplies the crowd with bad basic information.
But information about our behaviour is also, increasingly, available to us. It is possible to become what some psychologists call ‘test-conscious’: aware of the tricks, and therefore while not immune to them at least capable of understanding that they are being played and looking for ways of compensating. Once again, we’re discussing a kind of feedback, and once again, it comes from our new ability to watch ourselves in the digital mirror at various different levels – as individuals or as large groups. We can learn to use our knowledge of what will influence us to nudge ourselves towards good decisions, thus creating a virtuous cycle. We need, ultimately, to favour the creation of systems around us that foster informed, intelligent decisions, not the ones that increase the profit margin of companies at the expense of our own good outcomes or bow automatically to the will of a government that may or may not know what it’s doing. In the shorter term, we need to learn to make good choices even though we are also irrational, and to educate and train ourselves in the habit of decision.