Calum Chace's Blog, page 11

December 9, 2018

Reviewing last year’s AI-related forecasts

[image error]


As usual, I made some forecasts this time last year about how AI would change, and how it would change us. It’s time to look back and see how those forecasts for 2018 panned out. The result: a 50% success rate, by my reckoning. Better than the previous year, but lots of room for improvement. Here are the forecasts, with my verdicts in italics.


1. Non-tech companies will work hard to deploy AI – and to be seen to be doing so. One consequence will be the growth of “insights-as-a-service”, where external consultants are hired to apply machine learning to corporate data. Some of these consultants will be employees of Google, Microsoft and Amazon, looking to make their open source tools the default option (e.g. Google’s TensorFlow, Microsoft’s CNTK, Amazon’s MXNet).


Yes. The conversation among senior business people at the events I speak at has moved from “What is this AI thing?” to “Are we moving fast enough?”


2. The first big science breakthrough that could not have been made without AI will be announced. (I stole this from DeepMind’s Demis Hassabis. Well, I want to get at least one prediction right!)


Yes. In May, an AI system called Eve helped researchers at Manchester University discover that triclosan, an ingredient commonly found in toothpaste, could be a powerful anti-malarial drug. The research was published in the journal Scientific Reports (here) .


3. There will be media reports of people being amazed to discover that a customer service assistant they have been exchanging messages with is a chatbot.


Yes. Google Duplex


4. Voice recognition won’t be quite good enough for most of us to use it to dictate emails and reports – but it will become evident that the day is not far off.


Yes. Alexa is pretty good, but not yet a reliable stenographer. (Other brands of AI assistant are available.)


5. Some companies will appoint Chief Artificial Intelligence Officers (CAIOs).


Not sure. I don’t know of any, but I bet some exist.


6. Capsule networks will become a buzz word. These are a refinement of deep learning, and are being hailed as a breakthrough by Geoff Hinton, the man who created the AI Big Bang in 2012.


Not as far as I know.


7. Breakthroughs will be announced in systems that transfer learning from one domain to another, avoiding the issue of “catastrophic forgetting”, and also in “explainable AI” – systems which are not opaque black boxes whose decision-making cannot be reverse engineered. These will not be solved problems, but encouraging progress will be demonstrated.


I think I’ve seen reports of progress, but nothing that could fairly be described as a major breakthrough.


8. There will be a little less Reverse Luddite Fallacism, and a little more willingness to contemplate the possibility that we are heading inexorably to a post-jobs world – and that we have to figure out how to make that a very good thing. (I say this more in hope than in anticipation.)


No, dammit.

 •  0 comments  •  flag
Share on Twitter
Published on December 09, 2018 10:51

November 18, 2018

Book review: “21 Lessons for the 21st Century”, by Yuval Harari

[image error]


The title of Yuval Harari’s latest best-seller is a misnomer: it asks many questions, but offers very few answers, and hardly any lessons. It is the least notable of his three major books, since most of its best ideas were introduced in the other two. But it is still worth reading. Harari delights in grandiloquent sweeping generalisations which irritate academics enormously, and part of the fun is precisely that you can so easily picture his colleagues seething with indignation that he is trampling on their turf. More important, some of his generalisations are acutely insightful.


The insight at the heart of “Sapiens”, his first book, was that humans dominate the planet not because we are logical, but because 70,000 or so years ago we developed the ability to agree to believe stories that we know are untrue. These stories are about religion, and political and economic organisation. The big insight in his second book, “Homo Deus” is that artificial intelligence and other technologies are about to transform our lives far more – and far more quickly – than almost anyone realises. Both these key ideas are reprised in “21 Lessons”, but they are big ideas which bear repeating.


Happily, he has toned down his idiosyncratic campaigns about religion and vegetarianism. In the previous books he encountered religion everywhere: capitalism and communism have passionate adherents, but they are not religions. The first third of “Homo Deus” is religious in a different way: it is a lengthy sermon about vegetarianism.


[image error]


“21 Lessons” is divided into five parts, of which the first is the most coherent and the best. It concerns the coming technological changes, which Harari first explored in “Homo Deus”. “Most people in Birmingham, Istanbul, St Petersburg and Mumbai are only dimly aware, if at all, of the rise of artificial intelligence and its potential impact on their lives. It is undoubtable, however, that the technological revolutions will gather momentum in the next few decades, and will confront humankind with the hardest trials we have ever encountered.”


He is refreshingly blunt about the possibility of technological unemployment: “It is dangerous just to assume that enough new jobs will appear to compensate for any losses. The fact that this has happened during previous waves of automation is absolutely no guarantee that it will happen again under the very different conditions of the twenty-first century. The potential social and political disruptions are so alarming that even if the probability of systemic mass unemployment is low, we should take it very seriously.”


Very well said, but this part of the book would be much more powerful if he had offered a fully worked-through argument for this claim, which in the last couple of years has been sneeringly dismissed by a procession of tech giant CEOs, economists, and politicians. Perhaps next year, the World Economic Forum could organise a debate on this question between Harari and a leading sceptic, such as David Autor.


It is also a shame that he offers no prescriptions, beyond categorising them: “Potential solutions fall into three main categories: what to do in order to prevent jobs from being lost; what to do in order to create enough new jobs; and what to do if, despite our best efforts, job losses significantly outstrip job creation.” Fair enough, but this should be the start of the discussion, not the end. Still, at least he doesn’t fall back on the usual panacea of universal basic income, and his warning about what happens if we fail to develop a plan is clear: “as the masses lose their economic importance … the state might lose at least some of the incentive to invest in their health, education and welfare. It’s very dangerous to be redundant.”


Harari is also more clear-sighted than most about the risk of algocracy – the situation which arises when we delegate decisions to machines because they make better ones than we do. “Once we begin to count on AI to decide what to study, where to work, and who to marry, human life will cease to be a drama of decision-making. … Imagine Anna Karenina taking out her smartphone and asking the Facebook algorithm whether she should stay married to Karenin or elope with the dashing Count Vronsky.” Warning about technological unemployment, he coined the brutal phrase “the gods and the useless”. Warning about algocracy, he suggests that humans could become mere “data cows”.


[image error]


The remaining four parts of the book contain much less that is original and striking. Harari is a liberal and an unapologetic globalist, pointing out reasonably enough that global problems like technological disruption require global solutions. He describes the EU as a “miracle machine”, which Brexit is throwing a spanner into. He does not see nationalism as a problem in itself, although he observes that for most of our history we have not had nations, and they are unnatural things and hard to build. In fact he thinks they can be very positive, but “the problem starts when benign patriotism morphs into chauvinistic ultra-nationalism.”


Although he sees nationalism as a possible problem, he also thinks it has already lost the game: “we are all members of a single rowdy global civilisation … People still have different religions and national identities. But when it comes to the practical stuff – how to build a state, an economy, a hospital, or a bomb –almost all of us belong to the same civilisation.” He supports this claim by pointing out that the Olympic Games, currently “organised by stable countries, each with boringly similar flags and national anthems,” could not have happened in mediaeval times, when there were no such things as nation states. And he argues that this is a very good thing: “For all the national pride people feel when their delegation wins a gold medal and their flag is raised, there is far greater reason to feel pride that humankind is capable of organising such an event.”


[image error]


He is even more dismissive of religion – especially monotheism – despite his obsession with it. “From an ethical perspective, monotheism was arguably one of the worst ideas in human history … What monotheism undoubtedly did was to make many people far more intolerant than before … the late Roman Empire was as diverse as Ashoka’s India, but when Christianity took over, the emperors adopted a very different approach to religion.” Religion, he says, has no answers to any of life’s important questions, which is why there is no great following for a Christian version of agriculture, or a Muslim version of economics. “We don’t need to invoke God’s name in order to live a moral life. Secularism can provide us with all the values we need.”


He seems to be applying for membership of the “new atheists” club, in which Richard Dawkins and Stephen Pinker deliberately goad the religious by diagnosing religion as a disease which can be cured. Harari suggests that “when a thousand people believe some made-up story for one month, that’s fake news. When a billion people believe it for a thousand years, that’s a religion.”


[image error]


Oddly, given his perceptive take on the future of AI, Harari is weak on science fiction, displaying a fundamental misunderstanding of both The Matrix and Ex Machina. He is stronger on terrorism, pointing out that it is much less of a threat than it seems, contrary to the deliberate mis-representations by populists: “Since 11 September 2001, every year terrorists have killed about fifty people in the European Union, about ten people in the USA, about seven people in China, and up to 25,000 people globally (mostly in Iraq, Afghanistan, Pakistan, Nigeria and Syria).  In contrast, each year traffic accidents kill about 80,000 Europeans, 40,000 Americans, 270,000 Chinese, and 1.25 million people altogether.” Terrorists “challenge the state to prove it can protect all its citizens all the time, which of course it can’t.” They are trying to make the state over-react, and populists are their eager accomplices.


The book seems to be building to a climax when it addresses the meaning of life. Here and elsewhere, Harari has said that humans create meaning – or at least the basis of power – by telling ourselves stories. So is he going to give us a story which will help us navigate the challenges of the 21st century?


Sadly not. The closest we get is a half-baked version of Buddhism.


“The Buddha taught that the three basic realities of the universe are that everything is constantly changing, nothing has any enduring essence, and nothing is completely satisfying.  Suffering emerges because people fail to appreciate this … The big question facing humans isn’t ‘what is the meaning of life?’ but rather, ‘how do we get out of suffering?’ … If you really know the truth about yourself and about the world, nothing can make you miserable. But that is of course much easier said than done.” Indeed.


[image error]


Harari has worked out his own salvation: “Having accepted that life has no meaning, I find meaning in explaining this truth to others.” Given his six-figure speaking fees, this makes perfect sense.


Harari also finds solace in meditation, which he practices for two hours every day, and a whole month or two every year. “21 Lessons” is a collection of essays written for newspapers and in response to questions. This shows in its disjointed, discursive, and inconclusive nature. If Harari had spent less time meditating, maybe he would have found more time to answer the questions he raises. It’s still definitely worth reading, though. 


This article first appeared in Forbes Magazine

 •  0 comments  •  flag
Share on Twitter
Published on November 18, 2018 02:36

September 14, 2018

Shooting the Messenger

[image error]“It was Facebook wot dunnit.”


Select the unpleasantness of your choice, and Facebook is almost certainly being blamed for it by someone, and probably a lot of someones. Also in the dock are YouTube and Twitter, with Instagram and Snapchat lurking about, keeping their heads down and hoping that nobody notices them.


The charge sheet is long. Facebook and the other social media have shortened our attention spans, leaving us easy prey to slick salesmen with plausible one-liners. They have corralled us all into echo chambers, so that we only ever hear voices telling us what we already think. We are now all isolated from the wider community. They sneakily deploy algorithms of such breathtaking sophistication that they can delve inside our neurons and suck out the information that constitutes their marrow, and then use that information against us like master hypnotists. OK, maybe I made that last bit up. But they definitely plug into our neocortex and inject a sort of digital heroin, forcing us, like enslaved machines, to spend hours online, clicking away to generate ad money.


It is because of all this skull-duggery (pun intended) that political discussion has become so heated, and people aren’t listening to each other any more.


Well, it has to be someone’s fault, doesn’t it, and the tech giants who own the social media are uniquely well-placed to attract universal opprobrium. For people on the political left, they are large companies which make a lot of money. For many on the the left, capitalism is a conspiracy against the masses, profit is a Bad Thing, and a large profit is a Very Bad Thing Indeed. For people on the political right, the tech giants are run by suspiciously hippy-ish types, who give away their money and talk about Universal Basic Income. Their employees are a bunch of snowflake lefties who cannot bear to work with the military, and who excoriate anyone who doesn’t share their hatred of the patriarchy, and who dares to question the wisdom of affirmative action hiring and training policies. And people from both political wings can hold hands while condemning the tech giants for not paying enough tax.


[image error]


The mainstream media is also furious with the tech giants, and understandably so: Google and Facebook stole their lunch. Local newspapers grew fat on a diet of classified ads, but these were the first casualty of the web. National newspapers depended much less on classified ads and more on display ads and cover prices, but these have also dwindled, as advertisers have discovered the charms of paying only for eyeballs which have actually scanned their messages, and which can be micro-targeted to make those messages more relevant.


This is not something to make light of. If today’s febrile political atmosphere tells us anything, it is that we need professional journalists who have genuinely mastered their craft, and who care about getting the story right as well as getting it first. We are far from figuring out all the business models we need in order to pay for this, and one way or another, pay we must.


I’m not here to defend the tech giants. They are very rich, they hire great lobbyists, and they can look after themselves. (By way of disclosure, I have never used Facebook, as I don’t trust myself not to lose whole afternoons, chatting with friends and looking at cat videos. I think Twitter and Reddit are fabulous, and LinkedIn is handy, although inexplicably clumsy.) But mis-diagnosing major social problems simply allows those problems to fester while causing new ones, and it is mis-diagnosis on a grand scale to blame social media for today’s vicious style of political debate. It implies that there was a halcyon past when the electorate took great care to inform itself of the arguments from both sides, thought deeply about the philosophical underpinnings of each position, and arrived at a sophisticated understanding of the issues of the day. In reality, before social media came along, people lived in the political echo chamber constructed by their newspaper of choice. Guardian readers didn’t check out the talking points promoted by the Daily Telegraph or vice versa, and likewise the Sun and the Daily Mirror. And if you’re looking for examples of the blatant exploitation of human appetites to boost sales, remember it was only in 2015 that The Sun stopped publishing photos of topless young women on page three.


The real reason for the bitterness of today’s political conversation is not the arrival of social media. It is the considerable success of the liberal social agenda.


The left is always angry. That is as it should be: if the left isn’t angry, then it’s not doing its job. The job of the left is to make people discontented with the current state of affairs, and to agitate to improve it. The world can never be equal enough or just enough, and it is the left’s job to keep pushing it in the right direction. The right, traditionally, is more relaxed. By and large it thinks that the world is in pretty good condition, that the institutions are doing a reasonable job, and that throwing all the cards in the air and risking anarchy is a terrible idea. Generally, both have a very good point. Societies should change: they should look for ways to solve problems, but they should do so in ways that will actually improve the lives of citizens, not to accommodate ideological whims. They should recognise that ordinary citizens today live better lives than the kings of a couple of centuries ago, and that over-zealous radicalism has caused at least as much misery as any other social force. Apart from religion, of course.


[image error]
But something changed at the end of the first decade of the twenty-first century. Left-wingers complain that global economic policy has long been dominated by right-wing neoclassical orthodoxy. Be that as it may, during the end of the 20th century and the start of the 21st century, social policy in the developed world was driven strongly in a liberal direction. Strongly and quickly. Governments in developed countries now mostly spend between a third and two-fifths of GDP, and much of that spend is on social and welfare support. The treatment of homosexuals, women, and minority races has been greatly improved. Citizens were relieved of interference in their intimate lives by church and state, while on the other hand, health and safety officers worked to make construction sites and consumer goods less likely to kill and maim people.


As the economies of developed countries grew, their people became wealthier. This made them have fewer children, and made them less willing to do menial work for low wages. This spurred new waves of immigration, and innumerable studies have shown that economic migration is a boon for both the country receiving the immigrant and the one they came from.


All this means change, and change is uncomfortable. And when immigration is from countries with sharply different cultures, and perhaps different skin colours, it is more obvious, and more uncomfortable. If women, gays, and ethnic minorities are advancing, the previously privileged populations might not do worse in absolute terms, but their privilege is undermined, and less reassuring.


[image error]It is no coincidence that the creation of the Tea Party in the US, the moment when the right started to get cross, was in 2009, the year of Obama’s inauguration. The election of a black president was perhaps the apogee of liberal values. It also coincided with the credit crunch and the start of the prolonged recession which that caused – a source of further discontent all round.


The extremists, the alt right and the Saundernistas, have a spring in their step, and the squishy middle which is more-or-less neoclassical in economics and liberal in social values, is looking weak. This is an important battle, and it will take some years for its fog to lift.


Meanwhile, what of social media? If we accept they are a messenger that shouldn’t be shot, we shouldn’t just sit back and relax. We can do much better than leaving the present undifferentiated mess of facts and lies, thoughtful opinion and rabid conspiracy theory to contend on equal terms for the attention of the unwary. Editorials in the mainstream media call for social media platforms to be treated like media and to be regulated as such. But media regulation, whether by government or by industry bodies, is ponderous and generally timid. Thanks to the magic of the web, we can do better.


Reddit gives an idea of how. Many of the posts of Reddit are based on links to newspapers, magazines and other outlets, to which readers add comments – sometimes dumb, often insightful, and occasionally hilarious. Reddit automatically rates each source, and readers vote each others’ comments up or down. The most highly up-voted posts appear at the top of the page. Wikipedia is another site that crowd-sources opinion very effectively to fact-check and verify.


Using these ideas and new ones, and with the thousands of clever technologists and user experience designers they employ, the tech giants can semi-automate fact-checking, and over time, make social media better than any media or platform we have known so far. The process of getting there won’t be fast, and it will be messy: the likes of TrustPilot and TripAdviser are afflicted with false reviews on an almost industrial scale. But if problems like this are soluble, and they probably are, then we can have media and platforms where readers can easily assess the veracity of any piece of content, guided by judgements which even out and transcend partisan opinion.


Looking ahead, the tech giants are going to have the mother of all PR problems when AI-powered automation starts causing job churn, but for the time being, if we can refrain from shooting the messenger, maybe we can take the mess out of the message.


[image error]

 •  0 comments  •  flag
Share on Twitter
Published on September 14, 2018 03:42

August 22, 2018

Algocracy

[image error]Powerful new technologies can produce great benefits, but they can often produce great harm. Artificial intelligence is no exception. People have numerous concerns about AI, including privacy, transparency, security, bias, inequality, isolation, oligopoly, and killer robots. One which perhaps gets less attention than it deserves is algocracy.


Decisions about the allocation of resources are being made all the time in societies, on scales both large and small. Because markets are highly efficient systems for allocating resources in economies characterised by scarcity, capitalism has proved highly effective at raising the living standards of societies which have adopted it. Paraphrasing Churchill, it is the worst possible economic system except for all the others.


Historically, markets have consisted of people. There may be lots of people on both sides of the transaction (flea markets are one example, eBay is another). Or there may be few buyers and many sellers (farmers selling to supermarket chains) or vice versa (supermarket chains selling to consumers). But typically, both buyers and sellers were humans. That is changing. 


Machine-made decisions


[image error]Algorithms now take many decisions which were formerly the responsibility of humans. They initiate and execute many of the trades on stock and commodity exchanges. They manage resources within organisations providing utilities like electricity, gas and water. They govern important parts of the supply chains which put food on supermarket shelves. This phenomenon will only increase.


As our machines get smarter, we will naturally delegate decisions to them which would seem surprising today. Imagine you walk into a bar and see two attractive people at the counter. Your eye is drawn to the blond but your digital assistant (located now in your glasses rather than your phone) notices that and whispers to you, “hang on a minute: I’ve profiled them both, and the red-head is a much better match for you. You share a lot of interests. Anyway, the blond is married.”


In his 2006 book “Virtual Migration”, Indian-American academic A. Aneesh coined the term “algocracy”. The difficulty with it has been explored in detail by the philosopher John Danaher, who sets the problem up as follows. Legitimate governance requires transparent decision-making processes which allow for involvement by the people affected. Algorithms are often not transparent and their decision-making processes do not admit human participation. Therefore algorithmic decision-making should be resisted.


Danaher thinks that algocracy poses a threat to democratic legitimacy, but does not think that it can be, or should be, resisted. He thinks there will be important costs to embracing algocracy and we need to decide whether we are comfortable with those costs.


What not to delegate?


[image error]Of course many of the decisions being delegated to algorithms are ones we would not want returned to human hands – partly because the machines make the decisions so much better, and partly because the intellectual activity involved is deathly boring. It is not particularly ennobling to be responsible for the decision whether to switch a city’s street lights on at 6.20 or 6.30 pm, but the decision could have a significant impact. The additional energy cost may or may not be offset by the improvement in road safety, and determining that equation could involve collating and analysing millions of data points. Much better work for a machine than a human, surely.


Other applications make us much less sanguine. Take law enforcement: a company called Intrado provides an AI scoring system to the police in Fresno, California. When an emergency call names a suspect, or a house, the police can “score” the danger level of the person or the location and tailor their response accordingly. Other forces use a “predictive policing” system called PredPol which forecasts the locations within a city where crime is most likely to be carried out in the coming few hours. Optimists would say this is an excellent way to deploy scarce resources. Pessimists would reply that Big Brother has arrived.


AI is already helping to administer justice after the event. In 2016 the San Francisco Superior Court began using an AI system called PSA to determine whether parole should be given to alleged offenders. They got the tool free from the John and Laura Arnold Foundation, a Texas-based charity focused on criminal justice reform. Academics studying this area have found it very hard to obtain information about how these systems work: they are often opaque by their nature, and they are also often subject to commercial confidentiality.


There are many decisions which machines could make better than humans, but we might feel less comfortable having them do so. The allocation of new housing stock, the best date for an important election, the cost ceiling for a powerful new drug, for instance. Arguments about which decisions should be made by machines, and which should be reserved for humans are going to become increasingly commonly and increasingly vehement. Regardless whether they make better decisions than we do, not everyone is going to be content (to paraphrase Grace Jones) to be a slave to the algorithm.


Information is power. Machines may intrude on our freedom without actually making decisions. In September 2017 a research team from Stanford University was reported to have developed an AI system which could do considerably more than just recognise faces. It could tell whether their owners were straight or gay. The idea of a machine with “gaydar” is startling; it becomes shocking when you consider the uses it might be put to – in countries where homosexuals are persecuted and even prosecuted, for instance. The Stanford professor who led the research later said that the technology would probably soon be able to predict with reasonable accuracy a person’s IQ, their political inclination, or their predisposition towards criminality.


Things are getting weird.



https://www.washingtonpost.com/local/public-safety/the-new-way-police-are-surveilling-you-calculating-your-threat-score/2016/01/10/e42bccac-8e15-11e5-baf4-bdf37355da0c_story.html




. https://www.wired.com/story/when-government-rules-by-software-citizens-are-left-in-the-dark/




. https://www.wired.com/story/when-government-rules-by-software-citizens-are-left-in-the-dark/




. https://www.theguardian.com/technology/2017/sep/07/new-artificial-intelligence-can-tell-whether-youre-gay-or-straight-from-a-photograph




http://philosophicaldisquisitions.blogspot.co.uk/2014/01/rule-by-algorithm-big-data-and-threat.html



 

 •  0 comments  •  flag
Share on Twitter
Published on August 22, 2018 06:05

August 7, 2018

Just Another Day in Utopia

a Guest Post by Stuart Armstrong

Stuart is a Fellow in Artificial Intelligence and Machine Learning at Oxford University’s Future of Humanity Institute. He works on how to map humanity’s partially defined values onto the potential goals of AI. He is probably best known for his collaboration with DeepMind on how to stop a developing AI resisting being switched off after showing signs of going rogue – see here .


He has also written some excellent short science fiction stories, such as the one this is excerpted from , which seems to me an admirable blend of Iain M. Banks and Roger Williams’ 1994 novella, “The Metamorphosis of Prime Intellect”. There are links to the full version, and to another of Stuart’s other stories at the end.


[image error]


Ishtar went to sleep in the arms of her lover Ted, and awoke locked in a safe, in a cargo hold of a triplane spiralling towards a collision with the reconstructed temple of Solomon.


Again! Sometimes she wished that a whole week would go by without something like that happening. But then, she had chosen a high excitement existence (not maximal excitement, of course – that was for complete masochists), so she couldn’t complain. She closed her eyes for a moment and let the thrill and the adrenaline warp her limbs and mind, until she felt transformed, yet again, into a demi-goddess of adventure. Drugs couldn’t have that effect on her, she knew; only real danger and challenge could do that.


Right. First, the safe. She gave the inner door a firm thud, felt it ring like a bell, heard the echo return – and felt the tumblers move. So, a sound-controlled lock, then. A search through her shoes produced a small pebble which sparked as she dashed it against the metal. Trying to ignore the ominous vibration as the triplane motor shook itself to pieces, she constructed a mental image of the safe’s inside from the brief flashes of light. Symmetric gold and gilded extravagances festooned her small prison – French Baroque decorations, but not yet Roccoco. So Louis XIV period. She gave the less visited parts of her mind a good dusting, trying to remember the tunes of Jean Batiste Lully, the period’s most influential composer. She hoped it wasn’t any of his ballets; she was much better with his operas. The decorations looked vaguely snake-like; so she guessed Lully’s ‘Persée’ opera, about the death of the Medusa.


The engine creaked to a worrying silence as she was half-way through humming the Gorgon theme from the opera. Rushing the rest of the composition, she felt the door shift, finally, to a ten-times speeded up version of Andromeda’s response to Perseus’s proposal. She kicked the door open, exploded from the safe, took in the view of the temple of Solomon rushing up towards her, seconds away, snatched a picture from the floor, grabbed an axe from the wall, hacked off one of the wings with three violent cuts, and jumped out of the plane after it.


Behind her, the plane disintegrated in mid-air as the temple lasers cut it to shreds and she fell through space, buffeted by the wind, not losing her grip on the mangled wing. She had maybe thirty seconds to tie herself to the wing, using the object’s own canvas as binding, and she rushed through that. The Machines wouldn’t allow the fall to kill her, of course, but it would hurt quite a bit (another of her choices – she’d allowed herself to feel moderate amounts of pain). It would put back her attempts to ever find Ted, and, most importantly of all, it would be crushingly embarrassing socially.


Once she was lashed to the plummeting piece of wood and canvas, and she was reasonably confident that the fall was slow enough, and her knots secure enough, she finally looked at the photograph she had grabbed during her explosive exit from the plane. It showed Ted, trussed up in chains but smiling and evidently enjoying the novel experience. Underneath was a finely engraved note: “If you ever want to see your lover again, bring me the missing Stradivarius by noon tomorrow. Nero the 2nd”. Each capital letter was beautifully decorated with heads on spikes.


So! It seemed that her magnificent enemy Nero had resorted to kidnapping in order to get his way. It wasn’t as if Nero could actually harm Ted – unlike Ishtar, her lover had never chosen to accept any level of pain above mild, brief discomfort. But if he was ‘killed’, Ted would feel honour-bound to never see her again, and she wasn’t prepared to accept that. On the other hand, if she gave Nero her last Stradivarius, he might destroy it for good. It was her own choice: she had requested that her adventures have real meaning, with real consequences. If she failed, and if Nero so choose, a piece of humanity’s cultural history could be destroyed forever, permanently stymieing her attempts to reconstruct Stradivarius’s violin-making techniques for the modern world. Culture or love, what to choose? Those were her final thoughts before she crashed into an oak tree shaped like a duck.


She returned to bleary consciousness fifteen minutes later. Her fainting was a sign that the Machines had judged her escape attempt to be only a partial success; she would have to try harder next time. In the meantime, however, she would have to deal with shotgun pressed into her face and the gorgeous man at the other side of it shouting “Get off my property!”.


[image error]


“Pause,” she said softly. The man nodded; she had temporarily paused her adventure, so that she wouldn’t have to deal with danger or pursuit for the next few minutes, and so that this guy wouldn’t have to get her away immediately to protect his property from collateral damage. Some Adventurers disdained the use of the pause, claiming it ruined the purity of their experience. But Ishtar liked it; it gave her the opportunity, as now, of getting to know the people she bumped into. And this person definitely seemed to be in the ‘worth getting to know’ category. He put down his shotgun without a word and picked up his paintbrush, applying a few more touches to the canvas in front of him.


After disengaging herself from both the mangled wing and the duck-shaped tree (she’d have a dramatic scar from that crash, if she chose to accept it), she worked her way round to what he was painting. It was a rather good neo-impressionistic portrait of her, unconscious in the tree, pieces of torn canvas around her, framed by broken branches and a convenient setting moon. Even with his main subject out of the frame, as it were, he still seemed intent on finishing his painting.


“Why did you splice your tree’s genes to make it look like a duck?” she asked, when the silence had gone on, in her estimation, for ten times as long as it should have. He had done a pretty good job with that oak, in fact; the feathers and the features were clear and distinct amongst the wood – or had been, until someone had crashed a triplane wing into the middle of it.


“I didn’t,” he said. “That’s normal oak; I just trim it and tie it.”


“But…” she looked at it again in astonishment; the amount of work involved to get that detail from natural wood was beyond belief. And oak wasn’t exactly a fast growing plant… “It must have taken you decades!”


“Two centuries,” he answered with dour satisfaction. “All natural, no help from the Machines.” He waved his hand up the side of the hill. “I’m making the perfect landscape. And then, I shall paint it.”


Her gaze followed his hand. The scenery was a tapestry of secret themes. Hedges, streams, tree-rows, pathways, ridges and twined lianas carved the landscape into hidden pockets of beauty. Each pocket was a private retreat, cut off from the others and from the rest of the world – and yet all were visible at once, the layout a cunning display of multiple intimacy. Here and there were formal gardens, with lines of flowers all at attention, row after row, shading across colour and size from huge orchids to tiny snowdrops. Some pockets were carefully dishevelled, mini deserts or prairies or jungles, perfect fragments of wild untamed nature that could only exist at the cost of supreme artifice. There were herb gardens, rock gardens, orchards, water parks and vineyards; modelled on ancient Persia, England, Japan, France, Korea, Spain, the Inca and Roman empires – and others she didn’t immediately recognise.


And then a few touches of fancy, such as the segment they were in, with the oaks shaped into animals. Further off, a dramatic slew of moss-coated sculptures, with water pouring out from every nook and cranny. Then a dynamic garden, with plants blasting each other with discharges of pollen, set-up in a simple eight-beat rhythm. And a massive Baobab, its limbs plated with a forest of tiny bonsai trees.


“What’s your safety level for all this?” she asked. If he’d chosen total safety, he wouldn’t have needed her off his property, as the Machines wouldn’t have allowed his creations to be damaged by her adventure. But surely he wouldn’t have left such artistic creation vulnerable to the fallout of Adventurers or random accidents…


 “Zero,” he said.


“What?” No-one choose zero safety; it just wasn’t done.


“As I said, no help from the Machines.” He looked at her somewhat shyly, as she stared in disbelief. “It’s been destroyed twice so far, but I’ll see it out to the end.”


No wonder he’d wanted her out… He only had himself to count on for protection, so he had to chase out any potential disturbances. She felt deeply moved by the whole grandiose, proud and quixotic project. Acting almost – almost – without thinking, she drew out a battered papyrus scroll: “Can you keep this for me?”


“What is it?” he asked, before frowning and tearing up his painting with a sigh. Only then did he look at the scroll, and at her.


“It’s my grandfather’s diary,” she said, “with my own annotations. It’s been of great use and significance to me.” Of course it had been – the Machines would have gone to great pains to integrate such a personal and significant item deeply into her adventures. “Could you keep it for my children?” When she finally found the right person to have them with, she added mentally. Ever since her split with Albert… No, that was definitely not what she needed to be thinking right now. Focus instead on this gorgeous painter, name still unknown, and his impossible dreams.


“What was he like?” he asked.


“My grandfather? Odd, and a bit traditional. He brought me up. And when we were all grown up, all his grandchildren, he decided we needed, like in ancient times, to lose our eldest generation.”


“He died?” The painter sounded sceptical; there were still a few people choosing to die, of course, but those events were immensely rare and widely publicised.


“No, he had his intelligence boosted. Recursively. Then he withdrew from human society, to have direct philosophical conversations with the Machines.”


He thought for a while, then took the scroll from her, deliberately brushing her fingers as he did so. “I’ll keep this. And I’m sure your children will find their ways to me.” An artefact, handed down and annotated through the generations, and entrusted in a quirky landscape artist who laboured obsessively with zero safety level? It was such a beautiful story hook, there was no way the Machines wouldn’t make use of it. As long as one of her children had the slightest adventurous streak, they’d end up here.


“This feels rather planned,” he said. “I expect it’s not exactly a coincidence you ended up here.”


“Of course not.” He was reclusive, brilliant, prickly; Ishtar realised a subtle seduction would be a waste of time. “Shall we make love?”, she asked directly.


“Of course.” He motioned her towards a bed of soft blue moss that grew in the midst of the orchids. “I have to warn you, I insist that the pleasure-enhancing drugs we use be entirely natural, and picked from my garden. Let me show you around first, and you can make your choice.” They wandered together through the garden, shedding their clothes and choosing their pleasures.


Later, after love, she murmured “unpause” before the moment could fade. “Get off my property!” he murmured, then kissed her for the last time. She dived away, running from the vineyard and onto the street, bullets exploding overhead and at her feet.


Three robot gangsters roared through the street in a 1920 vintage car, spraying bullets from their Tommy guns. The bullets ricocheted off the crystal pavement and gently moving wind-houses, causing the passers-by (all of whom had opted for slight excitement that week) to duck enthusiastically to the floor, with the bullets barely but carefully missing them. Diving round a conveniently placed market stall a few seconds before it exploded in a hail of hurtling lead, she remembered a call she needed to return. She murmured her friend’s name, and a virtual screen opened with the corresponding face.


“Sigsimund, bit busy to talk now, but can you meet me in the Temple of Tea in about five…” a laser beam from a circling drone sliced off the pavement she was standing on, while three robot samurai rose to bar her passage, katanas drawn (many humans were eager and enthusiastic to have a go at being evil masterminds, but few would settle for being minions). “…in about ten minutes? Lovely, see ya there!”


[image error]


The full version of this story is here: https://www.lesswrong.com/posts/sMsvcdxbK2Xqx8EHr/just-another-day-in-utopia 


A longer one set in a similar universe is here: https://www.lesswrong.com/posts/Ybp6Wg6yy9DWRcBiR/the-adventure-a-new-utopia-story

 •  0 comments  •  flag
Share on Twitter
Published on August 07, 2018 10:03

July 4, 2018

Time for Europe to step up

[image error]It is widely known that investment in artificial intelligence (AI) is concentrated in two Pacific Rim countries, China and the USA. But the strength of this duopoly is not generally appreciated, so here are two recent data points which throw it into sharp relief.


First, we have learned that Amazon’s R&D spend has reached 50% of the total spent by the UK on R&D. This means all the spend on any kind of R&D by the UK government and all the UK’s companies and universities. This astonishing fact becomes even more significant when you consider that a great deal of Amazon’s R&D spend is on AI, whereas not much of the UK’s R&D spend is on AI.


The second data point is a piece of research revealed by Jeffrey Ding, a Sinologist at the Future of Humanity Institute in Oxford, at the CogX conference in June. He started with the Chinese government’s widely-reported plan to invest $150bn in AI by 2020, and wondered if this included investments by China’s municipal and regional authorities. He discovered that it didn’t, and he managed to obtain data for much of that spend as well. It brings China’s total planned government investment in AI by 2020 to a breath-taking $429bn.


[image error]
Earlier at the same conference, Matt Hancock, who, as minister for digital in the UK’s Department for Culture Media and Sport (DCMS), is the nearest thing the UK has to a minister for AI, claimed that with a projected investment of £1bn, the UK was “in the front rank of AI countries.” This frankly preposterous claim was echoed by the Mayor of London, on the publication of a report which trumpeted London as the centre of AI within Europe, which is not unlike claiming to be the best champagne producer in Maidstone, Kent.


It is misleading to describe the development of AI as a race, partly because there is no fixed point when the process stops and one team is declared the winner, but mostly because the enormous benefits that will flow from the development of better and better AI will accrue to people all over the world. Just like the development of the smartphone did. However, there are at least three powerful arguments why Europe really should make more of a contribution to the global project of developing AI.


The first argument is that AI will deliver enormous benefits to the world, and the faster we reap those benefits, the better. To cite just two of many examples, AI will improve healthcare so that people who would otherwise suffer or die will remain healthy, and self-driving cars will stop the appalling holocaust of 1.2 million deaths each year (and 50 million maimings) which human drivers cause. Europe has great wealth, great universities, and millions of smart and energetic people; it can and should be contributing more to realising these benefits.


[image error]
The second argument is gloomier, but perhaps also more compelling. Europeans should not feel relaxed about the development of AI, humanitys’s most powerful technology, being so heavily concentrated elsewhere.


Jeffrey Ding argues that there is a far more lively debate in China about the government infringing on individual privacy than we in the West usually think. If so, this is great news, but it is hard to believe that China’s current approach to the development of AI would be acceptable in Europe. Most people here would be uncomfortable with schools using face recognition and other AI techniques to check whether the children are paying attention in class, and the way AI is being used to control the Uygur population in the western Chinese province of Xinjiang would also raise serious objections.


Many Europeans are also feeling slightly nervous about the great AI power to their west. So far, the development of AI in the USA has been a project for the private sector, but the government is showing signs of waking up to its importance, particularly with regard to military applications. The USA is currently a vibrant democracy, and has long been an invaluable ally. But things can change. President Trump’s ruminations about NATO and his willingness to initiate a trade war against the EU mean that Europe cannot be certain that America will always share the benefits of its AI prowess.


[image error]
The third reason is that AI might well be the source of much of the value in the economy in two or three decades’ time. Countries and regions which play only a minor role in developing the technology are likely to find themselves enfeebled.


To be clear, these are not arguments for autarky or self-reliant isolationism. We will all do better if the countries of the world collaborate to develop AI together, and share its benefits openly. That is the approach which Europe should champion. But sometimes, while planning for the best, it is wise to have a backup plan for the worst.


Jurgen Schmidhuber, one of the foundation figures of modern AI, argues that AI is currently dominated by the Pacific Rim countries for two main reasons: they both have huge single markets, and they both pursued muscular industrial strategies to promote the development of their technology industries. (Silicon Valley got started as a tech hub because of the sinking of the Titanic. To prevent a recurrence of that tragedy, the authorities decided that all ships must have powerful ship-to-shore radios, and it so happened that Silicon Valley was the home of a nascent radio industry. Later, the military research organisation DARPA funded cutting-edge tech research there, especially after America’s Sputnik moment – which was, of course, Sputnik.)


[image error]
Schmidhuber urges that Europe should strengthen its single market (woops, Brexit – yet another reason for the People’s Vote), and that it should enact similarly forward-thinking industrial policies. He also observes that while the Pacific Rim countries clearly dominate the internet of humans, the internet of things (IoT) is still up for grabs. He argues that Europe is home to the leading companies in the development and manufacture of many of the component parts of the IoT, so the game is still wide open.


It is time for Europe to step up.


[image error]

 •  0 comments  •  flag
Share on Twitter
Published on July 04, 2018 07:07

June 10, 2018

Sunny side up

[image error]

Satisfying stories feature a hero or heroine facing jeopardy and triumphing over adversity. This explains why most science fiction is dystopian: that’s where the jeopardy is.


This gives us a problem. Science fiction provides the metaphors we use to think about and discuss the future, and unfortunately, for every Star Trek there are multiple Star Wars and Terminators. Fear of the future stops many of us from thinking about it seriously. Maybe we should offset the likes of Black Mirror with some White Mirror.


So here is a description of the world in which AI has turned everything upside down – for the good. It’s a scenario, not a forecast, but maybe if we’re smart we can get there.


2025: panic averted
[image error]

Vehicles without human drivers are becoming a common sight in cities and towns all over the world. Professional drivers are starting to be laid off, and it is clear to everyone that most of them will be redundant within a few years. At the same time, employment in call centres and the retail industry is hollowing out as increasingly sophisticated digital assistants are able to handle customer enquiries, and the move to online shopping accelerates. The picking function in warehouses has been cost-effectively automated, and we are starting to see factories which normally have no lights on because no humans are working there.


Farmers are experimenting with robots for both crops and animal husbandry. Small wheeled devices patrol rows of vegetables, interrogating plants which don’t appear to be healthy specimens, and eliminating individual weeds with herbicides, or targeted jets of very hot water. Cattle are entirely content to be milked by robots, so fewer members of the declining population of farm workers still have to get up before daybreak every day.


Construction firms are experimenting with pre-fabricated units, but most construction projects remain subject to great variability of on-site conditions. Robots which can handle this unpredictability are still too expensive to replace human construction workers.


The tedious jobs which traditionally provided training wheels for accountants and lawyers (“ticking and bashing” for auditors and “disclosure” or “discovery” for lawyers) are increasingly being handled by machines. Skeptics about technological unemployment point out that the amount of work carried out by professional firms has actually increased, as whole categories of previously uneconomic jobs have become possible, and that professionals are kept busy because the machines still need training on each new data set. But fewer trainees are being hired, and thoughtful practitioners are asking where tomorrow’s qualified lawyers and accountants will come from.


Many companies have laid off some workers, but most have reduced their headcount primarily by natural wastage: not replacing people who retired or moved on. As a result, there have been fewer headlines about massive redundancies than some people feared, but at the same time it has become much harder for people to find new jobs. The unemployment rate among new graduates is at historically high levels.


But instead of panicking, the populations of most countries are reassured by the consensus which evolved when governments and philanthropists started to sponsor serious work on the issue of technological unemployment in the late 2010s. Slowly at first, and then rapidly, it became conventional wisdom that the Star Trek economy is achievable.


Meanwhile, governments are investing heavily in retraining to help workers cope with rapid job churn. AI personal tutor systems show promise, but are still rudimentary.


2035: Transition

[image error]


Large numbers of people are now unemployed, and welfare systems are swollen. Universal Basic Income was not the solution: there was no point paying generous welfare to the many people who remained in lucrative employment, and a basic income was insufficient for the rest. The new welfare has many incarnations, and interesting new experiments are springing up all the time. Concepts like PCI (Progressive Comfortable Income), and HELP (Human Elective Leisure Programme) are being discussed not only in the think tanks, but in kitchens, bars and restaurants all over the world.


Professional drivers are now extremely rare. Their disappearance was resisted for a while – road rage against the machines was ferocious in places. Autonomous vehicles were attacked, their cameras and LIDARs sprayed with paint. Some high-profile arrests and jail sentences quickly put a stop to the practice. Most jurisdictions now have roads which are off-limits to human drivers. Insurance premiums have plummeted, and fears about self-driving cars being routinely hacked have not been realised.


Deliveries of fast food and small parcels in major cities are now mostly carried out by autonomous drones, operating within their own designated level of airspace. Sometimes the last mile of a delivery is carried out by autonomous wheeled containers. For a while, teenagers delighted in “bot-tipping”, but with all the cameras and other sensory equipment protecting the bots, the risk of detection and punishment became too high.


In manufacturing, 3D printing has advanced less quickly than many expected, as it remained more expensive than mass production. But it is common in niche applications, like urgently required parts, components with complex designs, and situations where products are bespoke, as in parts of the construction industry. They have an impact on businesses and the economy far greater than their modest output level would suggest.


On construction sites, human supervision is still the norm for laying foundations, but pre-fabricated (often 3D-printed) walls, roofs and whole building units are becoming common. Robot labour and humans in exoskeletons are increasingly used to assemble them. Drones populate the air above construction sites, tracking progress and enabling real-time adjustments to plans and activities.


The Internet of Things has materialised, with everyone receiving messages continuously from thousands of sensors and devices implanted in vehicles, roads, trees, buildings, etc. Fortunately, the messages are intermediated by personal digital assistants, which have acquired the generic name of “Friends”, but whose owners often endow them with pet names. New types of relationship and etiquette are evolving to govern how people interact with their own and other peoples’ Friends, and what personalities the Friends should present.


There is lively debate about the best ways to communicate with Friends and other computers. One promising technology is tattoos worn on the face and around the throat which have micro-sensors to detect and interpret the tiny movements when people sub-vocalise, i.e., speak without actually making a noise. The tattoos are usually invisible, but some people have visible ones which give them a cyborg appearance. Brain-Computer Interfaces (BCI) have made less progress than their early enthusiasts expected.


A growing amount of entertainment and personal interaction is mediated through virtual reality. Good immersive VR equipment is now found in most homes, and it is increasingly rare to see an adolescent in public outside school hours. All major movies made by Netflix, Hollywood and Bollywood are now produced in VR, along with all major video games. To general surprise, levels of literacy – and indeed book sales – have not fallen. Many people now have more time and energy for reading. In a number of genre categories, especially romance and crime, the most popular books are written by AIs.


Major sporting competitions have three strands: robots, augmented humans, and un-augmented humans. Audiences for the latter category are dwindling.


Dating sites have become surprisingly effective. They analyse videos of their users, and allocate them to “types” in order to match them better. They also require their members to provide clothing samples from which they extract data about their smells and their pheromones. The discovery that relationship outcomes can be predicted with surprising accuracy with these kinds of data has slashed divorce rates.


Opposition to the smartphone medical revolution has subsided in most countries, and most people obtain diagnoses and routine health check-ups from their “Friends” several times a week. Automated nurses are becoming increasingly popular, especially in elder care.


Several powerful genetic manipulation technologies are now proved beyond reasonable doubt to be effective, but backed by public unease, regulators continue to hold up their deployment. Cognitive enhancement pharmaceuticals are available in some countries under highly regulated circumstances, but are proving less effective than expected. There are persistent rumours that they are deliberately being engineered that way.


Ageing is increasingly seen as an enemy which can be defeated.


Education is finally undergoing its digital revolution. Customised learning plans based on continuous data analysis and personal AI tutors are becoming the norm. Teachers are becoming coaches and mentors rather than instructors.


2045: The Star Trek Economy


[image error]


Artificial intelligence has made companies so efficient that the cost of most non-luxury goods and services is close to zero. Few people pay more than a token amount for entertainment or information services, which means that education and world-class healthcare are also much improved in quality and universally available. The cost of energy is dramatically reduced also, as solar power can now be harvested, stored and transmitted almost for free. Transportation involves almost no human labour, so with energy costs low, people can travel pretty much wherever they want, whenever they want. The impressive environments available in virtual reality do a great deal to offset the huge demand for travel that this might otherwise have created.


Food production is almost entirely automated, and the use of land for agriculture is astonishingly efficient. Vertical farms play an important role in high-density areas, and wastage is hugely reduced. The quality of the housing stock, appliances and furniture is being continuously upgraded. A good standard of accommodation is guaranteed to all citizens in most developed countries, although of course there are always complaints about the time it takes to arrive. Personalised or more luxurious versions are available at very reasonable prices to those still earning extra income. Almost no-one in developed countries now lives in cramped, damp, squalid or noisy conditions. Elsewhere in the world, conditions are catching up fast.


Other physical goods like clothes, jewellery and other personal accessories, equipment for hobbies and sports, and a bewildering array of electronic equipment are all available at astonishingly low cost. The Star Trek economy is almost mature. But access to goods and some services is still rationed by price. Nobody – in the developed world at least – wants for the necessities of civilised life, but nobody who is not employed can afford absolutely everything they might wish for. It is generally accepted that this is actually a good thing, as it means the market remains the mechanism for determining what goods are produced, and when.


Unemployment has passed 75% in most developed countries. Among those still working, nobody hates their job: people only do work that they enjoy. Everyone else receives an income from the state, and there is no stigma attached to being unemployed, or partially employed. In most countries the citizens’ is funded by taxes levied on the minority of wealthy people who own most of the productive capital in the economy, and in particular on those who own the AI infrastructure. The income is sufficient to afford a very high standard of living, with access to almost all digital goods being free, and most physical goods being extremely inexpensive.


In many countries, some of the wealthy people have agreed to transfer the means of production and exchange into communally owned, decentralised networks using blockchain technology. Those who do this enjoy the sort of celebrity and popularity previously reserved for film and sports stars.


Some countries mandated these transfers early on by effectively nationalising the assets within their legislative reach, but found that their economies were stagnating, as many of their most innovative and energetic people emigrated. Worldwide, the idea is gaining ground that private ownership of key productive assets is distasteful. Most people do not see it as morally wrong, and don’t want it to be made illegal, but it is often likened to smoking in the presence of non-smokers. This applies particularly to the ownership of facilities which manufacture basic human needs, like food and clothing, and to the ownership of organisations which develop and deploy the most essential technology – the technology which adds most of the value in every industry sector: artificial intelligence.


The gap in income and wealth between rich and poor countries has closed dramatically. This did not happen because of a transfer of assets from the West to the rest, but thanks instead to the adoption of effective economic policies, the eradication of corruption, and the benign impact of technology in the poorer countries.


Another concern which has been allayed is that life without work would deprive the majority of people of a sense of meaning in their lives. Just as amateur artists were always happy to paint despite knowing that they could never equal the output of a Vermeer, so people now are happy to play sport, write books, give lectures and design buildings even though they know that an AI could do any of those things better than them.


Not everyone is at ease in this brave new world, however. Around 10% of the population in most countries suffers from a profound sense of frustration and loss, and either succumbs to drugs or indulges almost permanently in escapist VR entertainment. A wide range of experiments is under way around the world, finding ways to help these people join their friends and families in less destructive or limiting lifestyles. Huge numbers of people outside that 10% have occasional recourse to therapy services when they feel their lives becoming slightly aimless.


Governments and voters in a few countries resisted the economic singularity, seeing it as a de-humanising surrender to machine rule. Although they found economically viable alternatives at first, their citizens’ standard of living quickly fell far behind. Most of these governments have now collapsed like the communist regimes of Eastern Europe in the early 1990s, and there are persistent rumours that President-for-life Putin met a very grisly end in 2036. The other hold-outs look set to follow – hopefully without violence.


Significant funds are now allocated to radical age extension research, and there is talk of “longevity escape velocity” being within reach – the point when each year, science adds a year to your life expectancy. Most forms of disability are now offset by implants and exoskeletons, and cognitive enhancements through pharmaceuticals and brain-computer interface techniques are showing considerable promise.


The education sector has ballooned, and is vacational rather than vocational. Most education is provided by AIs.


Safeguards have now been found to enable direct democracy to be implemented in many areas. Professional politicians are now rare.


In London, DeepMind announces that it expects to unveil the first artificial general intelligence. With bated breath, the world awaits the arrival of Earth’s first superintelligence.



[image error]
 •  0 comments  •  flag
Share on Twitter
Published on June 10, 2018 07:58

May 20, 2018

New book: Artificial Intelligence and the Two Singularities

[image error]


My latest book has just been published by CRC Press, an imprint of the academic publishing house Taylor and Francis.  It updates and expands on my previous non-fiction AI books, especially the sections on the economic singularity.


Having been published in the past by Random House, and having also self-published, it is exciting to complete the trio of publishing options with an academic publisher, and it is an honour to be picked up by CRC.  (This does have an impact on pricing beyond my control.)


The book has attracted some gratifying reviews, including these:


“The arrival of super-intelligent AI and the economic replacement of humans by machines on a global scale are among the greatest challenges we face. This book is an excellent introduction to both. It is thoroughly researched and persuasive. Chace’s principal argument seems to be correct: we need to prepare now for the economic singularity or face a serious disruption of our civilization.”

Stuart Russell, Professor of Computer Science and Smith-Zadeh Professor in Engineering, University of California, Berkeley


“A brilliantly lucid guide to the current state of artificial intelligence and its possibilities and potential – including the likelihood that it will eliminate employment as we currently know it. What makes this a must-read is what comes next – this is no dystopian alarm but a clarion call to start thinking about how to cope with this eventuality – and an intriguing guide to some possible destinations. We should all be thinking about these potential problems right now – thankfully Calum has already provided an entertaining and thoughtful roadmap.”

Mark Mardell, Presenter, BBC Radio Four

 •  0 comments  •  flag
Share on Twitter
Published on May 20, 2018 03:28

May 3, 2018

AI: it’s not a race

Giant cogs are starting to turn


[image error]
Western governments are finally waking up to the significance of artificial intelligence. It’s a slow process, with much further to go, and most politicians have yet to grasp the significance of AI, but the gradual setting into motion of giant cogs is to be welcomed.


In the November 2017 budget the UK’s Chancellor excitedly announced expenditure of £75m on AI. In March 2018, France’s President Macron put that to shame by announcing a spend of €1.5bn by 2020, primarily to halt the brain drain of French computer scientists. The following month Germany’s Chancellor Merkel declared that Germany would be at the forefront of AI development, although she refrained from quoting any numbers. Shortly afterwards the UK announced a substantial upward revision of its investment, to £1bn, most of it coming from the private sector.


These are large sums of money, but they are modest compared to what the Chinese government is spending. The announcement in January of a $2bn AI industrial park in a mountainous province in the west of the country is just one example of the government’s ambitions. Alibaba’s announcement of $15bn to be invested in AI over three years gives a better idea of the scale of China’s activity.


China’s declared aim is to overtake the US as the predominant provider of AI development by 2030, and it has made astonishing progress in the short time since its “Sputnik” moment in March 2016, when DeepMind’s AlphaGo beat the world’s best human Go player. It now attracts more VC investment into AI firms than the US, for instance.


The US presents a curious dichotomy. Its tech giants still dominate the AI landscape: Google, Facebook, Amazon, Microsoft and Apple are increasingly AI companies, and they are the largest companies in the world (followed immediately by Alibaba and Tencent). Astonishingly, Amazon spends on R&D half of what the whole of the UK spends – including all UK government departments, all UK companies, and all the UK’s universities and NGOs. But the US government seems blithely unaware of the growing importance of AI. When Facebook founder Mark Zuckerberg was grilled by the US Senate in April, the senators revealed how little understanding they had of what kind of company it is. And the US president is too busy reassuring unemployed coal miners that he will bring their jobs back to have noticed the likelihood of massive waves of job disruption from machines.


The “Three Wise Men” moment


[image error]
Still, the big picture is that politicians are slowly waking up to the importance of AI. It is reminiscent of the “three wise men” moment in 2014, when Stephen Hawking, Elon Musk and Bill Gates all declared that strong AI is probably coming, and will either be the best or the worst thing ever to happen to humanity. The initial result was pictures of the Terminator all over the newspapers, but fairly soon a more balanced discussion emerged.


If the giant cogs of the government machines are indeed finally beginning to turn, that is to be welcomed. We need politicians and policy makers to be fully engaged if we are to grasp the amazing opportunities that AI offers – and to meet the serious challenges it poses. We need governments to help train AI researchers and engineers, to engage with the tech giants on issues such as privacy and bias, and to regulate them if necessary. Above all, we need governments to prepare for the possibility of technological unemployment.


These activities are not zero-sum between countries. Humanity will get the best out of AI only if it collaborates across borders. That is the natural inclination of scientists and business people, but not necessarily of voters and governments. And so the nascent discussion about AI and government is being framed as a race. Does that actually make any sense?


It’s not a race…


[image error]
The most profoundly influential technology of the last decade was arguably the smartphone – itself powered increasingly by AI. The race to sell them was won by Apple, which earns almost all of the profit in the smartphone industry. But that does not mean that only Americans benefit from smartphones: they have become essential tools for fully half the world’s population – they are constant companions for people in all countries and at all income levels. The same will be true of new forms of AI: self-driving cars and digital assistants, for instance. As soon as they are ready for prime time they will spread around the planet like the rising sun.


Of course the governments and citizens of the countries which produce the most financially successful AI companies will benefit the most from their tax payments, and their workforces will benefit most from the ecosystem of jobs and skills they build. But building companies (and universities) is not a fixed-term process where a winner surges through the tape to claim the prize and everyone else loses. Assuming we meet and overcome the challenges that AI poses, when country A produces more and better AI, the people in countries B, C and D are beneficiaries too.


…but there is a duopoly


[image error]
The idea of leadership in AI is less incoherent than the idea of an AI race. The problem here is the bizarre failure by politicians to recognise the enormous difference in scale between the AI activities of different countries. Leaders in the UK, France and Germany all declare their countries to be leaders in AI, whereas the truth is they are bit players compared to the duopoly of China and the US. The report published by the UK’s House of Lords in April was a refreshing exception: “vague statements about the UK “leading” in AI are unrealistic and unhelpful, especially given the vast scale of investment in AI by both the US and China.”


The lords went on to propose a more nuanced role for the UK. Citing its strong tradition of computer science research (Alan Turing, among many others), successful startups (notably DeepMind), globally respected institutions like the BBC and the British legal system, they suggested the UK could help “convene, guide and shape the international debates which need to happen.”


Even setting aside the question of how much harm Brexit is doing and will do to Britain’s standing in the world, is this noble vision realistic? Perhaps it is – but probably only so long as there are no serious and lasting differences of opinion between the world’s regions about how AI should evolve. If China cleaves to its view that the privacy of individuals is of little account compared to the safety of society at large – and the power of the communist party – then why should they listen when self-appointed leadership organisations in the UK and elsewhere tell them otherwise? Certainly, they will have to obey the rules of the countries they sell their products and services in, but consumers in those countries will demand the latest and the best technology, wherever it originates, and they may not look kindly on governments that get in the way.


Senior figures within the EU are now talking about investing €20bn into AI, but there is no realistic prospect yet of that happening. Absent such a massive change, it is hard to see why effective leadership in ethical frameworks will reside anywhere other than the same place as the leadership in R&D and commercial implementation. And that is in the AI duopoly of the US and China.

 •  0 comments  •  flag
Share on Twitter
Published on May 03, 2018 03:48

April 25, 2018

Aggravating algorithms

[image error]


You’ll have noticed there is something of a backlash against the tech giants these days. In the wake of the scandal over Cambridge Analytica’s alleged unauthorised use of personal data about millions of Facebook users, Mark Zuckerberg was subjected to an intensive grilling by the US Senate. (In the event, the questions were about as lacerating as candy floss because it turns out that most Senators have a pretty modest level of understanding about how social media works.) More seriously, perhaps, the share prices of the tech giants have tumbled in recent weeks, although Mr Zuckerberg’s day in Congress raised Facebook’s sufficiently to increase its CEO’s personal wealth by around $3bn. Not a bad day’s pay.


The tech firms are in the dog house for a number of reasons, but one of the most pressing is the widespread perception that they are using – and enabling others to use – algorithms which are poorly understood, and causing harm.


[image error]


The word “algorithm” comes from the name of a ninth-century Persian mathematician, Al-Khwarizmi, and is surprisingly hard to explain. It means a set of rules or instructions for a computer to follow, but not the precise, step-by-step instructions which a computer programme contains. A machine learning algorithm uses an initial data set to build an internal model which it uses to make predictions; it tests these predictions against additional data and uses the results to refine the model.


If the initial data set is not representative of the population (of people, for instance) which it will offer decisions about, then those decisions can be prejudiced and harmful. When asked to provide pictures of hands, algorithms trained on partial data sets have returned only pictures of white hands. In 2015, a Google algorithm reviewing photographs labelled pictures of black people as gorillas.


When this kind of error is made when dealing with decisions about who should get a loan, or be sent to jail, the consequences can obviously be serious. It is not enough to say (although it is true) that the humans being replaced by these systems are often woefully prejudiced. We have to do better.


[image error]


Algorithms’ answers to our questions are only as good as the data we give them to work with. To find the best needles you need really big haystacks. And not just big: you need them to be diverse and representative too. Machine language researchers are very well aware of the danger of GIGO – garbage in, garbage out – and all sorts of efforts and initiatives are under way in the tech giants and elsewhere to create bigger and better datasets.


[image error]
Society has to walk a tightrope regarding the use of data. Machine learning and other AI techniques already provide great products and services: Google Search provides us with something like omniscience, and intelligent maps tell you how long your journey will be at different times of day, and divert you if an accident causes a blockage. In the future they will provide many more wonders, like the self-driving cars which will finally end the holocaust taking place continuously on our roads, killing 1.2 million people each year and maiming another 50 million or so.


The data to fuel these marvels is generated by the fast-growing number of sensors we place everywhere, and by the fact that more and more of our lives are digital, and we leave digital breadcrumb trails everywhere we go. It would be a tragedy if the occasionally hysterical backlash against the tech giants we are seeing today ended up throttling the availability of data which is needed to let ML systems weave their magic – and do so with less bias than we humans harbour.


Left to their own devices, it is likely that most of us would carry on using Facebook and similar services with the same blithe disregard for our own privacy as they always have done. But the decisions about how our data is used are not ours alone to make. Quite rightly, governments and regulators will make their opinions known. The European Union’s GDPR (General Data Protection Regulation), which comes into force in May, is a powerful example. It requires everyone who stores or processes the personal data (liberally defined) of any EU citizens to do so only when necessary, and to provide the subjects of that data with access to it on demand.


This is all well and good, but as we call for “something to be done” to curb the power of the tech giants, let’s bear a few things in mind. First, regulation is a blunt weapon: regulators, like generals, often fight the war that has just finished, and as technology accelerates, this effect will become more pronounced. Second, regulation frequently benefits incumbents by raising barriers to entry, and by doing so curbs innovation. In general, regulation should be used sparingly, and to combat harms which are either proven, or virtually inevitable.


[image error]


Third, the issue is who controls or has access to the data, not who owns it. “Ownership” of a non-rivalrous, weightless good like data is a nebulous idea. But access and control are important. More regulation generally means that the state gains more access and control over our data, and we should think carefully before we rush in that direction. China’s terrifying Social Credit system shows us where that road can lead: the state knows everything about you, and gives you a score according to how you behave on a range of metrics (including who your friends are and what they say on social media). That score determines whether you have access to a wide range of benefits and privileges – or whether you are punished.


As our data-fuelled future unfolds, there will be plenty more developments which oblige us to discuss all this. At the moment the data we generate mostly concerns places we go, people we talk to, and things we buy. In future it will get more and more personal as well as more and more detailed. Increasingly we’ll be generating – and looking to control the use of – data about our bodies and our minds. Then the debates about privacy, transparency and bias will get really interesting.

 •  0 comments  •  flag
Share on Twitter
Published on April 25, 2018 09:26