🔥 Rogue AI; blockchain & the poor; exponential climate change; synthetic life; thoughts on the future++ #55

Revue
 
How AI goes rogue and what we learn from it. What the blockchain will mean for the world's poor. Self
 
March 27 - Issue #55
The Exponential View
How AI goes rogue and what we learn from it. What the blockchain will mean for the world’s poor. Self-driving cars & segregated cities. The exponential rise in sea levels. 34 thoughts about the future. Apple shouldn’t give in to the FBI. And much more in our Easter special edition. 🐇 🌷 
Please recommend: Twitter (preferred) | Facebook (also good

Dept of the near future
😀 What the blockchain will mean for the world’s poor: better property rights, remittances, identity, trust & supply chains. EXCELLENT.
🚗 Will self-driving cars lead to cities whose roads and planning are even more attuned to the needs of vehicles rather than humans? THOUGHT PROVOKING: these algorithmic beasts will have better navigation, awareness and coordination capabilities than two-legged wetware.
⚙ What is a robot? “Self-operating machines are permeating every dimension of society, so that humans find themselves interacting more frequently with robots than ever before—often without even realizing it.”
🎁 Farhad Manjoo: Can the Uber model translate to other businesses?
💰How to build a billion dollar business. RATHER GOOD presentation by Sarah Tavel of Greylock Partners.
Dept of artificial intelligence
A Microsoft chatbot went rogue on Twitter and started spewing Nazi epithets. It is an extremely helpful case study in outlining some of the key issues around the application of machine learning and AI to everyday tasks. What isn’t interesting is not that Tay, the bot, was taught to say rude things. What is interesting is what this tells us about designing AI systems that operate in the real world and how their learning can be hacked.
AIs are going to need to learn and interact somewhere akin to the real world. Equally, if we allow AI-systems unexpurgated access to the ‘real world’ while they are learning, there could be ramifications. 
The first: we don’t have common ethical standards within societies or cultures nor across them. Philippa Foot’s trolley experiment never results in unanimity within cultures. (Click here if you don’t know the Trolley problem.)
💡 However, this recent paper looks at cultural differences and variations of the Trolley problem. It finds that ordinary British people will pull a switch between 63 and 91% of the time but Chinese people would do so between 33% and 71% of the time. SUPER ACCESSIBLE, worth reading. Should future systems follow British or Chinese ethical standards? And who decides?
The Chalmers-Bourguet survey of Anglo-centric philosophers revealed that philosophers would switch 68% of the time, suggesting they might reflect their own politico-cultural heritage.
The second: we already have similarly poorly designed optimisation systems in the world. And they affect hundreds of millions of people, only they are less transparent and the people operating them are less responsive than Microsoft has been with Tay. 
Examples including credit scoring algorithms or indeed complaints-handling protocols amongst large firms. Those algorithms are often simple regressions built-in small datasets. Or in the case of complaints protocols often very simple, inflexible decision trees applied by a human. Sure Tay is more advanced, nuanced and embarrassing but no real harm is done compared to a poorly designed decision tree which might determine your access to a financial product or insurance. 
How many algorithmic approaches make it out into the real world without adequate testing or understanding of their ramifications or worse, their unintended consequences?
The third: Microsoft should have known better. In the experimental crucible that is the Internet, users will push and pull an experiment to its limits. Experience dictated that we, the collective, could have known this. ☠️ Vice asks some bot experts how to design a bot that doesn’t go 'haywire’. 
I’m extremely surprised Microsoft didn’t do what DeepMind did with AlphaGo. In that case, DeepMind seemed to keep AlphaGo under wraps until they were reasonably clear that AlphaGo was going to do an amazing job in the field. Microsoft appears to have taken the opposite tack. Strikes me as weird. You can read their statement here
The fourth: It’s not really clear how much harm was done by Tay. It’s embarrassing for Microsoft on a corporate level but it’s really not clear that Tay influenced anyone to become more xenophobic or racist. But we have all had a great lesson in some of the issues around training algorithmic systems. (Remember Google and the gorillas?) 
The fifth: Tay appears to support the notion of the 'unreasonable effectiveness of data’. That large amounts of training data determine the effectiveness of an AI-based system more so than the smarts of the algorithms.
Dept of climate change
No much good news here, but at least better informed means better decisions. 
🔥🔥 We are approaching exponential rates of glacial melt and sea-level rises. EXCELLENT, and alarming review of a new paper by James Hansen, the NASA scientist, on the non-linear acceleration of climate degradation and a possible imminent Heinrich event.
🔥 What we are doing to the environment has no parallel in the last 66m years. Our models are totally in uncharted territory. 
And now, at last, 96% of US meteorologists accepted climate change is real.
The Rockefeller family divests from Exxon Mobil. Their wealth stemmed from Standard Oil, a precursor to Exxon. Scotland closed its last coal-powered station 115 years after it began use the fossil-fuel for electrical generation. Oregon becomes the latest US state to ban coal suggesting that it is game-over for the commodity.
Short morsels to appear smart at dinner parties
The relationship between maths and physics and modernist art. Lovely brief essay.
🍾 Everything we’ve learnt about platform business models we learnt from mediaeval France and the Duke of Champagne.
😧 Near half of all mobile revenues comes from just 0.19% of users/losers. Invokes some ethical questions: is that industry taken advantage of a disadvantaged micro-minority? (Those with lower will power/high addictive potentiality?)
Anti-vaxxers have successfully brought measles and whooping cough back to the US.
Can you believe video? Extremely realistic facial reenactment tech lets you clone your facial expression onto another person. 
What EV members are up to
I am speaking on Artifical Intelligence at the awesome Business of Software Europe conference in May. This is run by Mark Littlewood & there is a special ‘EV’ discount.
Tom Glocer, angel investor, former attorney & former CEO of Thomson Reuters: Why Apple should not be forced to open a back-door to iPhones.
Pete Flint, founder of Trulia: “34 Remarkable things about the future
John Havens has composed a song about the philosophical ramifications of AI. We’d love to see it set to music (and sung.)
End note
I have a few days with my family over Easter so will be very slow on email. If you have holidays in your part of the world, I hope you enjoy them too!
P.S. Still hiring Product Managers for my marketplaces team based in London. Send qualified referrals my way.
Did you enjoy this issue?
Thumbs up 1ae5a7bdfcd3220e2b376aa0c1607bc5edaba758e5dd83b482d03965219a220b Thumbs down e13779fa29e2935b47488fb8f82977fedcf689a0cc0cc3c19fa3c6bb14d1493b
Carefully curated by The Exponential View with Revue.
If you were forwarded this newsletter and you like it, you can subscribe here.
If you don't want these updates anymore, please unsubscribe here.