University of Oxford
Oli Scarff/Getty Images
Oxford and Cambridge, the oldest universities in Britain and two of the oldest in the world, are keeping a watchful eye on the buzzy field of artificial intelligence (AI), which has been hailed as a technology that will bring about a new industrial revolution and change the world as we know it.
Over the last few years, each of the centuries-old institutions have pumped millions of pounds into researching the possible risks associated with machines of the future.
Clever algorithms can already outperform humans at certain tasks. For example, they can beat the best human players in the world at incredibly complex games like chess and Go, and they're able to spot cancerous tumors in a mammogram far quicker than a human clinician can. Machines can also tell the difference between a cat and a dog, or determine a random person's identity just by looking at a photo of their face. They can also translate languages, drive cars, and keep your home at the right temperature. But generally speaking, they're still nowhere near as smart as the average 7-year-old.
The main issue is that AI can't multitask. For example, a game-playing AI can't yet paint a picture. In other words, AI today is very "narrow" in its intelligence. However, computer scientists at the the likes of Google and Facebook are aiming to make AI more "general" in the years ahead, and that's got some big thinkers deeply concerned.
Nick Bostrom, a 47-year-old Swedish born philosopher and polymath, founded the Future of Humanity Institute (FHI) at the University of Oxford in 2005 to assess how dangerous AI and other potential threats might be to the human species.
In the main foyer of the institute, complex equations beyond most people's comprehension are scribbled on whiteboards next to words like "AI safety" and "AI governance." Pensive students from other departments pop in and out as they go about daily routines.
It's rare to get an interview with Bostrom, a transhumanist who believes that we can and should augment our bodies with technology to help eliminate ageing as a cause of death.
"I'm quite protective about research and thinking time so I'm kind of semi-allergic to scheduling too many meetings," he says.
Tall, skinny and clean shaven, Bostrom has riled some AI researchers with his openness to entertain the idea that one day in the not so distant future, machines will be the top dog on Earth. He doesn't go as far as to say when that day will be, but he thinks that it's potentially close enough for us to be worrying about it.
Swedish philosopher Nick Bostrom is a polymath and the author of "Superintelligence."
The Future of Humanity Institute
If and when machines possess human-level artificial general intelligence, Bostrom thinks they could quickly go on to make themselves even smarter and become superintelligent. At this point, it's anyone's guess what happens next.
The optimist says the superintelligent machines will free up humans from work and allow them to live in some sort of utopia where there's an abundance of everything they could ever desire. The pessimist says they'll decide humans are no longer necessary and wipe them all out.Billionare Elon Musk, who has a complex relationship with AI researchers, recommended Bostrom's book "Superintelligence" on Twitter.
Bostrom's institute has been backed with roughly $20 million since its inception. Around $14 million of that coming from the Open Philanthropy Project, a San Francisco-headquartered research and grant-making foundation. The rest of the money has come from the likes of Musk and the European Research Council.
Located in an unassuming building down a winding road off Oxford's main shopping street, the institute is full of mathematicians, computer scientists, physicians, neuroscientists, philosophers, engineers and political scientists.
Eccentric thinkers from all over the world come here to have conversations over cups of tea about what might lie ahead. "A lot of people have some kind of polymath and they are often interested in more than one field," says Bostrom.
The FHI team has scaled from four people to about 60 people over the years. "In a year, or a year and a half, we will be approaching 100 (people)," says Bostrom. The culture at the institute is a blend of academia, start-up and NGO, according to Bostrom, who says it results in an "interesting creative space of possibilities" where there is "a sense of mission and urgency."
If AI somehow became much more powerful, there are three main ways in which it could end up causing harm, according to Bostrom. They are:
"Each of these categories is a plausible place where things could go wrong," says Bostrom.
With regards to machines turning against humans, Bostrom says that if AI becomes really powerful then "there's a potential risk from the AI itself that it does something different than anybody intended that could then be detrimental."
In terms of humans doing bad things to other humans with AI, there's already a precedent there as humans have used other technological discoveries for the purpose of war or oppression. Just look at the atomic bombings of Hiroshima and Nagasaki, for example. Figuring out how to reduce the risk of this happening with AI is worthwhile, Bostrom says, adding that it's easier said than done.
I think there is now less need to emphasize primarily the downsides of AI.
Asked if he is more or less worried about the arrival of superintelligent machines than he was when his book was published in 2014, Bostrom says the timelines have contracted.
"I think progress has been faster than expected over the last six years with the whole deep learning revolution and everything," he says.
When Bostrom wrote the book, there weren't many people in the world seriously researching the potential dangers of AI. "Now there is this thriving small, but thriving field of AI safety work with a number of groups," he says.
While there's potential for things to go wrong, Bostrom says it's important to remember that there are exciting upsides to AI and he doesn't want to be viewed as the person predicting the end of the world.
"I think there is now less need to emphasize primarily the downsides of AI," he says, stressing that his views on AI are complex and multifaceted.
Bostrom says the aim of FHI is "to apply careful thinking to big picture questions for humanity." The institute is not just looking at the next year or the next 10 years, it's looking at everything in perpetuity.
"AI has been an interest since the beginning and for me, I mean, all the way back to the 90s," says Bostrom. "It is a big focus, you could say obsession almost."
The rise of technology is one of several plausible ways that could cause the "human condition" to change in Bostrom's view. AI is one of those technologies but there are groups at the FHI looking at biosecurity (viruses etc), molecular nanotechnology, surveillance tech, genetics, and biotech (human enhancement).
A scene from 'Ex Machina.'
Source: Universal Pictures | YouTube
When it comes to AI, the FHI has two groups; one does technical work on the AI alignment problem and the other looks at governance issuesthat will arise as machine intelligence becomes increasingly powerful.
The AI alignment group is developing algorithms and trying to figure out how to ensure complex intelligent systems behave as we intend them to behave. That involves aligning them with "human preferences," says Bostrom.
Roughly 66 miles away at the University of Cambridge, academics are also looking at threats to human existence, albeit through a slightly different lens.
Researchers at the Center for the Study of Existential Risk (CSER) are assessing biological weapons, pandemics, and, of course, AI.
We are dedicated to the study and mitigation of risks that could lead to human extinction or civilization collapse.
Centre for the Study of Existential Risk (CSER)
"One of the most active areas of activities has been on AI," said CSER co-founder Lord Martin Rees from his sizable quarters at Trinity College in an earlier interview.
Rees, a renowned cosmologist and astrophysicist who was the president of the prestigious Royal Society from 2005 to 2010, is retired so his CSER role is voluntary, but he remains highly involved.
It's important that any algorithm deciding the fate of human beings can be explained to human beings, according to Rees. "If you are put in prison or deprived of your credit by some algorithm then you are entitled to have an explanation so you can understand. Of course, that's the problem at the moment because the remarkable thing about these algorithms like AlphaGo (Google DeepMind's Go-playing algorithm) is that the creators of the program don't understand how it actually operates. This is a genuine dilemma and they're aware of this."
The idea for CSER was conceived in the summer of 2011 during a conversation in the back of a Copenhagen cab between Cambridge academic Huw Price and Skype co-founder Jaan Tallinn, whose donations account for 7-8% of the center's overall funding and equate to hundreds of thousands of pounds.
"I shared a taxi with a man who thought his chance of dying in an artificial intelligence-related accident was as high as that of heart disease or cancer," Price wrote of his taxi ride with Tallinn. "I'd never met anyone who regarded it as such a pressing cause for concern let alone anyone with their feet so firmly on the ground in the software business."
University of Cambridge
Geography Photos/UIG via Getty Images
CSER is studying how AI could be used in warfare, as well as analyzing some of the longer term concerns that people like Bostrom have written about. It is also looking at how AI can turbocharge climate science and agricultural food supply chains.
"We try to look at both the positives and negatives of the technology because our real aim is making the world more secure," says Sen higeartaigh, executive director at CSER and a former colleague of Bostrom's. higeartaigh, who holds a PhD in genomics from Trinity College Dublin, says CSER currently has three joint projects on the go with FHI.
External advisors include Bostrom and Musk, as well as other AI experts like Stuart Russell and DeepMind's Murray Shanahan. The late Stephen Hawking was also an advisor when he was alive.
The Leverhulme Center for the Future of Intelligence (CFI) was opened at Cambridge in 2016 and today it sits in the same building as CSER, a stone's throw from the punting boats on the River Cam. The building isn't the only thing the centers share staff overlap too and there's a lot of research that spans both departments.
Backed with over 10 million from the grant-making Leverhulme Foundation, the center is designed to support "innovative blue skies thinking," according to higeartaigh, its co-developer.
Was there really a need for another one of these research centers? higeartaigh thinks so. "It was becoming clear that there would be, as well as the technical opportunities and challenges, legal topics to explore, economic topics, social science topics," he says.
"How do we make sure that artificial intelligence benefits everyone in a global society? You look at issues like who's involved in the development process? Who is consulted? How does the governance work? How do we make sure that marginalized communities have a voice?"
The aim of CFI is to get computer scientists and machine-learning experts working hand in hand with people from policy, social science, risk and governance, ethics, culture, critical theory and so on. As a result, the center should be able to take a broad view of the range of opportunities and challenges that AI poses to societies.
"By bringing together people who think about these things from different angles, we're able to figure out what might be properly plausible scenarios that are worth trying to mitigate against," said higeartaigh.
Go here to see the original:
How Britain's oldest universities are trying to protect humanity from risky A.I. - CNBC
- Pioneering Transhumanism: a conversation with Natasha Vita-More - February 6th, 2023
- Transhumanist Party - Wikipedia - December 18th, 2022
- Transhumanism: advances in technology could already put evolution into ... - November 24th, 2022
- Bidens transhumanist EO calls for natural biology to be programmed ... - October 4th, 2022
- Nvidia and the 3D Future of the Internet - TechNewsWorld - October 4th, 2022
- Doomsday bunkers, Mars and 'The Mindset': the tech bros trying to outsmart the end of the world - The Conversation - October 4th, 2022
- Leo Hohmann: Biden's Executive Order Designed to Release Transhumanist ... - September 16th, 2022
- Prebunking Disinformation | prebunking disinformation - Patheos - September 8th, 2022
- WEF agenda envisions an augmented society ruled by Internet of Bodies, digital ID - The Sociable - September 8th, 2022
- Transhumanist Bioweapon & Deactivating the Bluetooth Chips - Hope ... - August 22nd, 2022
- History of a hard man: Neil Balme memoir stands out from the pack - Sydney Morning Herald - August 22nd, 2022
- Economism vs Common Good Part Three | Demythicizing Economism - Patheos - August 22nd, 2022
- The Importance of Medical Ethics Highlighted During the 75Th Anniversary of the Nuremberg Code - The Epoch Times - August 22nd, 2022
- Fringe: TV Has Never Been The Same Without This Sci-Fi Show - Fortress of Solitude - August 22nd, 2022
- Why death matters - Big Think - August 14th, 2022
- Horror stories of cryonics: The gruesome fates of frozen bodies - Big Think - August 5th, 2022
- What Kathleen Stock gets wrong about the Tories, trans and feminism - CapX - July 27th, 2022
- Tyranny by Numbers | John Waters - First Things - July 27th, 2022
- There Is No Such Thing As A Lightning Wallet - Bitcoin Magazine - July 27th, 2022
- What Is A Human? - The American Conservative - July 19th, 2022
- Exposing The Agenda To Implement Global Totalitarianism Based On ... - July 3rd, 2022
- There exists an Agenda to implement Global Totalitarianism based on ... - July 3rd, 2022
- What is Truth? | What is truth? - Patheos - July 3rd, 2022
- Dr. Zelenko Crimes Against Humanity and the Transhumanist Agenda - Ben Stein ... - April 19th, 2022
- Transhumanism: The Plot to Control Your Life! - Jonathan Brentner - April 19th, 2022
- Facebook and radicalisation: how can we regulate the internet to prevent harm? - Stuff - April 6th, 2022
- The Power of Pain | Peter J. Leithart - First Things - April 6th, 2022
- Values of the U.S. Transhumanist Party U.S ... - December 24th, 2021
- U.S. Transhumanist Party Official Website U.S ... - December 24th, 2021
- Another Year Without A Family Christmas? - newsconcerns - December 24th, 2021
- Will We Ever Cheat Death and Become Immortal With Mind Uploading? - Interesting Engineering - December 10th, 2021
- Pig Kidney Successfully Transplanted From Hog to Human - November 22nd, 2021
- As Facebook evolves to Meta, what is the future of consciousness and control? - Baptist News Global - November 8th, 2021
- International Symposium on Sorgner's "We Have Always Been Cyborgs"Back to Events - Institute for Ethics and Emerging Technologies - November 8th, 2021
- Crypto leaders are obsessed with life extension. Here's why Cointelegraph Magazine - Cointelegraph - August 18th, 2021
- Inside the Wild Ride of "Titane," the French Body-Horror Film That Just Won the Palme d'Or - InsideHook - July 21st, 2021
- We Can't Cheat Aging and Death, Claims New Study - Reason - June 23rd, 2021
- Third parties in the U.S. What options do voters have? - Deseret News - June 6th, 2021
- Building the Ark - Architecture - E-Flux - June 6th, 2021
- A Closer Look at the AI Hype Machine: Who Really Benefits? - Common Dreams - February 5th, 2021
- Deadpool's Monster Army and the X-Men's Nation Share a Surprising Tactic - Screen Rant - February 5th, 2021
- Moral advice straight from the computer: is it time for a virtual Socrates? - Innovation Origins - September 5th, 2020
- Masks Against the Coronavirus: How the Rejection of Mask Use, Unites the Extreme Right and the Extreme Left The Costa Rica News - The Costa Rica News - September 5th, 2020
- The Honorable Dr. Dale Layman, Founder of Robowatch, LLC, is Recognized as the 2020 Humanitarian of the Year by Top 100 Registry, Inc. - IT News... - September 4th, 2020
- Frontrunner for the VA GOP's 2021 Gubernatorial Nomination Rallies in Honor of Far-Right Paramilitary Group Member; As Del. Jay Jones Points Out, the... - July 7th, 2020
- Why Humanize? A New Effort to Defend the Unique Dignity of Human Beings - Discovery Institute - June 3rd, 2020
- The Proto-Communist Plan to Resurrect Everyone Who Ever Lived - VICE - April 22nd, 2020
- Is It Moral To Work For A Tech Giant? - Institute for Ethics and Emerging Technologies - April 19th, 2020
- First human to live 1,000 years may already be alive says scientist fighting to end ageing - Express.co.uk - March 1st, 2020
- 5 things to know about Miss Anthropocene by Grimes - Vancouver Sun - February 25th, 2020
- Information on all 50 NH Primary candidates (including the Hartford Whalers guy) - Manchester Ink Link - February 12th, 2020
- The tree family - Frontline - February 2nd, 2020
- 'Cyberpunk 2077' Delayed as CD Projekt Red Polishes 'Crowning Achievement' Over 'Witcher 3: Wild Hunt' - Newsweek - January 20th, 2020
- Transhumanism, AI, gaming and human biology to feature at Mumbrella MSIX with new session announced - mUmBRELLA* - November 5th, 2019
- Religious and spiritual online forums consist of chaotic, impactful ideas - Lamron - November 5th, 2019
- Publishers are going to live or die based on their relationship with readers: How Quartz is rethinking its membership offerings - Nieman Journalism... - October 19th, 2019
- Sports Man of the Future - The Good Men Project - October 14th, 2019
- Could a robot be prime minister? Machines will soon be smart enough to run the world, says futurist - CBC.ca - October 8th, 2019
- Heavy data protection regulation looms in Labour plans for post-Brexit flows and IoT devices - The Register - October 8th, 2019
- Education and Enhancement in a Transhuman Future - Patheos - October 8th, 2019
- Transhumanism - The Next Step to Super Humans? - April 21st, 2019
- Transhumanist politics - Wikipedia - April 12th, 2019
- Christian Transhumanist Conference 2018 Tickets, Sat, Aug ... - March 29th, 2019
- Nanochips & Smart Dust: New Face of the Human ... - March 29th, 2019
- Yudkowsky - Simplified Humanism - March 18th, 2019
- FM-2030 - Wikipedia - March 18th, 2019
- Space Fence: Connecting the Surveillance and Transhumanist ... - March 18th, 2019
- Biological immortality - Wikipedia - February 28th, 2019
- Transhumanism - Ascension Glossary - December 31st, 2018
- Transhumanism - H+Pedia - hpluspedia.org - December 31st, 2018
- Nick Bostrom - Wikipedia - December 4th, 2018
- Leadership U.S. Transhumanist Party Official Website - November 8th, 2018
- 2045: A New Era for Humanity - YouTube - August 16th, 2018
- The Transhumanist Wager: Zoltan Istvan: 9780988616110 ... - July 24th, 2018
- George Dvorsky - Wikipedia - June 18th, 2018
- Stronger, smarter, happier - what if a drug could make you a better version of yourself? - CBC.ca - August 23rd, 2017
- Bloodborne, Transhumanism and Cosmic Cyberpunk - Kotaku UK (blog) - August 22nd, 2017
- Ambrosia: the startup harvesting the blood of the young | Society ... - The Guardian - August 22nd, 2017
- Immortality: Silicon Valley's latest obsession ushers in the transhumanist era - South China Morning Post - August 17th, 2017
- Transhumanism Is Not Libertarian, It's an Abomination - The American Conservative - August 17th, 2017