Last time I attempted to grapple with R. Scott Bakker’s intriguing essay on what kinds of philosophy aliens might practice and remaining dizzied by questions.
Luckily, I had a book in my possession which seemed to offer me the answers, a book that had nothing to do with the a modern preoccupation like question of alien philosophers at all, but rather a metaphysical problem that had been barred from philosophy except among seminary students since Darwin; namely, whether or not there was such a thing as moral truth if God didn’t exist.
IEET co-founder Nick Bostrom, IEET Fellow Wendell Wallach and Affiliate Scholar Seth Baum are Principal Investigators on projects n funded by Elon Musk and the Open Philanthropy Project and administered by the Future of Life Institute.
Hackers calling themselves “The Impact Team” recently stole the customer data of Ashley Madison, an online dating service for people who are married or in committed relationships. Ashley Madison employs a slogan that says it all: “Life is short. Have an affair.”
During July and August, customer data was released online by the hackers: the upshot is that it’s now possible to identify many individuals who held Ashley Madison accounts. This includes such intimate details as their sexual fetishes and proclivities.
The new television show Humans raises some important ethical questions for a not-too-distant future society where human-looking domestic robots are commonplace. The 8 part series, shown on AMC in the US and Channel 4 in the UK, is based on the Swedish series Äkta människor (“Real Humans”) and is set in modern day London with the only discernible difference being that a company is manufacturing and selling “synths” – multi-purpose robots designed to look like humans and work as direct replacements for them. The drama tackles a wide range of questions from how synths would be treated and their impact on society, alongside the main story line of what happens if the artificially intelligent humanoids gain true self awareness and consciousness.
I recently binge watched my first TV series, Humans, which airs Sunday nights on AMC. As a science fiction writer myself, many people have been suggesting I check it out for weeks now. Finally I gave in, sat down on the couch, and watched the first six episodes over the course of two days. Not bad for a mother of two. And now I have to wait two whole days to see what happens next!!!!!
When I started to work on this map of AI safety solutions, I wanted to illustrate the excellent 2013 article “Responses to Catastrophic AGI Risk: A Survey” by Kaj Sotala and IEET Affiliate Scholar Roman V. Yampolskiy, which I strongly recommend. However, during the process I had a number of ideas to expand the classification of the proposed ways to create safe AI.
This map shows that AI failure resulting in human extinction could happen on different levels of AI development, namely:
1. before it starts self-improvement (which is unlikely but we still can envision several failure modes),
2. during its take off, when it uses different instruments to break out from its initial confinement, and
3. after its successful takeover of the world, when it starts to implement its goal system which could be unfriendly, or its friendliness may be flawed.
I’ve been stumping for some time about how Artificial Intelligence will provide the shortest path to curing aging forever. In fact, without it, I’m convinced we won’t solve aging in our lifetime. I’m glad to hear Peter Diamandis describe AI as the most important technology we’re developing this decade.
Peter goes on to say it’s a massive opportunity for humanity, not a threat, as well as the following:
One thing I promise, when we do politics here. It won’t be stuff you are reading anywhere else.
Cranking back NSA spying…?
Topmost in the news, recently, the shocking ability of the U.S. Congress to actually pass a compromise bill, one that dials back a few of the powers given (since 9/11) to our Professional Protector Caste (PPC) in the Patriot Act.
On Friday March 6, 2015, more than 3,000 people attended the ASU Emerge event. This is where Eric Kingsbury, futurist, founder of KITEBA, cofounder of the Confluence Project, launched “You Have Been Inventoried”. I helped with some of the content for the project, along with others from the Confluence Project.
Everywhere you look in the world you can see pessimism, gloom, doom and negativity. No matter where you live, it seems many are convinced that there’s just no hope. Many people have stopped trying to do anything, while they “wait for god” or “wait for the Singularity.” Or simply wait, period.
The negativity is everywhere.
So, here’s one of my rants, against that negativity.
This article examines the risks posed by “unknown unknowns,” which I call monsters. It then introduces a taxonomy of the unknowable, and argues that one category of this taxonomy in particular should lead us to inflate our prior probability estimates of annihilation, whatever they happen to be. The lesson here is ultimately the same as the Doomsday Argument, except the reasoning is far more robust.
Even if they aren’t flesh, “mindclones” deserve protection.
For much of the 20th century, capital punishment was carried out in most countries. During the preceding century many, like England, had daily public hangings. Today, even Russia, with a mountainous history of government-ordered executions, has a capital-punishment moratorium. Since 1996, it has not executed a criminal through the judicial system.
If we can learn to protect the lives of serial killers, child mutilators, and terrorists, surely we can learn to protect the lives of peace-loving model citizens known as mind clones and bemans—even if they initially seem odd or weird to us.
An advanced artificial intelligence (a “superintelligence”) could pose a significant existential risk to humanity. Several research institutes have been set-up to address those risks. And there is an increasing number of academic publications analysing and evaluating their seriousness. Nick Bostrom’s Superintelligence: Paths, Dangers, Strategies represents the apotheosis of this trend. In this article, I argue that in defending the credibility of AI risk, Bostrom makes an epistemic move that is analogous to one made by so-called sceptical theists in the debate about the existence of God. And while this analogy is interesting in its own right, what is more interesting is its potential implication. It has been repeatedly argued that sceptical theism has devastating effects on our beliefs and practices. Could it be that AI-doomsaying has similar effects? I argue that it could. Specifically, and somewhat paradoxically, I argue that it could lead to either a reductio of the doomsayers position, or an important and additional reason to join their cause. I use this paradox to suggest that the modal standards for argument in the superintelligence debate need to be addressed.
Are there ideas which could prove so incendiary, and so provocative, that it would be better to shut them down? Should some concepts be permanently locked into a Pandora’s box, lest they fly off and cause too much chaos in the world?
On January 20, the Electronic Frontier Foundation (EFF) announced the Apollo 1201 project, an effort to eradicate digital rights management (DRM) schemes from the world of Internet commerce. Led by well-known activist Cory Doctorow, the project aims to “accelerate the movement to repeal laws protecting DRM” and “kick-start a vibrant market in viable, legal alternatives to digital locks.” According to EFF, DRM technologies “threaten users’ security and privacy, distort markets, undermine innovation,” and don’t effectively protect so-called “intellectual property.”
Sony hacks, barbarians with FaceBook pages, troll armies, ministries of “truth”- it wasn’t supposed to be like this. When the early pioneers of what we now call the Internet freed the network from the US military they were hoping for a network of mutual trust and sharing- a network like the scientific communities in which they worked where minds were brought into communion from every corner of the world. It didn’t take long for some of the witnesses to the global Internet’s birth to see in it the beginnings of a global civilization, the unification, at last, of all of humanity under one roof brought together in dialogue by the miracle of a network that seemed to eliminate the parochialism of space and time.
I am really looking forward to Frank Pasquale’s new book The Black Box Society: The Secret Algorithms that Control Money and Information. The book looks to examine and critique the ways in which big data is being used to analyse, predict and control our behaviour. Unfortunately, it is not out until January 2015. In the meantime, I’m trying to distract myself with some of Pasquale’s previously published material.
On August 31 of this year, nearly 200 celebrities had their private images hacked and released for the entire world to see. These images ranged from the normal day-to-day activities, to their utmost private moments – from nudity to sex. This event hit both mainstream and social media airwaves, flooding the online sphere under the hashtags #Celebgate and the #Fappening.
It is a risky business trying to predict the future, and although it makes some sense to try to get a handle on what the world might be like in one’s lifetime, one might wonder what’s even the point of all this prophecy that stretches out beyond the decades one is expected to live? The answer I think is that no one who engages in futurism is really trying to predict the future so much as shape it, or at the very least, inspire Noah like preparations for disaster.
Google Inc.’s 2013 book The New Digital Age, authored by Google chairman Eric Schmidt and Google Ideas director Jared Cohen, was showered with praise by many, but attacked in a review by Julian Assange for the New York Times, where it is described as a “love song” from Google to the US state. Also addressed in Assange’s subsequent book When Google Met WikiLeaks, Google’s book makes an unconvincing effort to depict the internet as a double-edged sword, both empowering (p. 6) and threatening our lives (p. 7).
Julian Assange’s 2014 book When Google Met WikiLeaks consists of essays authored by Assange and, more significantly, the transcript of a discussion between Assange and Google’s Eric Schmidt and Jared Cohen.