When I started to work on this map of AI safety solutions, I wanted to illustrate the excellent 2013 article “Responses to Catastrophic AGI Risk: A Survey” by Kaj Sotala and IEET Affiliate Scholar Roman V. Yampolskiy, which I strongly recommend. However, during the process I had a number of ideas to expand the classification of the proposed ways to create safe AI.
This map shows that AI failure resulting in human extinction could happen on different levels of AI development, namely:
1. before it starts self-improvement (which is unlikely but we still can envision several failure modes),
2. during its take off, when it uses different instruments to break out from its initial confinement, and
3. after its successful takeover of the world, when it starts to implement its goal system which could be unfriendly, or its friendliness may be flawed.
I’ve been stumping for some time about how Artificial Intelligence will provide the shortest path to curing aging forever. In fact, without it, I’m convinced we won’t solve aging in our lifetime. I’m glad to hear Peter Diamandis describe AI as the most important technology we’re developing this decade.
Peter goes on to say it’s a massive opportunity for humanity, not a threat, as well as the following:
One thing I promise, when we do politics here. It won’t be stuff you are reading anywhere else.
Cranking back NSA spying…?
Topmost in the news, recently, the shocking ability of the U.S. Congress to actually pass a compromise bill, one that dials back a few of the powers given (since 9/11) to our Professional Protector Caste (PPC) in the Patriot Act.
On Friday March 6, 2015, more than 3,000 people attended the ASU Emerge event. This is where Eric Kingsbury, futurist, founder of KITEBA, cofounder of the Confluence Project, launched “You Have Been Inventoried”. I helped with some of the content for the project, along with others from the Confluence Project.
Everywhere you look in the world you can see pessimism, gloom, doom and negativity. No matter where you live, it seems many are convinced that there’s just no hope. Many people have stopped trying to do anything, while they “wait for god” or “wait for the Singularity.” Or simply wait, period.
The negativity is everywhere.
So, here’s one of my rants, against that negativity.
This article examines the risks posed by “unknown unknowns,” which I call monsters. It then introduces a taxonomy of the unknowable, and argues that one category of this taxonomy in particular should lead us to inflate our prior probability estimates of annihilation, whatever they happen to be. The lesson here is ultimately the same as the Doomsday Argument, except the reasoning is far more robust.
Even if they aren’t flesh, “mindclones” deserve protection.
For much of the 20th century, capital punishment was carried out in most countries. During the preceding century many, like England, had daily public hangings. Today, even Russia, with a mountainous history of government-ordered executions, has a capital-punishment moratorium. Since 1996, it has not executed a criminal through the judicial system.
If we can learn to protect the lives of serial killers, child mutilators, and terrorists, surely we can learn to protect the lives of peace-loving model citizens known as mind clones and bemans—even if they initially seem odd or weird to us.
An advanced artificial intelligence (a “superintelligence”) could pose a significant existential risk to humanity. Several research institutes have been set-up to address those risks. And there is an increasing number of academic publications analysing and evaluating their seriousness. Nick Bostrom’s Superintelligence: Paths, Dangers, Strategies represents the apotheosis of this trend. In this article, I argue that in defending the credibility of AI risk, Bostrom makes an epistemic move that is analogous to one made by so-called sceptical theists in the debate about the existence of God. And while this analogy is interesting in its own right, what is more interesting is its potential implication. It has been repeatedly argued that sceptical theism has devastating effects on our beliefs and practices. Could it be that AI-doomsaying has similar effects? I argue that it could. Specifically, and somewhat paradoxically, I argue that it could lead to either a reductio of the doomsayers position, or an important and additional reason to join their cause. I use this paradox to suggest that the modal standards for argument in the superintelligence debate need to be addressed.
Are there ideas which could prove so incendiary, and so provocative, that it would be better to shut them down? Should some concepts be permanently locked into a Pandora’s box, lest they fly off and cause too much chaos in the world?
On January 20, the Electronic Frontier Foundation (EFF) announced the Apollo 1201 project, an effort to eradicate digital rights management (DRM) schemes from the world of Internet commerce. Led by well-known activist Cory Doctorow, the project aims to “accelerate the movement to repeal laws protecting DRM” and “kick-start a vibrant market in viable, legal alternatives to digital locks.” According to EFF, DRM technologies “threaten users’ security and privacy, distort markets, undermine innovation,” and don’t effectively protect so-called “intellectual property.”
Sony hacks, barbarians with FaceBook pages, troll armies, ministries of “truth”- it wasn’t supposed to be like this. When the early pioneers of what we now call the Internet freed the network from the US military they were hoping for a network of mutual trust and sharing- a network like the scientific communities in which they worked where minds were brought into communion from every corner of the world. It didn’t take long for some of the witnesses to the global Internet’s birth to see in it the beginnings of a global civilization, the unification, at last, of all of humanity under one roof brought together in dialogue by the miracle of a network that seemed to eliminate the parochialism of space and time.
I am really looking forward to Frank Pasquale’s new book The Black Box Society: The Secret Algorithms that Control Money and Information. The book looks to examine and critique the ways in which big data is being used to analyse, predict and control our behaviour. Unfortunately, it is not out until January 2015. In the meantime, I’m trying to distract myself with some of Pasquale’s previously published material.
On August 31 of this year, nearly 200 celebrities had their private images hacked and released for the entire world to see. These images ranged from the normal day-to-day activities, to their utmost private moments – from nudity to sex. This event hit both mainstream and social media airwaves, flooding the online sphere under the hashtags #Celebgate and the #Fappening.
It is a risky business trying to predict the future, and although it makes some sense to try to get a handle on what the world might be like in one’s lifetime, one might wonder what’s even the point of all this prophecy that stretches out beyond the decades one is expected to live? The answer I think is that no one who engages in futurism is really trying to predict the future so much as shape it, or at the very least, inspire Noah like preparations for disaster.
Google Inc.’s 2013 book The New Digital Age, authored by Google chairman Eric Schmidt and Google Ideas director Jared Cohen, was showered with praise by many, but attacked in a review by Julian Assange for the New York Times, where it is described as a “love song” from Google to the US state. Also addressed in Assange’s subsequent book When Google Met WikiLeaks, Google’s book makes an unconvincing effort to depict the internet as a double-edged sword, both empowering (p. 6) and threatening our lives (p. 7).
Julian Assange’s 2014 book When Google Met WikiLeaks consists of essays authored by Assange and, more significantly, the transcript of a discussion between Assange and Google’s Eric Schmidt and Jared Cohen.
Between us and the future stands an almost impregnable wall that cannot be scaled. We cannot see over it,or under it, or through it, no matter how hard we try. Sometimes the best way to see the future is by using the same tools we use in understanding the present which is also, at least partly, hidden from direct view by the dilemma inherent in our use of language.
Don’t believe the hype. You’re the customer, whether you pay directly or by seeing ads. Tell me if you’ve heard this one before: “On the internet, if you’re not paying for something, then you’re not the customer. You’re the product.”
Wired has a long form interview with Edward Snowden: The Most-Wanted Man in the World. A must-read… as far as it goes. Only keep ahold of your ability to parse complexities and contradictions, because my reflex is always to point out aspects that were never raised. I refuse to choose one "side's" purist reflex. So should you.
The growing body of work in the new field of “affective robotics” involves both theoretical and practical ways to instill – or at least imitate – human emotion in Artificial Intelligence (AI), and also to induce emotions toward AI in humans. The aim of this is to guarantee that as AI becomes smarter and more powerful, it will remain tractable and attractive to us. Inducing emotions is important to this effort to create safer and more attractive AI because it is hoped that instantiation of emotions will eventually lead to robots that have moral and ethical codes, making them safer; and also that humans and AI will be able to develop mutual emotional attachments, facilitating the use of robots as human companions and helpers. This paper discusses some of the more significant of these recent efforts and addresses some important ethical questions that arise relative to these endeavors.
Without clear rules for cyberwarfare, technology workers could find themselves fair game in enemy attacks and counterattacks. If they participate in military cyberoperations—intentionally or not—employees at Facebook, Google, Apple, Microsoft, Yahoo!, Sprint, AT&T, Vodaphone, and many other companies may find themselves considered “civilians directly participating in hostilities” and therefore legitimate targets of war, according to the legal definitions of the Geneva Conventions and their Additional Protocols.
Within the next few years, autonomous vehicles—alias robot cars—could be weaponized, the US Federal Bureau of Investigation (FBI) fears. In a recently disclosed report, FBI experts wrote that they believe that robot cars would be “game changing” for law enforcement. The self-driving machines could be professional getaway drivers, to name one possibility. Given the pace of developments on autonomous cars, this doesn’t seem implausible.
The resilience of our entire civilization is increasingly reliant on a fragile network of cell phone towers, which are the first things to fail in any crisis, e.g. a hurricane or other natural disaster… or else deliberate (e.g. EMP or hacker) sabotage.