Sony hacks, barbarians with FaceBook pages, troll armies, ministries of “truth”- it wasn’t supposed to be like this. When the early pioneers of what we now call the Internet freed the network from the US military they were hoping for a network of mutual trust and sharing- a network like the scientific communities in which they worked where minds were brought into communion from every corner of the world. It didn’t take long for some of the witnesses to the global Internet’s birth to see in it the beginnings of a global civilization, the unification, at last, of all of humanity under one roof brought together in dialogue by the miracle of a network that seemed to eliminate the parochialism of space and time.
I am really looking forward to Frank Pasquale’s new book The Black Box Society: The Secret Algorithms that Control Money and Information. The book looks to examine and critique the ways in which big data is being used to analyse, predict and control our behaviour. Unfortunately, it is not out until January 2015. In the meantime, I’m trying to distract myself with some of Pasquale’s previously published material.
On August 31 of this year, nearly 200 celebrities had their private images hacked and released for the entire world to see. These images ranged from the normal day-to-day activities, to their utmost private moments – from nudity to sex. This event hit both mainstream and social media airwaves, flooding the online sphere under the hashtags #Celebgate and the #Fappening.
It is a risky business trying to predict the future, and although it makes some sense to try to get a handle on what the world might be like in one’s lifetime, one might wonder what’s even the point of all this prophecy that stretches out beyond the decades one is expected to live? The answer I think is that no one who engages in futurism is really trying to predict the future so much as shape it, or at the very least, inspire Noah like preparations for disaster.
Google Inc.’s 2013 book The New Digital Age, authored by Google chairman Eric Schmidt and Google Ideas director Jared Cohen, was showered with praise by many, but attacked in a review by Julian Assange for the New York Times, where it is described as a “love song” from Google to the US state. Also addressed in Assange’s subsequent book When Google Met WikiLeaks, Google’s book makes an unconvincing effort to depict the internet as a double-edged sword, both empowering (p. 6) and threatening our lives (p. 7).
Julian Assange’s 2014 book When Google Met WikiLeaks consists of essays authored by Assange and, more significantly, the transcript of a discussion between Assange and Google’s Eric Schmidt and Jared Cohen.
Between us and the future stands an almost impregnable wall that cannot be scaled. We cannot see over it,or under it, or through it, no matter how hard we try. Sometimes the best way to see the future is by using the same tools we use in understanding the present which is also, at least partly, hidden from direct view by the dilemma inherent in our use of language.
Don’t believe the hype. You’re the customer, whether you pay directly or by seeing ads. Tell me if you’ve heard this one before: “On the internet, if you’re not paying for something, then you’re not the customer. You’re the product.”
Wired has a long form interview with Edward Snowden: The Most-Wanted Man in the World. A must-read… as far as it goes. Only keep ahold of your ability to parse complexities and contradictions, because my reflex is always to point out aspects that were never raised. I refuse to choose one "side's" purist reflex. So should you.
The growing body of work in the new field of “affective robotics” involves both theoretical and practical ways to instill – or at least imitate – human emotion in Artificial Intelligence (AI), and also to induce emotions toward AI in humans. The aim of this is to guarantee that as AI becomes smarter and more powerful, it will remain tractable and attractive to us. Inducing emotions is important to this effort to create safer and more attractive AI because it is hoped that instantiation of emotions will eventually lead to robots that have moral and ethical codes, making them safer; and also that humans and AI will be able to develop mutual emotional attachments, facilitating the use of robots as human companions and helpers. This paper discusses some of the more significant of these recent efforts and addresses some important ethical questions that arise relative to these endeavors.
Without clear rules for cyberwarfare, technology workers could find themselves fair game in enemy attacks and counterattacks. If they participate in military cyberoperations—intentionally or not—employees at Facebook, Google, Apple, Microsoft, Yahoo!, Sprint, AT&T, Vodaphone, and many other companies may find themselves considered “civilians directly participating in hostilities” and therefore legitimate targets of war, according to the legal definitions of the Geneva Conventions and their Additional Protocols.
Within the next few years, autonomous vehicles—alias robot cars—could be weaponized, the US Federal Bureau of Investigation (FBI) fears. In a recently disclosed report, FBI experts wrote that they believe that robot cars would be “game changing” for law enforcement. The self-driving machines could be professional getaway drivers, to name one possibility. Given the pace of developments on autonomous cars, this doesn’t seem implausible.
The resilience of our entire civilization is increasingly reliant on a fragile network of cell phone towers, which are the first things to fail in any crisis, e.g. a hurricane or other natural disaster… or else deliberate (e.g. EMP or hacker) sabotage.
My goal in this article is to demolish the AI Doomsday scenarios that are being heavily publicized by the Machine Intelligence Research Institute, the Future of Humanity Institute, and others, and which have now found their way into the farthest corners of the popular press. These doomsday scenarios are logically incoherent at such a fundamental level that they can be dismissed as extremely implausible - they require the AI to be so unstable that it could never reach the level of intelligence at which it would become dangerous. On a more constructive and optimistic note, I will argue that even if someone did try to build the kind of unstable AI system that might lead to one of the doomsday behaviors, the system itself would immediately detect the offending logical contradiction in its design, and spontaneously self-modify to make itself safe.
If you get just old enough, one of the lessons living through history throws you is that dreams take a long time to die. Depending on how you date it, communism took anywhere from 74 to 143 years to pass into the dustbin of history, though some might say it is still kicking. The Ptolemaic model of the universe lasted from 100 AD into the 1600′s. Perhaps even more dreams than not simply refuse to die, they hang on like ghost, or ghouls, zombies or vampires, or whatever freakish version of the undead suits your fancy. Naming them would take up more room than I can post, and would no doubt start one too many arguments, all of our lists being different. Here, I just want to make an argument for the inclusion of one dream on our list of zombies knowing full well the dream I’ll declare dead will have its defenders.
Last post we observed the dynamics of the collective in the terms of a small tribe, and indicated that at this size, things worked pretty well. That is not to say that error modes were not possible, but that when error modes arose, there were mechanisms in place to deal with those errors. Essentially, at this scale, the ability of individuals to veil their actions in a wall of secrecy did not exist. While it is certainly possible for the individual to lie, cheat, steal and deceive, such actions could only be carried out to a limited extent, and carried repercussions that were deleterious to that individuals long term well being.
Intrigued by IEET Fellow Patrick Lin’s essay “The Ethics of Autonomous Cars” we asked “Should your robot car sacrifice your life if it will save more lives?” A third of of the 196 of you who responded said no, a third said they should and a third said it should be the driver’s option.
Tyler Cowen points to this great Marc Andreessen interview in the Washington Post that features him saying the following about net neutrality: So, I think the net neutrality issue is very difficult. I think it’s a lose-lose. It’s a good idea in theory because it basically appeals to this very powerful idea of permissionless innovation. But at the same time, I think that a pure net neutrality view is difficult to sustain if you also want to have continued investment in broadband networks.
Most of us in the west were raised with legends, myths and movies that taught Suspicion of Authority (SoA). Thanks to the great science fiction author, George Orwell, we share a compelling metaphor— Big Brother —propelling our fears about a future that may be dominated by tyrants.
For anyone thinking about the future relationship between nature-man-machines I’d like to make the case for the inclusion of an insightful piece of fiction to the canon. All of us have heard of H.G. Wells, Isaac Asimov or Arthur C. Clarke. And many, though perhaps fewer, of us have likely heard of fiction authors from the other side of the nature/technology fence, writers like Mary Shelley, or Ursula Le Guin, or nowadays, Paolo Bacigalupi, but certainly almost none of us have heard of Samuel Butler, or better, read his most famous novel Erewhon (pronounced with 3 short syllables E-re-Whon.)
IEET Fellow Kevin Lagrandeur’s book Androids and Intelligent Networks in Early Modern Literature and Culture: Artificial Slaves has been awarded an Honorable Mention by the Science Fiction and Technoculture Studies .
I am sure you have heard it constantly. "Google is (insert fear term here.)" They want to take over the internet, they are building skynet, they are invading our privacy, they are trying to become big brother, etc, etc, ad nausem. Be it Glass, or their recent acquisition of numerous robotics firms, to even hiring Ray Kurzweil, Google has recently been in the news a lot, usually as the big bad boogieman of whatever news story you are reading.
I’ve looked at data-mining andpredictive analytics before on this blog. As you know, there are many concerns about this type of technology and the increasing role it plays in our lives. Thus, for example, people are concerned about the oftentimes hidden way in which our data is collected prior to being “mined”. And they are concerned about how it is used by governments and corporations to guide their decision-making processes. Will we be unfairly targetted by the data-mining algorithms? Will they exercise too much control over socially important decision-making processes? I’ve reviewed some of these concerns before.
We asked whether “artificial general intelligence with self-awareness” or “uploaded personalities or emulations of human brains” were more of a threat to human beings. Almost three times as many of you thought AGI was more of a threat than uploaded personalities, and overall 62% of the 245 respondents thought one or the other or both were a threat.