Materials and how we use them are inextricably linked to the development of human society. Yet amazing as historic achievements using stone, wood, metals and other substances seem, these are unbelievably crude compared to the full potential of what could be achieved with designer materials.
This is the sixth part in my series on Nick Bostrom’s recent book Superintelligence: Paths, Dangers, Strategies. The series is covering those parts of the book that most interest me. This includes the sections setting out the basic argument for thinking that the creation of superintelligent AI could threaten human existence, and the proposed methods for dealing with that threat.
“Every office full of ambitious people has them. And we have all worked with at least one—the co-worker with an inexplicable ability to rise in the ranks,” wrote the Wall Street Journal recently in an article entitled What Corporate Climbers Can Teach Us. “‘How do they do it?’ we may ask ourselves or whisper to friends at work,” it continued. “They don't have more experience. They don't seem that brilliant.”
Within the next few years, autonomous vehicles—alias robot cars—could be weaponized, the US Federal Bureau of Investigation (FBI) fears. In a recently disclosed report, FBI experts wrote that they believe that robot cars would be “game changing” for law enforcement. The self-driving machines could be professional getaway drivers, to name one possibility. Given the pace of developments on autonomous cars, this doesn’t seem implausible.
The transfer of used military equipment from the armed forces to police departments around the country has been accompanied, at least to a certain extent, by a shift in public thinking. The news media have played a critical part in that shift, both in its coverage and in what it chooses not to cover.
On August 9, at around 12 in the afternoon, Michael Brown and his friend Dorian Johnson were attacked by Ferguson, Missouri police officer Darren Wilson. With his hands in the air, telling Officer Wilson that he was unarmed, the officer shot Brown several times, killing him as a result. This was the eyewitness account told by Brown’s friend Dorian.
Positive moods are a virtue, both in enabling enjoyment of life and in supporting prosocial behavior. But it is not the only kind of happiness, and in excess can be quite excessive. Along with positive moods we also want to cultivate flourishing, a sense that overall our lives are meaningful and going well. What are the public policies and life behaviors that support positive moods and flourishing lives? As we enter a “hedonistic imperative” future in which we are able to tweak our moods with “happy-people-pills-for-all” how will we find the right balance of positive mood to achieve flourishing lives?
The WHO medical ethics panel convened Monday to discuss the ethics of using experimental treatments for Ebola in West African nations affected by the disease. I am relieved to note that this morning they released their unanimous recommendation: “it is ethical to offer unproven interventions with as yet unknown efficacy and adverse effects, as potential treatment or prevention.”
Human beings seem to have an innate need to predict the future. We’ve read the entrails of animals, thrown bones, tried to use the regularity or lack of it in the night sky as a projection of the future and omen of things to come, along with a thousand others kinds of divination few of us have ever heard of. This need to predict the future makes perfect sense for a creature whose knowledge bias is towards the present and the past. Survival means seeing enough ahead to avoid dangers, so that an animal that could successfully predict what was around the next corner could avoid being eaten or suffering famine.
The resilience of our entire civilization is increasingly reliant on a fragile network of cell phone towers, which are the first things to fail in any crisis, e.g. a hurricane or other natural disaster… or else deliberate (e.g. EMP or hacker) sabotage.
Debate about the merits of enhancement tends to pretty binary. There are some — generally called bioconservatives — who are opposed to it; and others — transhumanists, libertarians and the like — who embrace it wholeheartedly. Is there any hope for an intermediate approach? One that doesn’t fall into the extremes of reactionary reject or uncritical endorsement?
Machine ethics is a term used in different ways. The basic use is in the sense of people attempting to instill some sort of human-centric ethics or morality in the machines we build like robots, self-driving vehicles, and artificial intelligence (Wallach 2010) so that machines do not harm humans either maliciously or unintentionally.
I just finished a thrilling little book about the first machine war. The author writes of a war set off by a terrorist attack where the very speed of machines being put into action,and the near light speed of telecommunications whipping up public opinion to do something now, drives countries into a world war. In his vision whole new theaters of war, amounting to fourth and fifth dimensions, have been invented. Amid a storm of steel huge hulking machines roam across the landscape and literally shred human beings in their path to pieces. Low flying avions fill the sky taking out individual targets or help calibrate precision attacks from incredible distances beyond. Wireless communications connect soldiers and machine together in a kind of world-net…
Transhumanists as a rule may prefer to contemplate implants and genetic engineering, but few if any violations of morphological freedom exceed being torn to pieces by shrapnel or dashed against concrete by an overpressure wave. In this piece I argue that the settler-colonial violence in occupied Palestine relates to core aspects of modernity and demands futurist attention both emotionally and intellectually.
Robot cars and military robots have more in common than you’d think. Some accidents with self-driving cars will result in fatalities, and this may be troubling in ways that human-caused fatalities are not. But is it really worse to be killed by a robot than by a drunk driver—or by a renegade soldier?
This is the fourth post of my series on Nick Bostrom’s recent book Superintelligence: Paths, Dangers, Strategies. In the previous post, I started my discussion of Bostrom’s argument for an AI doomsday scenario. Today, I continue this discussion by looking at another criticism of that argument, along with Bostrom’s response.
What has the Maker Movement got to do with public health? Quite a lot as it turns out, as I explore in the latest Risk Bites video. This in turn was inspired by being invited to talk at the inaugural We Make Health Fest in Ann Arbor (August 16 – please join us if you can!).
This is the second post in my series on Nick Bostrom’s recent book Superintelligence: Paths, Dangers, Strategies. In the previous post, I looked at Bostrom’s defence of the orthogonality thesis. This thesis claimed that pretty much any level of intelligence — when “intelligence” is understood as skill at means-end reasoning — is compatible with pretty much any (final) goal. Thus, an artificial agent could have a very high level of intelligence, and nevertheless use that intelligence to pursue very odd final goals, including goals that are inimical to the survival of human beings. In other words, there is no guarantee that high levels of intelligence among AIs will lead to a better world for us.
In this entry, I take a look at Bostrom’s orthogonality thesis. As we shall see, this thesis is central to his claim that superintelligent AIs could pose profound existential risks to human beings. But what does the thesis mean and how plausible is it?
Lately, I’ve been enjoying reruns of the relatively new BBC series Sherlock, starring Benedict Cumberbatch, which imagines Arthur Conan Doyle’s famous detective in our 21st century world. The thing I really enjoy about the show is that it’s the first time I can recall that anyone has managed to make Sherlock Holmes funny without at the same time undermining the whole premise of a character whose purely logical style of thinking make him seem more a robot than a human being.
There may be as many as 80,000 American prisoners currently locked-up in a SHU, or segregated housing unit. Solitary confinement in a SHU can cause irreversible psychological effects in as little as 15 days. Here’s what social isolation does to your brain, and why it should be considered torture.
My goal in this article is to demolish the AI Doomsday scenarios that are being heavily publicized by the Machine Intelligence Research Institute, the Future of Humanity Institute, and others, and which have now found their way into the farthest corners of the popular press. These doomsday scenarios are logically incoherent at such a fundamental level that they can be dismissed as extremely implausible - they require the AI to be so unstable that it could never reach the level of intelligence at which it would become dangerous. On a more constructive and optimistic note, I will argue that even if someone did try to build the kind of unstable AI system that might lead to one of the doomsday behaviors, the system itself would immediately detect the offending logical contradiction in its design, and spontaneously self-modify to make itself safe.
So I finally got around to reading Max Tegmark’s book Our Mathematical Universe, and while the book answered the question that had led me to read it, namely, how one might reconcile Plato’s idea of eternal mathematical forms with the concept of multiple universes, it also threw up a whole host of new questions. This beautifully written and thought provoking book made me wonder about the future of science and the scientific method, the limits to human knowledge, and the scientific, philosophical and moral meaning of various ideas of the multiverse.
Over the past few weeks, revelations of potentially dangerous errors in US federal labs handling pathogens have placed health and safety high on the national agenda. In June, the US Centers for Disease Control and Prevention (CDC) announced as many as 75 of its staff may have been exposed to anthrax due to safety issues at one of its labs. At the beginning of July, vials of smallpox virus were found in an unsecured room at the National Institutes of Health (NIH). Then earlier this week came the revelation that in the same room were over 300 vials containing pathogens such as dengue virus, influenza, and the bacterium that causes Q fever.
Voltaire once said that “work saves a man from three great evils: boredom, vice and need.” Many people endorse this sentiment. Indeed, the ability to seek and secure paid employment is often viewed as an essential part of a well-lived life. Those who do not work are reminded of the fact. They are said to be missing out on a valuable and fulfilling human experience. The sentiment is so pervasive that some of the foundational documents of international human rights law — including the UN Declaration of Human Rights (UDHR Art. 23) and the International Covenant on Economic, Social and Cultural Rights (ICESCR Art. 6) — recognise and enshrine the “right to work”.
Most broadly, Social Futurism stands for positive social change through technology; i.e. to address social justice issues in radically new ways which are only just now becoming possible thanks to technological innovation. If you would like some introduction to Social Futurist ideas, you can read the introduction page at wavism.net and there are links to articles at http://IEET.org listed at the top of this post. In this post I will discuss the Social Futurist alternative to Liberal Democratic and Authoritarian states, how that model fits with our views on decentralization and subsidiarity, and its relevance to the political concept of a “Third Way“.