Institute for Ethics and Emerging Technologies


The IEET is a 501(c)3 non-profit, tax-exempt organization registered in the State of Connecticut in the United States. Please give as you are able, and help support our work for a brighter future.


Search the IEET
Subscribe and Contribute to:


Technoprogressive? BioConservative? Huh?
Quick overview of biopolitical points of view




whats new at ieet

The Super-Shulgin Academy: A Singularity I Can Believe In

Network Society: the coming socio-economic phase transformation

Daniel Dennett: In Defense of Robotic Consciousness

Stefan Sorgner speaking on Transhumanism and Star Wars

Against Transhuman Separatism: Breakaway Cultures Become Broken Cultures

Will Reproductive Rights Advocates Stand Up for Men?


ieet books

Keywords for Environmental Studies
Author
eds. Joni Adamson, William A. Gleason, David N. Pellow


comments

spud100 on 'The Super-Shulgin Academy: A Singularity I Can Believe In' (Feb 11, 2016)

randomname on '"McMindfulness": is Buddhism contaminated by capitalism?" - interview with Terry Hyland' (Feb 10, 2016)

Giulio Prisco on 'The Small Mammal Brain Preservation Prize Has Been Won' (Feb 10, 2016)

Giulio Prisco on 'The Small Mammal Brain Preservation Prize Has Been Won' (Feb 10, 2016)

Giulio Prisco on 'The Small Mammal Brain Preservation Prize Has Been Won' (Feb 10, 2016)

almostvoid on 'John Searle’s Critique of Ray Kurzweil' (Feb 10, 2016)

almostvoid on 'Apple Search Finally Stops Directing People Seeking Abortions to Adoption Centers' (Feb 10, 2016)







Subscribe to IEET News Lists

Daily News Feed

Longevity Dividend List

Catastrophic Risks List

Biopolitics of Popular Culture List

Technoprogressive List

Trans-Spirit List



JET

Enframing the Flesh: Heidegger, Transhumanism, and the Body as “Standing Reserve”

Moral Enhancement and Political Realism

Intelligent Technologies and Lost Life

Hottest Articles of the Last Month


Basic Income Guarantee will allow us to move up the Maslow Pyramid - interview with Gerd Leonhard
Feb 7, 2016
(13083) Hits
(0) Comments

Becoming the First Transhuman: A Call For The Right Stuff
Jan 17, 2016
(7409) Hits
(2) Comments

“McMindfulness”: is Buddhism contaminated by capitalism?” - interview with Terry Hyland
Feb 6, 2016
(6228) Hits
(3) Comments

The one percent discovers transhumanism: Davos 2016
Feb 6, 2016
(5490) Hits
(1) Comments



Comment on this entry

The Blackjack Generation


David Brin


Ethical Technology

January 27, 2012

In this second selection of speculative fiction, and excerpt from a forthcoming novel, David Brin asks how we will keep our machine mind progeny loyal.


...

Complete entry


COMMENTS



Posted by Greg Hullender  on  01/28  at  02:52 PM

I’ve worked on machine learning and natural language since about 1979. I recently went back to school to get a masters in computational linguistics. I made a study of Dialogue Systems, which are software systems meant to be conversational. A “General Dialogue System” would, by definition, pass the Turing test. “Practical Dialogue Systems” exist today (I actually wrote one for a class project) and let you converse as long as you stick to tasks the computer can do. (E.g. make a plane reservation.) If you haven’t, you ought to read about dialogue systems.

That said, I spent a lot of time thinking about why we can’t make a general dialogue system. I concluded that the problem is with volition. WHY does the system do or say anything? I could create a system that correctly parsed and “understood” everything a person said, but then what? What does it do with that info and why? The only answer I could find is that it will only “want” to do what people have programmed it to.

So it would never be a crime to create an AI—just to give one motivations that are contrary to human interests. Much as it’s no crime to use dynamite, as long as you don’t try to hurt people with it.

I could visualize highly paid teams of “Volition Designers,” who construct the rules that govern deep behavior of an AI. (Asimov’s laws of Robotics are a very, very simple form of this; the Robot’s only volition is to a) save people b) obey orders c) protect itself. Otherwise, it’ll just sit there.)

So could you have a self-modifying AI? Meaning one which could change its own volition template? I almost want to claim that that’s logically impossible; whatever code decides when and why to change the template is now the real volition system.

So could a system kill people? Sure. It could have bugs (like HAL 9000—everyone’s favorite General Dialogue System), or a malicious programmer could build one, but I can’t see one evolving. Would they get rights? I hope not, but people can be dumb. Someone could easily make an AI that mimed human feelings, swaying the public into pushing for such rights. (I already see people get attached to characters in video games, and those have almost no ability to communicate.)

One thing’s for sure; we have made zero progress toward a machine that “wants” to do anything. Something about human intelligence is very, very different from anything we have ever put on any machine ever made. It may not seem that way to someone outside the field, but everyone doing serious work in the area knows it.

 





Posted by nicholasjh1  on  12/12  at  05:27 PM

You’re asumming the conversational programs have anything at all to do with AI. I would say the problem of getting AI is largely a hardware one. We will never get a general intelligence AI with Von neuman Machines, however the hardware for AI is currently being worked on. Complexity for General AI as good as ours my be difficult though, We do have a lot of underlieing programming.






Add your comment here:


Name:

Email:

Location:

Remember my personal information

Notify me of follow-up comments?

HOME | ABOUT | FELLOWS | STAFF | EVENTS | SUPPORT  | CONTACT US
SECURING THE FUTURE | LONGER HEALTHIER LIFE | RIGHTS OF THE PERSON | ENVISIONING THE FUTURE
CYBORG BUDDHA PROJECT | AFRICAN FUTURES PROJECT | JOURNAL OF EVOLUTION AND TECHNOLOGY

RSSIEET Blog | email list | newsletter |
The IEET is a 501(c)3 non-profit, tax-exempt organization registered in the State of Connecticut in the United States.

East Coast Contact: Executive Director, Dr. James J. Hughes,
56 Daleville School Rd., Willington CT 06279 USA 
Email: director @ ieet.org     phone: 860-428-1837

West Coast Contact: Managing Director, Hank Pellissier
425 Moraga Avenue, Piedmont, CA 94611
Email: hank @ ieet.org