Video (starts 27 minutes in) * Powerpoint (unfortunately not visible in the video) presented March 8, 2010 at The Third Conference on Artificial General Intelligence (Lugano, Switzerland)
Inspired by a question after the previous day’s Designing a Safe Motivational System for Intelligent Machines presentation. Immediately followed by a rebuttal by and debate/Q&A with SIAI’s Roko Mijic.
Video * Presentation Powerpoint * Proceedings Paper presented March 7, 2010 at The Third Conference on Artificial General Intelligence (Lugano, Switzerland)
As machines become more intelligent, more flexible, more autonomous and more powerful, the questions of how they should choose their actions and what goals they should pursue become critically important. Drawing upon the examples of and lessons learned from humans and lesser creatures, we propose a hierarchical motivational system flowing from an abstract invariant super-goal that is optimal for all (including the machines themselves) to low-level reflexive “sensations, emotions, and attentional effects” and other enforcing biases to ensure reasonably “correct” behavior even under conditions of uncertainty, immaturity,error, malfunction, and even sabotage.
Presentation Powerpoint * Proceedings PDF presented Nov. 6, 2009 at the AAAI 2009 Fall Symposium on Biologically Inspired Cognitive Architectures (Washington, DC)
As machines become more intelligent and take on more responsibilities, their decision-making capabilities must be informed and constrained by a coherent, integrated moral/ethical structure with no internal inconsistencies for everyone’s safety and well-being. Unfortunately, no such structure is currently agreed upon to exist. We propose to solve this problem by a) drawing upon experimental evidence and lessons learned from evolution and economics to show that morality is actually objective and derivable from first principles; b) presenting a coherent, integrated, platonic ethical system with no internal inconsistencies that flows naturally from a single high-level logically-derived Kantian imperative to low-level reflexive “rules of thumb” that match current human sensibilities; and c) suggesting a biologically-inspired architecture which supports and enforces this system which can be relatively easily implemented.