SLIDE 1
1
A peer-reviewed electronic journal published by the Institute for Ethics and Emerging Technologies ISSN 1541-0099
22(1) – February 2012
Nine Ways to Bias Open-Source AGI Toward Friendliness
Ben Goertzel and Joel Pitt Novamente LLC ben@goertzel.org
Journal of Evolution and Technology - Vol. 22 Issue 1 – February 2012 - pgs 116-131
Abstract While it seems unlikely that any method of guaranteeing human-friendliness (“Friendliness”) on the part of advanced Artificial General Intelligence (AGI) systems will be possible, this doesn’t mean the only alternatives are throttling AGI development to safeguard humanity, or plunging recklessly into the complete unknown. Without denying the presence of a certain irreducible uncertainty in such matters, it is still sensible to explore ways of biasing the odds in a favorable way, such that newly created AI systems are significantly more likely than not to be Friendly. Several potential methods of effecting such biasing are explored here, with a particular but non- exclusive focus on those that are relevant to open-source AGI projects, and with illustrative examples drawn from the OpenCog open-source AGI project. Issues regarding the relative safety
- f open versus closed approaches to AGI are discussed and then nine techniques for biasing AGIs
in favor of Friendliness are presented:
- 1. Engineer the capability to acquire integrated ethical knowledge.
- 2. Provide rich ethical interaction and instruction, respecting developmental stages.
- 3. Develop stable, hierarchical goal systems.
- 4. Ensure that the early stages of recursive self-improvement occur relatively slowly and
with rich human involvement.
- 5. Tightly link AGI with the Global Brain.
- 6. Foster deep, consensus-building interactions between divergent viewpoints.
- 7. Create a mutually supportive community of AGIs.
- 8. Encourage measured co-advancement of AGI software and AGI ethics theory
- 9. Develop advanced AGI sooner not later.