From: Eliezer S. Yudkowsky (sentience@pobox.com)
Date: Sun May 18 2003 - 09:07:58 MDT
Philip Sutton wrote:
> Eliezer said:
> 
>>That's the problem with outsiders making up security precautions for
>>the project to take; at least one of them will, accidentally, end up
>>ruling out successful Friendliness development. 
> 
> I don't see a problem with outsiders making up precautions or proposed 
> solutions to help achieve the creation of safe AGIs.  Making up is one 
> thing and unilaterally imposing is another.
I stand corrected.
-- Eliezer S. Yudkowsky http://intelligence.org/ Research Fellow, Singularity Institute for Artificial Intelligence
This archive was generated by hypermail 2.1.5 : Tue Feb 21 2006 - 04:22:21 MST