From: Eliezer Yudkowsky (sentience@pobox.com)
Date: Wed Jun 02 2004 - 07:46:19 MDT
Ben Goertzel wrote:
>
> Philip Sutton wrote:
>
>>Eliezer,
>>
>>It appears that you now are wanting to create a non-sentient super,
>>general AI (optimisation process) rather than a sentient
>>super, general AI (optimisation process).
>
> I'm not sure that his proposed optimization process has to be all that
> *general*. Perhaps it has to be general in its potential applicability,
> but it seems to me that, in accordance with his proposal, it has to be
> applied only to some specific problems, viz
>
> -- inferring the "collective volition" of humanity
> -- estimating which actions are most likely to lead to the fulfillment
> of this collective volition
This obviously requires general intelligence, and a huge amount of power
that can only come from recursive self-improvement. But I suppose that is
what you meant by general applicability but not general application. There
is really no such thing as a "general application"; every optimization
process is doing *something*.
-- Eliezer S. Yudkowsky http://intelligence.org/ Research Fellow, Singularity Institute for Artificial Intelligence
This archive was generated by hypermail 2.1.5 : Tue Feb 21 2006 - 04:22:38 MST