From: Eliezer Yudkowsky (sentience@pobox.com)
Date: Sat Oct 23 2004 - 07:01:55 MDT
Jeff Medina wrote:
> Robin Lee Powell wrote: "IIRC, Eliezer is not allowed to put Ph.D.
> after his name. That pretty much rules out this avenue of approach."
>
> That absolutely *does not* rule out this avenue of approach.
Correct.
> Many
> respected journals and conferences in the relevant areas are
> blind-reviewed (such that the academic credentials of the author of
> the paper is made irrelevant, because the author's identity & other
> info is kept secret), and even among the many which are not quality
> submissions are never rejected or looked down upon simply due to the
> lack of a Ph.D. by the author.
Papers which are *not* blind-reviewed show a *decided* bias toward known,
prestigious researchers and institutions. That is *why* some journals are
blind-reviewed. Most aren't, and it doesn't help that other journals are
blind-reviewed if the one you want to target isn't. I've read studies
assessing the bias, but though I googled on "effectiveness of peer review"
I failed to track them down. I did find other interesting material
including, to pick an arbitrary example, a study by Rothwell and Martyn
(2000) showing that peer review in two neuroscience journals was not
reproducible; that is, agreement between reviewers is not significantly
greater than chance. My recollection is that this result is widespread in
studies of this kind. I include this tidbit by way of saying that
experimental study of peer review has produced surprising and alarming
results, so be sure to check out peer-reviewed studies of peer review
before praising its effectiveness.
(Rothwell PM, Martyn CN. Reproducibility of peer review in clinical
neuroscience. Is agreement between reviewers any greater than would be
expected by chance alone? Brain 2000;123:1964-9.)
This has nothing to do with the reason I don't write additional academic
papers. I just thought I'd mention it.
**
First, you'll note that I say "write additional papers". I allocated one
month in early 2002 to write "Levels of Organization in General
Intelligence" (LOGI) in the best academic style I could manage. It
actually took four months. Since then the draft has been online at
http://intelligence.org/LOGI/. In 2005 this paper will finally appear in
"Artificial General Intelligence", eds. Goertzel and Pennachin, to be
published by Springer-Verlag in 2005. (The three-year delay was for the
entire book, not my own paper; *I* turned in my homework on time.)
The fact that none of the people plaguing me to write papers have even
*noticed* "Levels of Organization in General Intelligence", speaking
instead as if I haven't written *any* papers, is indeed related to the
reason I am not more involved with academia.
I've come a long way over the eight years since 1996. People said to me:
Write up your ideas about AI in a web page. In 1998 I did. Then new
people came along and they said: You'll never get anywhere with this, no
one will be interested enough to pay you to do this. In 2000, thanks to
Brian Atkins, the Singularity Institute started up. Possibly that
impressed a few people who never thought I'd get that far. Then new people
came along, to whom Eliezer had *always been* a part of the Singularity
Institute, so it wasn't impressive, and they said: No one will ever pay
attention to you unless you do as we say and write some kind of paper
targeted at academia and get it published. In 2002 I did. I didn't expect
anyone to notice, and no one did, but the effort of writing the LOGI paper
served to help me unify my ideas and force me to read relevant literature
and therefore I account it a partial success. And lo, the people said:
What you really need, Eliezer, is to write some kind of paper targeted at
academia.
Someone always thinks there's just one more thing you need to do. *That*
never changes, no matter how many times you fulfill the request. They just
find something else for you to do. Often it's something you've already
done. I wasn't puzzled by this. I expected it. Thus the particular
things that I did were selected strictly on the basis of their needing
doing, rather than to one-up naysayers.
Case in point: Dr. Eric Drexler and _Nanosystems_.
Before: Eric Drexler has no PhD and hasn't written up his ideas in great
gory technical detail. People tell him: Eric, no one will pay attention
to you if you don't have a PhD. People tell him: Eric, you need to write
up your technical ideas in great gory detail in a way that a wide audience
can understand.
Eric spends six years writing _Nanosystems_ and making it presentable to
any technical reader without demanding a specific background in chemistry,
physics, or computer science. Eric defends _Nanosystems_ as his thesis and
receives the world's first PhD in nanotechnology from MIT.
Afterward: None of the naysayers read _Nanosystems_ or even mention it
exists. No one pays any more attention to Drexler than before. They just
shift their criterion to something else Eric hasn't done yet. Often they
indignantly proclaim that Drexler hasn't given any technical presentation
of his ideas - complete indifference to the work already accomplished. The
same people who liked Drexler before still like him. The kind of people
who objected to Drexler before find something different to which to object.
I suspect those who objected to nanotechnology did not say: "Hm... I have
no idea whether I like this or not... but wait! Drexler doesn't have a
PhD! Okay, now I've decided that nanotechnology is impossible and Drexler
is scaring our children." The causal sequence of events is more like,
"Eek! Too weird! Hm, it seems that I disbelieve in nanotechnology. I
wonder why I disbelieve in nanotechnology? (Searches for reason.) It must
be because Drexler doesn't have a PhD, hey, yeah, that's it." After
Drexler got a PhD, exactly the same process took place, only the
rationalization search terminated elsewhere.
Drexler has a personality far better suited to academia than I'll ever be.
He's humble. He did everything by the book, the way he was supposed to.
Academia... to put it bluntly, they spit in his face. And Drexler had a
vastly easier problem to explain, in a field with all the underlying
physical equations established and agreed upon. If Drexler didn't make it
in academia there's no chance in hell that I could do so. Friendly AI
would be two orders of magnitude harder to sell to academia than molecular
nanotechnology. I pointed out that last part to Drexler, by the way; he
agreed. And come to think, while he didn't say a word to me against
academia or the academic system, Dr. Eric Drexler is *not* on the list of
people whose advice to me included getting a PhD.
I don't want to sound like I'm criticizing Drexler's intelligence. Drexler
did not have Drexler's case to warn him. Drexler's choices were different;
he may have had nothing better to try than getting a PhD and spending six
years writing a technical book.
But people seem to be absurdly optimistic about how easy it is for the
actors on stage to carry out the helpful advice shouted from the audience.
Then again, as plenty of studies show, people are also absurdly
optimistic about the course of their own lives - except for the severely
depressed, who are sometimes properly calibrated with respect to outcomes,
a phenomenon known as "depressive realism". (I am not making this up.)
Part of the reason why people are absurdly optimistic is that they think:
I'll just do X, and then everything will be all right! Not: I'll try to
do X, it will take four times as long as I expect, I'll probably fail, and
even if I succeed, only one in ten successes of this kind have as great an
impact as the one I pleasantly imagined.
I remember meeting Chris Phoenix of CRN at a Foresight Gathering, and Chris
Phoenix spoke optimistically of the day when molecular manufacturing is
proved possible, and all the naysayers have to admit it... and I said:
"Yes, Chris, we can look forward to the fine day when the naysayers are
presented with a working example of mechanosynthesis, and they are finally
forced to stand up and say, in unison: 'Oh, but that isn't *really*
nanotechnology.'"
Did I get a Ph.D., nothing would change. I'd just hear: oh, but you aren't
an eminent scientist in the field, go write more papers.
If I were the sort of person who chased all over the map - starting
companies, getting PhDs, whatever - then I wouldn't be here in the first
place. My life would have happened to me while I was making other plans.
Antoine de Saint-Exupéry: "Perfection is achieved, not when there is
nothing left to add, but when there is nothing left to take away." People
overestimate conjunctive probabilities and underestimate disjunctive
probabilities; they overestimate the chance of many things going right in
sequence, underestimate the probability of a single thing going wrong. The
way to success is to remove everything from the plan that doesn't
absolutely *have* to be there. The way to have any chance at all of
finishing on time is to do nothing that is not absolutely necessary.
Is being a part of academia absolutely necessary to success? I don't think
so. No one's told me to get a PhD in something because in-depth technical
mastery of that subject is absolutely necessary to the creation of AI, and
yet that is *supposed* to be what PhDs are about. No one's said a word
about learning or knowledge. It's all about the impressiveness of some
letters after your name. I know I'm far from the first person to point out
the massive failure of the educational system, but it remains just as huge
a problem and just as horribly awry. The failure doesn't go away just
because someone has pointed it out before.
To tackle AI I've had to learn, at one time or another, evolutionary
psychology, evolutionary biology, population genetics, game theory,
information theory, Bayesian probability theory, mathematical logic,
functional neuroanatomy, computational neuroscience, anthropology,
computing in single neurons, cognitive psychology, the cognitive psychology
of categories, heuristics and biases, decision theory, visual neurology,
linguistics, linear algebra, physics, category theory, and probably a dozen
other fields I haven't thought of offhand. Sometimes, as with evolutionary
psychology, I know the field in enough depth to write papers in it. Other
times I know only the absolute barest embarassingly simple basics, as with
category theory, which I picked up less than a month ago because I needed
to read other papers written in the language of category theory. But the
point is that in academia, where crossbreeding two or three fields is
considered daring and interdisciplinary, and where people have to achieve
supreme depth in a single field in order to publish in its journals, that
kind of broad background is pretty rare.
I'm a competent computer programmer with strong C++, Java, and Python, and
I can read a dozen other programming languages.
I accumulated all that (except category theory) before I was twenty-five
years old, which is still young enough to have revolutionary ideas.
That's another thing academia doesn't do very well. By the time people
finish a Ph.D. in *one* field, they might be thirty years old, past their
annus mirabilis years. To do AI you need a dozen backgrounds and you need
them when you're young. Small wonder academia hasn't had much luck on AI.
Academia places an enormous mountain of unnecessary inconveniences and
little drains of time in the way of learning and getting the job done. Do
your homework, teach your classes, publish or perish, compose grant
proposals, write project reviews, suck up to the faculty... I'm not saying
it's all useless. Someone has to teach classes. But it is not absolutely
necessary to solving the problem of Friendly AI.
Nearly all academics are untrained in the way of rationality. Not
surprising; few academics are fifth-dan black belts and there are a lot
more fifth-dan black belts than fifth-dan rationalists. But if I were in
academia I would be subject to the authority of those who were not Bayesian
Masters. In the art of rationality, one seeks to attain the perception
that most of the things that appear to be reasons and arguments are not
Bayesian. Eliminate the distractions, silence the roar of cognitive noise,
and you can finally see the small plain trails of genuine evidence. One of
my academic friends once asked me to look at a paper on decision theory;
the paper described the conventional theory, presented a problem, and then
proposed several different individual patches to the conventional theory
and analyzed the patches individually, concluding that none of the
solutions were satisfactory. I replied by arguing that the conventional
theory actually contained *two* independent foundational errors, which
needed to be simultaneously refactored to solve the problem, and in fact,
he needed to look at this whole problem a different way. And the one said:
But I have to take the traditional theory as a point of departure and
then present changes to it, because that's what the reviewers will expect.
And I said: Okay, but for myself I don't have to give a damn about
reviewers, and so I plan to go on using the solution with two simultaneous
corrections. That bias against two simultaneous changes, owing to the need
to take the conventional theory as a point of departure, was justified as
necessary by pointing to social forces instead of Bayesian forces. That
makes it a distraction.
I refuse to accept that entire class of distractions. As an independent
scholar, I never have to give any reason for saying something or thinking
something that points to social forces instead of the facts of the matter.
I have the freedom to do the right thing, without the faintest bias
toward the academically acceptable thing except insofar as the academically
acceptable thing happens to be right. I have the luxury of giving no more
credence to an idea than the weight of Bayesian evidence calls for, even if
the idea has become fixed in academia through any of the non-Bayesian
processes that prevail there.
Now, most of the time, I don't second-guess academia - certainly not in
established fields with great weights of evidence, nor after learning just
the basics of something. Like I said, it's dangerous to be half a
rationalist; if you learn the skill of challenging conventional ideas,
you'd damn well better learn the skill of accepting conventional ideas, or
end up worse off than before. But sometimes, on the fringes, AI for
example, people just make stuff up that sounds cool, and it becomes fixed
because everyone repeats it. Look at Freudian analysis: not one scrap of
experimental evidence. It was a major academic field, with peer-reviewed
journals and everything, but not the faintest hint of science. If that's
the academic standard then academia's standards are too damn low. Or
sometimes the people in one field don't know about the results in another
field, and they say things that are silly and get past the reviewers,
because the people who could catch the mistake work in a different building
of the college campus. That likewise happens, *a lot*, in AI.
It seems to me that the secret of effectiveness is refusing to be
distracted. At one point in my life I did permit myself to be
distracted... by writing freelance programs, by planning to start a
company... eventually I noticed that the only projects in my life that had
ever done the slightest bit of good were the ones that were *directly* on
track to the Singularity. *Not* the distraction projects that I thought
would provide resources or whatever, but the projects that were directly
part of the critical path to AI. In 1998 I took one month out of my
all-important plots to accumulate Singularity resources to write "Coding a
Transhuman AI", and in the end CaTAI that was the only thing I did that
year that actually mattered. And that was very much the story of my life,
until the day I finally snapped and decided to concentrate solely on the
Singularity. Today I refuse to be distracted. Not by academia, not by
technology companies, not by anything. All I ask of myself is that I do
this one thing, solve this one challenge of Friendly AI.
> 2. If lacking a PhD really becomes a problem... well, why not get
> one? PhD students get living stipends and support for their research.
> So even being a PhD student may well put Eliezer in a better position
> to pursue his research than the current scenario allows. Further, if
> he (or anyone else involved) doesn't like the idea of being forced to
> take 2 years of coursework for the PhD, he could always pursue the PhD
> outside of the U.S., where PhDs are pure research degrees with no
> course requirements.
>
> 3. There are a couple of schools (e.g., The University of Technology,
> Sydney, in Australasia) that award PhDs by prior publication. After
> applying, you put together a portfolio of your research, and write an
> overarching paper that illustrates your contribution to the field of
> study, and if deemed PhD-level, you are granted a PhD. I've come
> across at least a few professors in the UK and elsewhere who have
> received their doctorates in this manner. (I've also seen quite a few
> professors with just Master's, but this falls back to point 2 above).
>
> Of course, most people aren't aware of some or all of the above
I wasn't aware. Thanks. If I don't need to spend eight years, that does
shift the cost/benefit ratio. But not far enough, I'm afraid.
-- Eliezer S. Yudkowsky http://intelligence.org/ Research Fellow, Singularity Institute for Artificial Intelligence
This archive was generated by hypermail 2.1.5 : Wed Jul 17 2013 - 04:00:49 MDT