<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8"></head><body style="word-wrap: break-word; -webkit-nbsp-mode: space; line-break: after-white-space;" class="">I imagine playing chess, or go, with a computer. As I play I have a very enlivening experience of playing. The computer seems to have no such thing. For me, in my engagement, “Every neuron is listening to the mutter of the crowd.” Jerry Lettvin, MIT. If The computer goes on to win it has nothing like the experience of winning. it just stops. I can’t imagine a computer saying, except by playing a pre recorded sound file, “that is disgusting.”<div class=""><br class=""></div><div class=""><br class=""><div><br class=""><blockquote type="cite" class=""><div class="">On Jul 27, 2020, at 8:12 PM, Frank Wimberly <<a href="mailto:wimberly3@gmail.com" class="">wimberly3@gmail.com</a>> wrote:</div><br class="Apple-interchange-newline"><div class=""><div dir="auto" class="">Smart computer (program). Nick, is the behaviorist position that if it behaves as though it thinks than it thinks?<br class=""><br class=""><div data-smartmail="gmail_signature" class="">---<br class="">Frank C. Wimberly<br class="">140 Calle Ojo Feliz, <br class="">Santa Fe, NM 87505<br class=""><br class="">505 670-9918<br class="">Santa Fe, NM</div></div><br class=""><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Mon, Jul 27, 2020, 8:28 PM Alexander Rasmus <<a href="mailto:alex.m.rasmus@gmail.com" class="">alex.m.rasmus@gmail.com</a>> wrote:<br class=""></div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div dir="ltr" class=""><div class="">There's also a lot of stuff that isn't poetry. He generates Dwarf Fortress patch notes in the Misc section. <br class=""></div><div class=""><br class=""></div><div class="">There's even a section where GPT-3 argues that it doesn't really understand anything (bold at the start is a shortened version of the prompt):</div><div class="">"<strong class="">AI will never be able to truly</strong>
<strong class=""></strong> think for themselves, to have consciousness, to feel
any of the richness and complexity of the world that we mere humans can
feel. Mere, unenlightened humans might be impressed by the abilities of
simple deep learning programs, but when looked at in a more holistic
manner, it all adds up to… well, nothing. They still don’t exhibit any
trace of consciousness. All of the available data support the notion
that humans feel and experience the world differently than computers do.
While a computer can beat a human master at chess or Go or some other
game of structured rules, it will never be able to truly think outside
of those rules, it will never be able to come up with its own new
strategies on the fly, it will never be able to feel, to react, the way a
human can. Artificial intelligence programs lack consciousness and
self-awareness. They will never be able to have a sense of humor. They
will never be able to appreciate art, or beauty, or love. They will
never feel lonely. They will never have empathy for other people, for
animals, for the environment. They will never enjoy music or fall in
love, or cry at the drop of a hat. Merely by existing, mere,
unenlightened humans are intellectually superior to computers, no matter
how good our computers get at winning games like Go or Jeopardy. We
don’t live by the rules of those games. Our minds are much, much bigger
than that.</div><div class=""><br class=""></div><div class="">
Compared to the awesome power of the human intellect, it’s obvious that
computers are just dense electrochemical machines with intricate wiring.
And while one day we might develop smarter (though still soulless)
computers that surpass the human intellect in the games mentioned above,
it will just end up being some more sophisticated way of doing the same
thing. Even if it’s more efficient than our brains at doing complex
mathematical operations, it will remain entirely in the dark about what
it all means. Mindless processing power alone is not the answer to our
existential questions. So, in the end, we don’t need to create another <span class="">HAL</span>
9000, or Skynet, or some evil global intelligent network of computers
that rule the world and try to exterminate humanity as we become
obsolete, which is a pretty cool subplot for a future Terminator movie.
We don’t need any of that. We’re humans, and we are already
intellectually superior to machines. End of story.</div><div class=""><br class=""></div><div class="">
To be sure, I’m not denying that there are hard problems in AI. Of
course, there are. All of these so called “advances” in this field are a
bunch of crap if we still don’t have a program that can putt around on a
miniaturized golf course all by itself, let alone actually play a
decent game of golf like the humans do, without going into meltdown if
it misses the hole five out of six times in a row. Can we get a movie
about that? If you ask me, we don’t even really know how to define
intelligence yet. How does it work, exactly? It’s a mess, really. If
we’re ever going to create another actual <span class="">HAL</span>
9000, or yet another Skynet (what fun), we’d better be all on the same
page when it comes to AI: what intelligence is, how it works, what
consciousness is, what it feels like, what it really means to be
self-aware. Without that common framework, trying to program yet another
AI that can play yet another game like Go is like trying to blow up
another Death Star with yet another way-too-large superlaser.</div><div class=""><br class=""></div><div class="">
I think one of the big mistakes that computer scientists are making is
that they are conflating intelligence with problem-solving. They’ve
gotten into this habit of creating intricate Turing test competitions:
give the computer a series of math problems, a chess board, etc., etc.,
give it a chat interface so you can interact with it like you would with
another human being, and then see if the machine can fool you into
thinking that it is a human. Once it does this, computers will have
passed the Turing test and achieved general AI. Really? Is that really
the way it works? I don’t see how. A computer has succeeded in faking it
until it makes it, in terms of passing a Turing test competition, only
if it has satisfied some pre-specified set of conditions that we know to
be what a human would do in the same situation. But that is no
guarantee that it has actually achieved intelligence! For all we know,
computers can imitate humans until they generate the most plausible
patterns of thought and behavior we know of, while all along remaining
as soulless as ever. Who’s to say that the computer doesn’t merely use
its programming to cheat the test? Who’s to say that it isn’t just
shuffling its data around in an effort to do the most computations
possible with the least amount of effort? It may succeed in conning us
into thinking that it is self-aware, but that doesn’t prove that it
actually is. It hasn’t actually passed the Turing test, unless we have
defined it in a way that pre-determines the outcome: i.e., if the human
pretends to be a computer, then it passes the test, but if the computer
pretends to be a human, then it doesn’t pass the test! To me, that just
doesn’t sound all that scientific."</div><div class=""><br class=""></div><div class="">Best,</div><div class="">Rasmus<br class=""></div></div><br class=""><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Mon, Jul 27, 2020 at 8:04 PM glen <<a href="mailto:gepropella@gmail.com" target="_blank" rel="noreferrer" class="">gepropella@gmail.com</a>> wrote:<br class=""></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">Excellent. Thanks! I'd seen the link to Gwern from Slate Star Codex. But I loathe poetry. Now that you've recommended it, I have no choice. 8^)<br class="">
<br class="">
On July 27, 2020 6:32:15 PM PDT, Alexander Rasmus <<a href="mailto:alex.m.rasmus@gmail.com" target="_blank" rel="noreferrer" class="">alex.m.rasmus@gmail.com</a>> wrote:<br class="">
>Glen,<br class="">
><br class="">
>Gwern has an extensive post on GPT-3 poetry experimentation here:<br class="">
><a href="https://www.gwern.net/GPT-3" rel="noreferrer noreferrer" target="_blank" class="">https://www.gwern.net/GPT-3</a><br class="">
><br class="">
>I strongly recommend the section on the Cyberiad, where GPT-3 stands in<br class="">
>for<br class="">
>Trurl's Electronic Bard:<br class="">
><a href="https://www.gwern.net/GPT-3#stanislaw-lems-cyberiad" rel="noreferrer noreferrer" target="_blank" class="">https://www.gwern.net/GPT-3#stanislaw-lems-cyberiad</a><br class="">
><br class="">
>There's some discussion of fine tuning input, but I think more cases<br class="">
>where<br class="">
>they keep the prompt fixed and show several different outputs.<br class="">
<br class="">
-- <br class="">
glen<br class="">
<br class="">
- .... . -..-. . -. -.. -..-. .. ... -..-. .... . .-. .<br class="">
FRIAM Applied Complexity Group listserv<br class="">
Zoom Fridays 9:30a-12p Mtn GMT-6 <a href="http://bit.ly/virtualfriam" rel="noreferrer noreferrer" target="_blank" class="">bit.ly/virtualfriam</a><br class="">
un/subscribe <a href="http://redfish.com/mailman/listinfo/friam_redfish.com" rel="noreferrer noreferrer" target="_blank" class="">http://redfish.com/mailman/listinfo/friam_redfish.com</a><br class="">
archives: <a href="http://friam.471366.n2.nabble.com/" rel="noreferrer noreferrer" target="_blank" class="">http://friam.471366.n2.nabble.com/</a><br class="">
FRIAM-COMIC <a href="http://friam-comic.blogspot.com/" rel="noreferrer noreferrer" target="_blank" class="">http://friam-comic.blogspot.com/</a> <br class="">
</blockquote></div>
- .... . -..-. . -. -.. -..-. .. ... -..-. .... . .-. .<br class="">
FRIAM Applied Complexity Group listserv<br class="">
Zoom Fridays 9:30a-12p Mtn GMT-6 <a href="http://bit.ly/virtualfriam" rel="noreferrer noreferrer" target="_blank" class="">bit.ly/virtualfriam</a><br class="">
un/subscribe <a href="http://redfish.com/mailman/listinfo/friam_redfish.com" rel="noreferrer noreferrer" target="_blank" class="">http://redfish.com/mailman/listinfo/friam_redfish.com</a><br class="">
archives: <a href="http://friam.471366.n2.nabble.com/" rel="noreferrer noreferrer" target="_blank" class="">http://friam.471366.n2.nabble.com/</a><br class="">
FRIAM-COMIC <a href="http://friam-comic.blogspot.com/" rel="noreferrer noreferrer" target="_blank" class="">http://friam-comic.blogspot.com/</a> <br class="">
</blockquote></div>
- .... . -..-. . -. -.. -..-. .. ... -..-. .... . .-. .<br class="">FRIAM Applied Complexity Group listserv<br class="">Zoom Fridays 9:30a-12p Mtn GMT-6 <a href="http://bit.ly/virtualfriam" class="">bit.ly/virtualfriam</a><br class="">un/subscribe <a href="http://redfish.com/mailman/listinfo/friam_redfish.com" class="">http://redfish.com/mailman/listinfo/friam_redfish.com</a><br class="">archives: <a href="http://friam.471366.n2.nabble.com/" class="">http://friam.471366.n2.nabble.com/</a><br class="">FRIAM-COMIC <a href="http://friam-comic.blogspot.com/" class="">http://friam-comic.blogspot.com/</a> <br class=""></div></blockquote></div><br class=""></div></body></html>