[Xmca-l] Re: Interesting article on robots and social learning
Andy Blunden
andyb@marxists.org
Sat Jul 14 05:58:27 PDT 2018
I understand that the Turing Test is one which AI people can
use to measure the success of their AI - if you can't tell
the difference between a computer and a human interaction
then the computer has passed the Turing test. I tend to rely
on a kind of anti-Turing Test, that is, that if you can tell
the difference between the computer and the human
interaction, then you have passed the anti-Turing test, that
is, you know something about humans.
Andy
------------------------------------------------------------
Andy Blunden
http://www.ethicalpolitics.org/ablunden/index.htm
On 14/07/2018 1:12 PM, Douglas Williams wrote:
> Hi--
>
> I think I'll come out of lurking for this one. Actually,
> what you're talking about with this pain algorithm system
> sounds like a modeling system that someone might need to
> develop what Alan Turing described as a P-type computing
> device. A P-type computer would receive its programming
> from inputs of pleasure and pain. It was probably derived
> from reading some of the behavioralist models of mind at
> the time. Turing thought that he was probably pretty close
> to being able to develop such a computing device, which,
> because its input was similar, could model human thought.
> The Eliza Rogersian analysis computer program was another
> early idea in which the goal was to model the patterns of
> human interaction, and gradually approach closer to human
> thought and interaction that way. And by the 2000's, the
> idea of the "singularity" was afloat, in which one could
> model human minds so well as to enable a human to be
> uploaded into a computer, and live forever as software
> (Kurzweil, 2005). But given that we barely had a
> sufficient model of mind to say Boo with at the time (what
> is consciousness? where does intention come from? What is
> the balance of nature/nurture in motivation? Speech
> utterances? and so on), and you're right, AI doesn't have
> much of a theory of emotion, either--the goal of computer
> software modeling human thought seemed very far away to me.
>
> At someone's request, I wrote a rather whimsical paper
> called "What is Artificial Intelligence?" back in 2006
> about such things. My argument was that statistical
> modeling of human interaction and capturing thought was
> not too easy after all, precisely because of the parts of
> mind we don't think of, and the social interactions that,
> at the time, were not a primary focus. I mused about that
> in the context of my trying to write a computer program by
> applying Chomsky's syntactic structures to interpret
> intention of a few simple questions--without, alas, in my
> case, a corpus-supported Markov chain logic to do it.
> Generative grammar would take care of it, right?Wrong.
>
> So as someone who had done a little primitive, incompetent
> attempt at speech modeling myself, and in the light of my
> later-acquired knowledge of CHAT, Burke, Bakhtin, Mead,
> and various other people in different fields, and of the
> tendency of people to interact through the world through
> cognitive biases, complexes, and embodied perceptions that
> were not readily available to artificial systems, I didn't
> think the singularity was so near.
>
> The terrible thing about computer programs is that they do
> just what you tell them to do, and no more. They have no
> drive to improve, except as programmed. When they do
> improve, their creativity is limited. And the approach now
> still substantially is pattern-recognition based. The
> current paradigm is something called Convolutional Neural
> Network Long Short-Term Memory Networks (CNN/LSTM) for
> speech recognition, in which the convolutional neural
> networks reduce the variants of speech input into
> manageable patterns, and temporal processing (temporal
> patterns of the real wold phenomena to which the AI system
> is responding). But while such systems combined with
> natural language processing can increasingly mimic human
> response, and "learn" on their own, and while they are
> approaching the "weak" form of artificial general
> intelligence (AGI), the intelligence needed for a machine
> to perform any intellectual task that a human being can,
> they are an awfully long way from "strong" AGI--that is,
> something approaching human consciousness. I think that's
> because they are a long way from capturing the kind of
> social embeddedness of almost all animal behavior, and the
> sense in which human cognition is embedded in the messy
> things, like emotion. A computer algorithm can recognize
> the patterns of emotion, but that's it. An AGI system that
> can experience emotions, or have motivation, is quite
> another thing entirely.
>
> I can tell you that AI confidence is still there. In
> raising questions about cultural and physical embodiment
> in artficial intelligence interations with someone in the
> field recently, he dismissed the idea as being that
> relevant. His thought was that "what I find essential is
> that we acknowledge that there's no obvious evidence
> supporting that the current paradigm of CNN/LSTM under
> various reinforcement algorithms isn't enough for A AGI
> and in particular for broad animal-like intelligence like
> that of ravens and dogs."
>
> But ravens and dogs are embedded in social interaction, in
> intentionality, in consciousness--qualitatively different
> than ours, maybe, but there. Dogs don't do what you ask
> them to, always. When they do things, they do them for
> their own intentionality, which may be to please you, or
> may be to do something you never asked the dog to do,
> which is either inherent in its nature, or an expression
> of social interactions with you or others, many of which
> you and they may not be consciously aware of. The deep
> structure of metaphor, the spatiotemporal relations of
> language that Langacker describes as being necessary for
> construal, the worlds of narrativized experience, are
> mostly outside of the reckoning, so far as I know (though
> I'm not an expert--I could be at least partly wrong) of
> the current CNN/LSTM paradigm.
>
> My old interlocutor in thinking about my language program,
> Noam Chomsky, has been a pretty sharp critic of the
> pattern recognition approach to artificial intelligence.
>
> Here's Chomsky's take on the idea:
>
> http://languagelog.ldc.upenn.edu/myl/PinkerChomskyMIT.html
>
> And here's Peter Norvig's response; he's a director of
> research at Google, where Kurzweil is, and where, I
> assume, they are as close to the strong version of
> artificial general intelligence as anyone out there...
>
> http://norvig.com/chomsky.html
>
> Frankly, I would be quite interested in what you think of
> these things. I'm merely an Isaiah Berlin fox, chasing to
> and fro at all the pretty ideas out there. But you, many
> of you, are, I suspect, the untapped hedgehogs whose ideas
> on these things would see more readily what I dimly grasp
> must be required, not just for achieving a strong AGI, but
> for achieving something that we would see as an ethical,
> reasonable artificial mind that expands human experience,
> rather than becomes a prison that reduces human
> interactions to its own level.
>
> My own thinking is that lately, Cognitive Metaphor Theory
> (CMT), which I knew more of in its earlier (now "standard
> model') days, is getting even more interesting than it
> was. I'd done a transfer term to UC Berkeley to study with
> George Lakoff, but we didn't hit it off well, perhaps I
> kept asking him questions about social embeddedness, and
> similarities to Vygotsky's theory of complex thought, and
> was too expressive about my interest in linking out from
> his approach than folding in. It seems that the idea I was
> rather woolily suggesting to Lakoff back then has caught
> on: namely, that utterances could be explored for cultural
> variation and historical embeddedness, a form ofsocial
> context to the narratives and metaphors and blended spaces
> that underlay speech utterances and thought; that there
> was a degree of social embodiment as well as physiological
> embodiment through which language operated. I thought
> then, and it looks like some other people now, are
> thinking that someone seeking to understand utterances (as
> a strong AGI system would need to do) really, would need
> to engage in internalizing and ventriloqusing a form of
> Geertz's thick description of interactions. In such forms,
> words do not mean what they say, and can have different
> affect that is a bit more complex than I think temporal
> processing currently addresses.
>
> I think these are the kind of things that artificial
> intelligence would need truly to advance, and that Bakhtin
> and Vygotsky and Leont'ev and in the visual world,
> Eisenstein were addressing all along...
>
> And, of course, you guys.
>
> Regards,
> Douglas Willams
>
>
>
> On Tuesday, July 3, 2018, 10:35:45 AM PDT, David H
> Kirshner <dkirsh@lsu.edu> wrote:
>
>
> The other side of the coin is that ineffable human
> experience is becoming more effable.
>
> Computers can now look at a human brain scan and determine
> the degree of subjectively experienced pain:
>
>
>
> In 2013, Tor Wager, a neuroscientist at the University of
> Colorado, Boulder, took the logical next step by creating
> an algorithm that could recognize pain’s distinctive
> patterns; today, it can pick out brains in pain with more
> than ninety-five-per-cent accuracy. When the algorithm is
> asked to sort activation maps by apparent intensity, its
> ranking matches participants’ subjective pain ratings. By
> analyzing neural activity, it can tell not just whether
> someone is in pain but also how intense the experience is.
>
>
>
> So, perhaps the computer can’t “feel our pain,” but it can
> sure “sense our pain!”
>
>
>
> Here’s the full article:
>
> https://www.newyorker.com/magazine/2018/07/02/the-neuroscience-of-pain
>
>
>
> David
>
>
>
> *From:*xmca-l-bounces@mailman.ucsd.edu
> <xmca-l-bounces@mailman.ucsd.edu> *On Behalf Of *Glassman,
> Michael
> *Sent:* Tuesday, July 3, 2018 8:16 AM
> *To:* eXtended Mind, Culture, Activity
> <xmca-l@mailman.ucsd.edu>
> *Subject:* [Xmca-l] Re: Interesting article on robots and
> social learning
>
>
>
> / /
>
>
>
> It seems like we are still having the same argument as
> when robots first came on the scene. In response to John
> McCarthy, who was claiming that eventually robots can have
> belief systems and motivations similar to humans through
> AI John Searle wrote the Chinese room. There have been a
> lot of responses to the Chinese room over the years and a
> number of digital philosopher claim it is no longer
> salient, but I don’t think anybody has ever effectively
> answered his central question.
>
>
>
> Just a quick recap. You come to a closed door and know
> there is a person on the other side. To communicate you
> decide the teacher the person on the other side Chinese.
> You do this by continuously exchanging rules systems under
> the door. After a while you are able to have a
> conversation with the individual in perfect Chinese. But
> does that person actually know Chinese just from the rule
> systems. I think Searle’s major point is are you really
> learning if you don’t know why you’re learning, or are you
> just repeating. Learning is embedded in the human
> condition and the reason it works so well and is adaptable
> is because we understand it when we use what we learn in
> the world in response to others. To put it in response to
> the post, does a bomb defusion robot really learn how to
> defuse a bomb if it does not know why it is doing it. It
> might cut the right wires at the right time but it doesn’t
> understand why and therefore is not doing the task just a
> series of steps it has been able to absorb. Is that the
> opposite of human learning?
>
>
>
> What the researcher did really isn’t that special at this
> point. Well I definitely couldn’t do it and it is
> amazing, but it is in essence a miniature version of
> Libratus (which beat experts at Texas Hold em) and Alphago
> (which beat the second best Go player in the world). My
> guess it is the same use of deep learning in which the
> program integrates new information into what it is already
> capable of. If machines can learn from interacting with
> other humans then they can learn from interacting with
> other machines. It is the same principle (though much,
> much simpler in this case). The question is what does it
> mean. As we defining learning down because of the
> zeitgeist. Greg started his post saying a socio-cultural
> theorist be interested in this research. I wonder if they
> might more likely to be the ones putting on the brakes,
> asking questions about it.
>
>
>
> Michael
>
>
>
> *From:*xmca-l-bounces@mailman.ucsd.edu
> <mailto:xmca-l-bounces@mailman.ucsd.edu>
> <xmca-l-bounces@mailman.ucsd.edu
> <mailto:xmca-l-bounces@mailman.ucsd.edu>> *On Behalf Of
> *Andy Blunden
> *Sent:* Tuesday, July 03, 2018 7:04 AM
> *To:* xmca-l@mailman.ucsd.edu <mailto:xmca-l@mailman.ucsd.edu>
> *Subject:* [Xmca-l] Re: Interesting article on robots and
> social learning
>
>
>
> Does a robot have "motivation"?
>
> andy
>
> ------------------------------------------------------------
>
> Andy Blunden
> http://www.ethicalpolitics.org/ablunden/index.htm
>
> On 3/07/2018 5:28 PM, Rod Parker-Rees wrote:
>
> Hi Greg,
>
>
>
> What is most interesting to me about the understanding
> of learning which informs most AI projects is that it
> seems to assume that affect is irrelevant. The role of
> caring, liking, worrying etc. in social learning seems
> to be almost universally overlooked because
> information is seen as something that can be ‘got’ and
> ‘given’ more than something that is distributed in
> relationships.
>
>
>
> Does anyone know about any AI projects which consider
> how machines might feel about what they learn?
>
>
>
> All the best,
>
>
> Rod
>
>
>
> *From:*xmca-l-bounces@mailman.ucsd.edu
> <mailto:xmca-l-bounces@mailman.ucsd.edu>
> <xmca-l-bounces@mailman.ucsd.edu>
> <mailto:xmca-l-bounces@mailman.ucsd.edu> *On Behalf Of
> *Greg Thompson
> *Sent:* 03 July 2018 02:50
> *To:* eXtended Mind, Culture, Activity
> <xmca-l@mailman.ucsd.edu> <mailto:xmca-l@mailman.ucsd.edu>
> *Subject:* [Xmca-l] Interesting article on robots and
> social learning
>
>
>
> I’m ambivalent about this project but I suspect that
> some young CHAT scholar out there could have a lot to
> contribute to a project like this one:
>
> https://www.sapiens.org/column/machinations/artificial-intelligence-culture/
>
>
>
> -Greg
>
> --
>
> Gregory A. Thompson, Ph.D.
>
> Assistant Professor
>
> Department of Anthropology
>
> 880 Spencer W. Kimball Tower
>
> Brigham Young University
>
> Provo, UT 84602
>
> WEBSITE: greg.a.thompson.byu.edu
> <http://greg.a.thompson.byu.edu>
> http://byu.academia.edu/GregoryThompson
>
> ------------------------------------------------------------
>
> Image removed by sender.
> <http://www.plymouth.ac.uk/worldclass>
>
> This email and any files with it are confidential and
> intended solely for the use of the recipient to whom
> it is addressed. If you are not the intended recipient
> then copying, distribution or other use of the
> information contained is strictly prohibited and you
> should not rely on it. If you have received this email
> in error please let the sender know immediately and
> delete it from your system(s). Internet emails are not
> necessarily secure. While we take every care,
> University of Plymouth accepts no responsibility for
> viruses and it is your responsibility to scan emails
> and their attachments. University of Plymouth does not
> accept responsibility for any changes made after it
> was sent. Nothing in this email or its attachments
> constitutes an order for goods or services unless
> accompanied by an official order form.
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.ucsd.edu/pipermail/xmca-l/attachments/20180714/81c74f8a/attachment.html
More information about the xmca-l
mailing list