[LINK] ChatGPT

Tom Worthington tom.worthington at tomw.net.au
Sun Jan 29 10:47:53 AEDT 2023


On 27/1/23 11:01, David wrote:

> ... I think most people are pretty quick to detect when they're 
> talking to a machine ...

ChatGPT is text based, making it much harder to tell your are 
communicating with a machine. With audio, it is easier to
tell its is a machine. But if it provides a cheap and convenient 
service, do you mind?

Recently I watched someone put dinner in the oven and start to walk out 
of the kitchen. I thought it odd they did not set the oven timer. But as 
they walked they said "Alexa, set an alarm for 30 minutes". Alexa's 
response was far from human sounding, but would you be willing to pay 
for a human butler to tell you when dinner was ready? Also with the 
"Uncanny valley" we don't want our machines to be almost like people.

> We also have to consider how much responsibility and authority an AI 
> system carries. ...

Organisations are run by rules. The teacher, judge, or police officer,
has only limited discretion. Most of the time they are running through a
complex rules base, much as AI does. The human has more flexibility, but 
are frequently using out of date rules, forgetting some, or acting on 
conscious, or unconscious, bias.

> Does the machine which allows a student an extension of time for 
> their end-of-semester submission ...

There is a danger in using AI to treat the symptoms of a problem, rather 
than the underlying cause. Applications for student extensions is an 
example. Rather than automate the process, is it better to improve the 
teaching and assessment design, so students rarely have to ask.

Teachers use "scaffolding", with the student doing an assignment a piece 
at a time. Those who are struggling can be identified, and provided with 
help, long before the end of semester. 
https://blog.highereducationwhisperer.com/2013/09/how-what-and-when-of-improving-student.html

At an AI workshop a few years I learned to build a TutorBot to respond 
to student requests for extensions. This used IBM Watson, to interpret 
what the student was asking. But whatever they asked, my Bot answered 
"No!". ;-)
https://blog.highereducationwhisperer.com/2018/12/chatbot-tutors-for-blended-learning.html

> ... explain their judgements to some human who ultimately carries
> the can?  Or will they not be given power to make those judgements?

Judging by the evidence given to the RoboDebt inquiry, AI would do a
better job of explaining its decisions than humans. AI would say: "The 
government wanted to get the support of rich people, by persecuting poor 
people, so that is what we did."

AI could be used to patiently explain the reasons for a decision. Of 
course the client should be able to appeal to a human, but just 
explaining why a decision was made would help in a lot of cases.

> It seems to me there's rather a divergence in our social licensing 
> here. ...

I am happy to have self-driving cars, when they are safer than human
drivers. Even my decade old car has automated systems which override
my inputs if it is going to skid, or not stop quickly. The car can do 
this better than I can.

> But I wouldn't like to try telling a bank manager they're personally
> responsible for the autonomous decisions of some AI system.

Have a look at the evidence to previous royal commissions into the
financial sector: they stole money from dead people. Could AI do worse?

More seriously, how often does a bank manager make a decision, based 
purely on their own judgement? The bank manager applies a set of rules, 
or just enters the details onto a system which applies the rules. Also, 
when is the last time you talked to a bank manger, for me it was about 
40 years ago.


-- 
Tom Worthington http://www.tomw.net.au


More information about the Link mailing list