The Academics That Think ChatGPT Is BS

Published Jul 19, 2024, 4:00 AM

In a paper released earlier this year, three academics from the University of Glasgow classified ChatGPT's outputs not as "lies," but as "BS" - as defined by philosopher Harry G. Frankfurt in "on BS" (and yes I'm censoring that) - and created one of the most enjoyable and prescient papers ever written. In this episode, Ed Zitron is joined by academics Michael Townsen Hicks, James Humphries and Joe Slater in a free-wheeling conversation about ChatGPT's mediocrity - and how it's not built to represent the world at all.

LINKS:

Paper: https://link.springer.com/article/10.1007/s10676-024-09775-5

Michael Townsen Hicks: https://www.townsenhicks.com/

Joe Slater: https://www.gla.ac.uk/schools/humanities/staff/joeslater/

Newsletter: wheresyoured.at

Reddit: http://www.reddit.com/r/betteroffline

Discord: chat.wheresyoured.at

Ed's Socials - http://www.twitter.com/edzitron instagram.com/edzitron https://bsky.app/profile/zitron.bsky.social https://www.threads.net/@edzitron

All Zone Media.

Hello and welcome to Better Offline. I'm your host ed z trun. In early June, three researchers from the University of Glasgow published a paper in the Ethics of Information and Technology Journal called chat GBT is Bullshit. And I just want to be clear. This is a great and thoroughly researched and well argued paper. This is not silly at all. It's actually great academia. And today I'm joined by the men who wrote it, academics Michael Townsend Hicks, James Humphries and Joe Slater, to talk about chat gbt's mediocrity and how it's not really built to represent the world at all. So, for the sake of argument, could you define bullshit for me?

So you are bullshitting if you are speaking without caring about the truth of what you say. So normally, if I'm telling you staff about the world in a good case, I'll be telling you something that's true, and like trying to tell you something that's true. If I'm lying to you, I'll be knowingly telling you something that's false or something I think is false. I'm bullshitting. I just don't care. I'm trying to get you to believe me. I don't really care about whether what I say is true. I might not have any particular view on whether it's true.

Or not right. And you define between like soft and hard bullshit? Can you also get into that as well? Can you also identify yourselves as well to.

R Yeah, yeah, and Joe. So the soft bullshit hard bullshit distinction is a very serious and technical distinction, right, So he came up with.

This because bullshit is in the technical philosophical sense. Comes from Harry Frankfurt, recently diseased but a really great philosopher.

And he talks about the about.

Bullshit that there is in a popular culture and just in general discourse these days. Some of the ways he talks about bullshit seemed to suggest that it needs to be accompanied by a sort of malign intention.

I'm doing something kind of bad.

I'm intending to mislead you about the enterprise of what I'm doing.

Yeah, maybe about who you are or what you know.

So you may be trying to portray yourself as someone who is knowledgeable about a particular subject. Maybe you're a student who showed up to class without doing the work. Maybe you're trying to portray yourself as someone who's virtuous in ways you're not. Maybe you're a politician who wants to seem like you care about your constituents, but actually you don't. So you're not trying to mislead somebody about what you're saying the content of your utterance. You're trying to mislead them instead about like why you're saying it's that's what we call hard bullshit.

Yeah, it's one of the things Frankfort talked about.

Yeah, so FRANKFOT doesn't make this hard bullshit soft bullshit the station, but we do because sometimes it seems like Frank for has this particular kind of intention in mind, but sometimes he's just a bit looser with it. And we want to say that chat GPT and other large language models they don't really have this intention to deceive because they're not people that don't have these intentions. They're not trying to mess with us in that kind of way, but they do lack this kind of caring about truth.

Well, I'm James, by the way, I suppose we strictly don't want to say that they aren't hard bullshitters, and we just think if you don't think that large language models are sapient. If you don't think they're kind of minds in any important way, then they're not hard bullshitters. So I think in the paper we don't take a position on whether or not they are. We just say if they are, this is the way in which they are. But minimally they're soft bullshitters. So they're kind of soft bullshites, as Joe says, doesn't require that speakers attempting to deceive the audience about the nature of the enterprise.

Hard bullshit does.

So if it turns out that large language models are sapient, which they're definitely not, like that's just technoproperty.

Yeah, that's nonsense.

Yeah, But if they are, then they're bullshitters and minimally their soft bullshitters or their bullshit machines.

So you also make the distinction in there the intention, so the very fabric of hard bullshit that you intentionally are bullshitting to someone, you kind of make this distinction that the intention of the designer and the involved prompting could make this hard bullshit. Because with a lot of these models and someone recently jail broke chat GPT and it listed all of the things that's prompted to do could prompting be considered intentional? That Sam Altman, CEO of open AI, could be intentionally bullshitting? I think it is.

Yeah, this again, I think it's something I don't know what the kind of hive mind consensus on this is. I'm sort of sympathetic to the idea that if you take this kind of purposive or teleological attitude towards right, what an intention is is an effort to do something, then maybe they do have intentions. But again, I think in the paper we just sort of wanted I mean, it's a standard philosophical move, right, just sort of go, look, here's all this controversial stuff as we can make it. Now we can hit you with the really controversial shit that we wanted to get to. So in the paper we sort of deliberately when maybe you might think it has intentions for it this reason, we kind of have no judgment on this efficiently, I'm sympathetic to the sort.

Of view that you're putting.

I think you're kind of sympathetic this as well.

Right, Yeah, so.

I Mike, there are a few ways that you can think of CHATCHYBT as having intentions.

I think, and we talked about a few of them.

One is by thinking of the designers who created it as kind of imbuing it with intentions. So they created it for a purpose and to do a particular task, and that task is to make people think that it's a normal sounding person.

Right.

It's to make people when they have a conversation with it, not be able to distinguish between what it's outputting and what a normal human would say.

Right.

Right, And that kind.

Of goal, if it amounts to an intention, is the kind of intention we think a bullshitter has.

Right.

It's not trying to deceive you about anything it's saying. It doesn't care whether what it's saying is true. That's not part of the goal. But the goal is to do is to make it seem as if it's something that it's not, like specifically a human interlocutor. And one source for that goal is the programmers who designed it. Another is the training method. So it was trained by being given sort of positive and negative feedback in order to achieve.

A specific thing, right, And that specific thing is just sounding normal.

And that's so similar to what our students are doing when they try to pretend that they read something they haven't read.

There is something very collegiate about the way it bullshits though. It reminds me of when I was in college when Penn State and Aporistwyth two very different institutes.

Right, both kind of in the middle of nowhere.

Yeah, both very sad. But the one thing you saw with like students who were doing like B plus homework is they were using words they didn't really understand. They were putting things together in a way that realists, like the intro body conclusion, there was a certain formula behind it, and it's just it feels exactly like it. But that kind of brings me to my next question, which is, how did you decide to write this? What inspired this?

Right?

I would feel this because whenever Mike tells the story, you get sort of sanitized version. We were in the pub whining about student essays. Perfect, yeah, like.

You were not long here, right, you're not long yeah.

Long in post of a bunch of us, or went to the pub on emotional let's let's welcome our new members of staff sort of event, and inevitably with hat about two points we were all pissing in moaning and do you think it might have been.

Neiled that kind of prompted this. No, I don't think he was the.

We talked about it after, right, Yeah, in a case like sort of came up.

We were talking about this sort of prevalence of chat GPT generated stuff and kind of what it said about how at least some students were kind of approaching the assessments, and you know, I forget who. Someone sort of just went off handly, yeah, it was all just Frank thirteen bullshit though, isn't it. And we sort of collectively went hello, because obviously, all having a background in plosphy, we've all read on bullshit. You know, we all went he we get to say bullshit and allegively serious academic work. So the start it was we've had this experience of having to read all of this kind of uncanny Valley stuff, and when prompted, we all went, oh, it really is like Frank thirty and bullshit in a way that we can probably get paper outselof.

Yeah, And at that point I think we were in the department meetings talking about how to deal with chat GPT written papers, and there were discussions going on all over the university, including kind of in high levels and the administration to come up with a policy, and we specifically wanted a stronger policy because our university is very interested in cutting into technology, right, and so they wanted the students to have an experience using and figuring out how to use whatever the freshest technology is.

Which as they should as they should, but not for essays, right, That's the.

Worry we had.

We thought, you know, if we're not very clear about this, the students will be using it in a way that will detract from the educational experience.

Right. And at the same time, it was becoming more.

Widely known how these machines work, like specifically how they're doing next token or next word prediction in order to come up with a digestible string of text. And when you know that that's how you're they're doing it, I mean, it seems so similar to what humans do when they have no idea what they're talking about. And so when we were talking about it just seemed like an obvious paper that someone was going to write, and we thought that a better bs. You know, eventually people are going to see this connection and it's.

A great papa.

I think it worked very well.

I mean, I think that it's the kind of thing where when I pitch this to other philosophers, it doesn't take them very long to just agree to be like, ah, yes.

That's right.

It's an interesting philosophical concept as well, because the way that people look at chat GPT in large language models is very much machine do this. But when you think about it, there are the ways where people are giving it consciousness. I just sort of tweet just now as someone talking about saying please with every request. It's like, no, I will abuse the computer in whatever manner I see fit. But it's it's curious because I think more people need to be having conversations like this. And one particular thing I like that you said, I actually would love you to go into more detail, is you said, the chat GPT in large language buddles, they're not designed to represent the world at all. They're not lying or misrepresenting. They're not designed to do that. What do you mean by that?

I mean kind of as background, I do philosophy of science, but my thoughts about something like CHATCHEPT are largely inspired by the fact that I also teach a class called Understanding Philosophy through Science Fiction.

Oh and and now we like talk.

About whether computers could be conscious and I don't know what you guys think. Actually I think they could, right, I just don't think this one is. And part of the reason I think they could but this one isn't is that I think that in order to sort of represent the world or have the kinds of things we have that are like beliefs, desires, thoughts that are about external things, you have to have internal states that are connected in some way to the external world. Usually causation. We're perceiving things, information is coming in. Then we've got some kind of state in our brain that's designed just to track these things in the external world.

Right.

That's a huge part of our cognitive lives. It is just tracking external world things and it's a very important part of childhood development when you figure out how to interact them.

It's semiotics, right, Daniel John Aberistwyth told me semiotics it's like an reception of the world.

And this is like theory of meaning stuff. So, yeah, semiotics is like theory of science. How is it that assign.

Can be folks, the representation of the thing and the thing itself.

That can happen, yeah, but not always.

Sometimes it's just the representation of the thing and there's a lot of philosophy as at figuring out how brain states or words on a page can be about external world things. And a big part of it, at least from my perspective, has to do with tracking those things, keeping tabs on them, changing as a result of seeing differences in the external world. And chatch ept is not doing any of that, right, That's not what it's designed to do. It's taking in a lot of data once and then using that to respond to text. But it's not remembering individuals, tracking things in the world in any way perceiving things in the world. It's just forming a sort of statistical model of what people say. And that's kind of so divorced from what most thinking beings do.

It's divorced from experience.

Yeah. I mean, as far as I can tell, it doesn't have anything like experience.

Yeah.

And that's one of the things that this I thinks of. In one way it comes down to is that if you sort of post this sufficiently far, someone is going to go, ah, isn't this just biochauvinism, right, Like, aren't you just assuming that unless something runs on like meat, it can't be sentient? And this isn't something we get into into paper part because we didn't really think it was worth addressing. But the sorts of things that seem like they're like never mind consciousness, right, but it seemed to be necessary in order for something to be trying to track the world, or in order to be corresponding the world, or to form beliefs about the world. Chat GPT just doesn't seem to meet any of them. Kind Of, it's if it does turn out that it's sapient, then chat GPT has got some profoundly serious executive function disorders, right, sapient, Right, so we don't have to worry about it. But kind of it's not the case that we've got some blundering proto general intelligence that's trying to figure out how to represent the world. It's not trying to represent the world at all. It's utterances are designed to look as if it's trying to represent the world, and then we just go that that's just bullshit.

This is a classic case of bullshit.

Yeah, it seems to be making stuff up, but making stuff up doesn't even seem to describe it. It's just throwing shit at a wall very accurately, but not accurately enough.

Yeah, it's got various guidelines that allow it to throw shit at the wall where a sort of reasonably high degree of Actually, no, you're right. I mean, one of the things that a human bullshitter could at least be characterized as doing is they'd have to try and kind of judge their audience, right, They'd have to try and make the bullshit plausible to the audience that they're speaking to, and chat GPT can't do that, right. All it can do is go on a statistically large enough model and it looks.

Like Z follows why follows X?

Right, it's not kind of got any does have any consciousness at all, of course, but it doesn't have any sensitivity to the sorts of things that people are in fact likely to find plausible. It just does a kind of brute number crunch. So it's more complicated than that, but I think it boies down effectively to kind of number crunching, right, kind of data and context.

In which probabilistic planning of what planning, it doesn't plan. That's the thing. It's interesting. There are these words you use to describe things that, when you think about it, are not accurate. You can't say it plans or things.

I think one thing.

Between when we came up with the idea and when we finish writing the paper, we spent some time reading about how it works and how it represents language and what the statistical.

Model is like.

And I was maybe more impressed than James about that, because it is like doing things that are similar to what we do when we understand language. It does seem to kind of locate words in a meaning space right right, and connect them to other words, and you know, show similarity and meaning, and it also does seem to be able to in some way understand context. But we don't know how similar that is for a variety of reasons, but mostly because it's too big of a system and we can only kind of probe it. And it's trained indirectly, right, so it's not programmed by individuals. And even though that's kind of a very impressive model of language and meaning and may some ways be similar to what we do when we understand language, we're doing a lot more things like planning, things like tracking things in the world. Just having desires and representing the way you want the world to be and thereby generating goals doesn't seem to be something that it has anything any.

Room for in its architects.

This is something of the time of it's just goods me and you were talking about the kind of in some ways it learns language the same way.

That we do.

I mean, it's got no grasp of expleted in fixation, right, this is one of chumps here.

What does that mean? Just for not for me? I definitely know.

Yeah, yeah, of course if I give you the sentence that's completely crazy, man, and tell you to put the word fucking into that sentence, there's a number of ways in which any language speaker is going to do it that they'll just go, yeah, of course.

That where it goes, right.

But we it seems that we've got a grasp on this incredibly early on in a way that doesn't look like it's the ways at least most of us that taught language, right, we get quite highly told off when we try.

And do expleted in fixation.

Yes, So this I think would be one of those cases where you could do a sort of disnalogy by cases. Right, you present chat GPT a sentence and say insert the word fucking correctly in this sentence.

And I don't think it would be very good at.

It I think it would be.

I think you reckon.

I mean, we probably couldn't tell we could, but we shouldn't.

Yeah.

One of the things that I thought was like kind of interesting about how it works is that it does learn language, probably differently from the way we do, but it does it all by examples, you know, So it's looking at all these pieces of text and thinking, ah, this is okay, that's okay. And one kind of interesting thing about how humans understand language is that we're able to kind of understand meaningless but grammatical sentences. It's not clear to me that chat GPT would understand those. That's another Chopski example. So you know, Chomsky has this example that's like colorless uss sleep furiously color it's colorless green ideas, right, And that's a meaningless sentence, but it's grammatically well formed, and we can understand that it's grammatically well formed, but also that it's meaningless. Because chat Gibt kind of combines different aspects of what philosophers of language, logicians, linguistics people see as like different components of meaning. It sees these as all kind of wrapped up.

In the same thing. It puts them in the same.

Big model I'm not sure it could differentiate between ungrammaticality and meaninglessness.

So we're doing real time science right now.

I just yeah, we should check the words get.

Into the following sentence in the correct way. Man, that's crazy. And I did it six times, and I would say fifty percent of the time it got it right. And it did. Man, that's fucking crazy. Man, that's crazy. Fucking Man, that's fucking crazy. Man that's crazy. Fucking My favorite is man, comma that's crazy, Comma fucking.

To be fair.

Reliable.

You take the commas out of that last one, and you've got a grammatical sentence.

Yeah.

Of course in Glasgow you can also start with fucking man, that's crazy.

West London as well. But the thing is, though it doesn't know what correct means there.

Yeah.

Now, when it's trained on this language, when it's trained on thousands of Internet posts that stole, it's not like it reads them and says, oh, I get this, like I see what they're going for. It just learns structures by looking, which is kind of how we learn language. But it kind of reminds me of like when I was a kid and I'd hear someone say something funny I'd to repeat it, and my dad, who's wonderful, would just say that doesn't make any sense, and you'd have to explain, because if you're learning everything through copying, you're not learning, you're just memorizing.

Yeah, yeah, exactly.

There's a I don't know if you have already talked to somebody about this, but there's a you know, a classic argument from child ski against behaviorism.

Right.

Behaviorism is the view that we learn everything through listen response.

Roughly, that's not exactly.

It, but I'm not a philosopher of mine, so I can get away with with that.

So Chomsky says, look, we don't get enough.

Stimulus to learn language as quickly as we do just through watching.

Other people's behavior and copying it.

We have to have some in built grammatical structures that language is latching onto. And there's been some papers arguing that chatchipt shows Chomsky was wrong because it doesn't have the inbuilt grammatical structure.

But one interesting thing.

Is it requires ten to one hundred times more data than a human child does when learning language. Right, So Chomsky's argument was we don't get enough stimulus, and chat GPT can kind of do it without the structure, but it's not quite doing it as well, and it gets like orders of magnitude more more input than a human does before a human learns language, which is kind of interesting.

You're going to do something as basically as putting the word.

Yeah, yeah.

It doesn't even see and it doesn't have the knowledge to say, request more context because it doesn't perceive context. And that's kind of the interesting thing. So there was another paper out of Oxver I think that's talking about cognition and chat GPT and all this thing, and it's just it doesn't feat chat GPT features in no way any of the things that the human mind is really involved in. It seems it's mostly just not even memorization, because it doesn't memorize. It's just guessing based on a very blood pile of stuff. But this actually does lead me to marvel question, which is, you don't like the term hallucination. Why is that hallucination?

It makes it sound a bit like I'm usually doing something right and I'm looking around seeing the world that's something like what it really is. And then one little bit of the feature for a visual hallucination, one feature of my visual field actually isn't represented in the real world, it's not actually there.

Everything else might well might well be right.

Imagine I hallucinate, there's a red balloon in front of me. I still see Mike, I still see James, I still see the laptop.

One bit is.

Wrong, everything else is right, and I'm doing the same thing that I'm usually doing, Like my eyes are still working pretty much normally, I think, I and this.

Is the way I usually get knowledge about the world.

This is a pretty reliable process for me learning from it right and representing the world in this way. So when talking about hallucinations, this suggests that Chatgypt and other similar things they're going through this process that is usually quite good at representing the world. And then oh, it's made a mistake this one time, but actually no, it's bullshitting the whole time, and like sometimes it gets things right right, bullshitting just like a politician.

Imagine a politician that bullshits all the time, if you.

Could possibly imagine it, Sometimes they might just get some things true, and we should still call them bullshitting because that's what they're doing. And this is what at GPT is doing every time it produces an output. So this is why we think bullshit is a better way of thinking about this. One of the reasons why we think bullshit's better way think about it.

I also kind of think that some of the ways we talk about chat gpt LIN, even when it makes mistakes, lind themselves to overhyping its ability or overestimating its abilities, and talking about it is hallucinating. Is one of these because when you say that it's hallucinating, as you're pointed out, you're giving the idea that it's representing the world in some way and then telling you what the.

Content and it has perception.

Yeah, exactly, like it has perceived something and like, oh no, it's taken some computer acid and now it's hallucinating like mastery things and yeah, just as you say, and.

That's not what it's doing.

And so when the kind of people who are trying to promote these these things as products talk about the AI hallucination problem, they're kind of selling a product that is a product that's representing the world usually checking things and occasionally makes mistakes. And if the mistakes were, like Joe said, we're a misfiring of a normally a reliable process or you know, something that normally represents going wrong in some way that would lend itself to certain solutions to them, and it would make you think there's an underlying reliable product here, right, which is exactly what somebody who's making a product to go on the market will.

Want you to think.

Right.

But if that's not what it's doing in a certain sense, they're misrepresenting what it's doing even when it gets things right.

And that's that's bad for all of us who are going to be using these.

Systems, especially since people, you know, most people don't know how this works. They're just understanding the product as it's described and using these kind of metaphors. So the way the metaphor describes it's going to really influence how they think about it and how they use it.

Yeah, just to sort of cat fire if I can.

Is one of the responses to some corners has been to sort of say, of us, look, you winch about people anthropomorphosizing chat GPT, but look, if you've got a bullshity, you're doing exactly the same thing. And I mean there might be some center which it's just really hard not to anthropomorphise it. I don't know why I picked the word that I can barely say, like we've been doing it constantly throughout this discussion. Right when we were talking through the kind of through the paper, we kept talking about chat GPT as if it had intention, right, as if it was thinking about anything. That might be another reason to call it bullshit. We go, look, if we have to treat it as if it's doing something like what we do, it's not hallucinating, it's not lying, it's not confabulating, it's bullshitting, right, And if we have to treat it as if it's behaving in some kind of human like way, here's the appropriate human like behavior to describe it.

I also think the language in this case, and one of the reasons they probably really like the large language model concept is language gives life too things when we describe the processes through which we interact with the world and interact with living beings, even cats, dogs, even we anthropomiz living things. But also when we communicate with something, languages life, and so it probably works out really fucking well for them. Samultman was saying a couple of weeks back, maybe a month or two. He was saying, oh, yeah, AI is not a creature with something he said, and it was just so obvious what he wanted people to do was say, but what if it was or are people saying this is a creature and it almost feels like just part of the con Yeah, yeah, yeah.

I hadn't thought about that as a reason for them to go for large language models as a way of kind of, I don't know, being the gateway into more investment.

Into fight consciousness.

Yeah.

Yeah, but I had thought about, like how this might have been caused by just like deep misunderstandings of returning test.

Right, go ahead, hear this one.

Yeah, yeah, so that like the Turing test, I think this is closer to what Turing was thinking. But the turning test is a way of getting evidence that something is conscious. Right, So you know, I'm not in your head, so I can't feel your feelings or think your thoughts directly. Right, I have to judge whether you're conscious based on how you interact with me, and the way I do it is by listening to what you say, right, and talking to you and turning sort.

Of was asked, you know, how would you know if a computer was conscious?

So, you know, we think that our brains are doing something similar to what computers do. That's a reason to think that maybe computers eventually could have thoughts.

Like ours, right, and some of us I think that I think it's possible. It's great.

Yeah, I didn't know if they thought it was possible, because not everybody thinks it's possible.

It's possible. I just don't know how.

Yeah, yeah, exactly. So Turing was kind of, you know, thinking like, how would we know? And one way we would know.

The obvious way is to do the same thing you do to humans, talk to it and see how it responds. And that's actually pretty good evidence if you don't have the ability to look deeper. But it's not constitutive of being conscious. It's not what makes something conscious or determines whether they're conscious or in any way like grounds their consciousness. Right, Their ability to talk to you is just evidence. It's just one signal you can get. And that's the way to think of the Turing test. So as a result of people thinking in a kind of behaviorist way, thinking, ah, passing the Turning test is just all it is to be a thinking thing. There have been, at least since the nineties, attempts to design chatbucks that can be the Turning test, right, and popularizations of these attempts and run throughs of the Turing test that talk as if, oh, if a computer finally beats the.

Turing test, I should say what the Turing test is?

Right? The way turns suggests that the test works as you have a computer and a person both chatting in some way with a judge, and the judge is also a person, right, And if the judge can't tell which of the people he's chatting with is a human, then the computer's one the right test because it's distinguishable from a human.

Right.

So people have taken this and it's been popularized as a way of determining sort of full stop, whether something's conscious.

But it's just a piece of evidence, and we have a lot.

More evidence, Like we know a lot more than Turning did about how the internal functioning of a mind works functionally, what it's doing, how representation works, how experience and belief works, how they are connected to action, and how they're connected to speech and thought. And once you know all that stuff, you have a lot of other avenues to get more evidence about whether the thing is conscious and whether it passes. The Turning test is just like a drop in the bucket compared to these, especially if you know how it's internal functioning.

Is the other the story is problem with the showing test, and I think to be fair, Cheuring did mention this. If not in the original then kind of later on. One problem with the chewing test is that it's like the de void camp in de androids gene of electric cheap right. Plenty of humans would fail the cheering test. Yeah, it is a piece of evidence, but it was never as Mike said, it was supposed to be constitutive. It wasn't like, if you can do this thing, your conscious kind of full stops. It was supposed to be here is a thing that might indicate that being you're talking to is conscious.

Happily, as my said, we've got the loads and loads of other evidence.

So these guys have made a machine that's just designed to do one thing, and that's past the Turing test.

I can give you one more annoying example. Are you familiar with Francoise Chole the abstraction and reasoning corpus, So this is going to make you laugh. So he's a Google engineer and he created this thing of the arc the abstraction reasoning corpus, the test whether a machine was intelligent, and someone created a model that could be it, and then he immediately went, Okay, you can't just train the model on the answers to the test.

This is why people, well I see people a fairly small subset of weird nerds, But this is why a small subset of weird nerds have been for the last twenty years emphasizing artificial general intelligence, right, and kind of like what we'll call it. When something really is a thinking being is when it's not specialized to do one and only one task, but rather when it's capable of applying reasoning to multiple kinds of different and this analogous cases. On the one hand, it does seem a little bit like the guy flipping the table and go, oh, for fox sake, now you've one. I'm changing the rules. But on the other I think he's got a point, right, Yeah, if you're training a thing to do very specific you know, like activate certain shiboleths, then unless you're some kind of mad hard behaviorist, then yeah, like it's not that doesn't demonstrate intelligence. It is one thing that might indicate intelligence.

It's the same problem with the chat GPT. It's built to resemble intelligence and resemble conscious This will resemble these things, but it isn't. It's almost like it's meaningless. On a very high end philosophical I find the whole generative a I think deeply nihilistic.

I mean, one thing that connects to this is how bad it is at reasoning. And this is kind of good for us, especially in philosophy, because our students when they use it to write papers, the papers have to have arguments, and chat GPT is very bad at doing reasoning. If it has to be you know, sort of extended argument or a proof or something like that, it's very bad at it. I think also that if there's one thing kind of we learned from chat GPT, it's that this is not the way to get to artificial general intelligence.

I was going to ask, do you think that this is getting to that?

No, partially because it's so subject specific, right, It's one it's trained to do one task, and it takes quite a lot of training to get it to do that task. Well, it's bad at many of the other tasks that we think are connected with intelligence. It's bad at logical and mathematical reasoning. I understand that open AI is trying to fix that. Some sometimes it sounds like they want to fix it by just connecting it to a database or a program that can do that for it. But either way, when you have with these kind of big base nets models, it's something that is really good at whatever you train it to do, but not going to be good at anything else. You know, it's got a lot of data on one thing. It finds patterns in that, it finds regularities in that, it represents those. The more data you feed it, the better it will be at that. But it's not going to have this kind of general ability, and it's not going to grow it out of learning how to speak English.

Have you heard the Terry Pratchett quote about It's quite early on.

He's talking about Hex, the kind of steampunk computer a make and unseen university, and it just has this off hand line. A computer program is exactly like a philosophy professor. Unless you ask it the question in exactly the right way, it will delight in giving you a perfectly accurate, completely unhelped answer.

Right. So if you abstract that justified you got a philosophy lecture.

Is there?

Basically what intelligent things do is go You can't have meant that, you must have meant this right, chat GPTs goes, I will take a question as read. I mean, of course it doesn't have an eye. There is nothing in the rest. Like I've just answered, formal play it again. But it's the same thing. And if trained to do attractively specific things, and you get the same problem as any program like garbage in Garbage album.

So have you found a lot of students using chat GBT, because this is a hard problem to quantify? Is it all the time?

I mean it's it's a lot.

I wouldn't say all the time.

I changed that you.

Thought that there were more seas indeed, yeah, so I think there are quite a few. And sometimes it is difficult for us to prove and if we can't prove it then at our university, then.

Well shucks, they kind of get away with it.

We can interview them, but unless we're like, unless we have the proof that you could take before like a court, then we are we're not able to really nail them for it, which is a bit of a shame. So suspicion, Yeah, we suspect that a lot of them are very I'm one hundred percent on some of them.

I just know what are the clues I want to hear from all of you on this one, like, what are the side?

It's the uncanny value.

I will not shut up about this, right you know that like you of course will be imagine a lot of your kind of the reader will be as well.

But the uncanny Valley.

Thing in respective humans is that there's a kind of a point up to which humans seem to trust artificial things more the more they look like a human. So like, we'll trust a robot if it's kind of bipedal or if it looks like a dog, But after a point, we really rapidly start to trust me. So like if it basically when it starts looking like data, some deep buried lizard brain goes danger.

Danger. This thing's trying to fuck with you somehow.

Yeah, right, and.

Chatter GPT writes exactly like that.

It writes almost, but not entirely, like a thing that is trying to convince you that it's human.

I mean, the dead giveway for me is that it never.

Makes any spelling mistakes, but it can't form out a paragraph to save its life. Normally, you would expect someone who didn't know how to kind of order their sentences to misspell the occasional word chat GPT spells everything correctly and doesn't know what like subject object agreement is.

It's like, it's.

So, can you just define that, please? Subject object the.

Beer that I drink, right, not the drink that I bear.

Right, Because it's interesting. It almost feels like there is just an entire way of processing and delivering information as a human being that we do not understand. There is something missing.

I mean, for me, like it's very good at summarizing, but when it responds, it responds in them in really trite and repetitive ways, or like you'll get a paper that summarizes a bunch of literature at length very effectively, and then respond by saying, well, you know, this person said this, that is doubtful. They should say more to support that, you know, which is basically saying nothing right, And that's pretty common.

It also does lists in formats things and kind of like a list.

And even if this like make it look less like a list, the paper still reads like a list.

Oh, because someone has asked give me a few thoughts on X. Yeah, yeah, that's very depressing.

Things like the introductions you get you'll get this is an interesting and complex question that parst of us have asked you the ages, and this is one of the things we shouted students not to write from first year.

And then you'll get this garbage right back at.

You, and then at the end it will be oh overall, this is a complicated and difficult question with many nuancewers.

It's a world of contrasts.

One of the things we tell them, don't ever fucking do this. This is terrible and right back at you in perfect English. That's one of English you'd expects a really good student might have written. But clearly they can't be a good student because otherwise they've listened to a fucking instruction.

I also, I had some papers that I suspected were chatept this year, but they were already failing, so I didn't okay, yeah, I didn't think it was worth it to pursue them, you know, as a plagiarism or Chatchepte case.

So it's never good papers. Then it's never like an eight, like a first.

No, absolutely not.

It's so I think that part of what goes on is you can get a passing grade with the Chatcheapte paper. Sometimes in the first couple of years, when the papers are shorter and we're not expecting as much. But then when you move into the what we call honors level here, which is like upper level classes in the US, like third and fourth.

Year, I've got first ever wrist with I know.

Yeah, yeah, exactly, great, well done.

Well you would not have gotten it with chat Gipt because you get dropped in these classes where we expect you to have gained writing skills minimal ones in your first two years and then we're going to build on that and have you do more complicated stuff, and chat Gipt doesn't build on that, right, it just stays where it was. So you go from writing kind of C grade passing papers, yeah, your F grade paper.

And it's also more obvious because the papers are longer, and chat gypt can write long text. But it gets very repetitive and noticeably repetitive, right, and so you're kind of lost, like you haven't done the work of figuring out how to write on your own and the tool that you've been using is not up to the task that you're now presented with. And so I think I have seen a few papers that I was suspicious of, but the papers that I was certain of were ones that were like senior thesis, the very clear person just had no way of writing culture.

That's insane at that stage.

Yeah, I mean it's fun because I mean one of the things that we get told about is like, oh, students have got to learn how to use you know, AI and large language model plagiarism machines responsibly and in a kind of positive way. Well, if they're using them in a way that means they don't learn how to write, then it's not positive. Visit like, yeah, it's fucking hard to write a good essay. Yes, it is hard to write. That's why we practice, That's why we have editors, that's why we do this collaboratively. And if you're using this as a sort of oh I don't know how to write, well, tough shit, you're never going to know how to write that. That doesn't seem a positive use of any of this.

Well that's the thing. Writing is also reading the consumption of information and then bouncing the ideas off of your brain allegedly.

Yeah, I worry about like, because CHATCHBT is good at summarizing, So I worry that one of the uses people will think, ah, this is a pretty good use for it is summarizing the paper that they're supposed to read, and it will do that effectively enough for them to discuss it in class if they're.

Willing to be here.

Yeah, right, but they're.

Not going to pick up a lot of the nuances in a lot of the kind of like stylistic ways of presenting ideas that you get when you actually do the reading.

And it's it's so frustrating as well, because like for this, for example, I've just got the printing off things that don't read PDFs anymore because I feel like you do need to focus.

Yeah, there's some evidence that reading physical copies makes you engage more.

Sorry I'm very old fashioned, I guess, but no, it's true though, But also reading that I wouldn't have really on a PDF. I wouldn't have given it as much attention. But also, going through this paper, you could see what you were doing, like you could see that you were lining up. Here are the qualities that we use to judge bullshit. But also summarizing a paper does not actually give you the argument, it gives you an answer. So what do you actually want students to do instead? Because I don't think there's any reason to use chat GPT for these reasons that it doesn't seem to do anything that's useful for them.

I don't have, no, I don't actually have any use for chat shabt that I can put to my students and say, here's.

What I think you should do with it.

We are like kind of developing strategies for keeping them from using it, So like building directly on what you're saying.

Like, in my class.

Next year, I'm going to have the students do regular assignments, which are argument summaries and not paper summaries. So the idea is they have to read the paper, find an argument, and tell me what are the premises, what's the conclusion? And that's something that chatchept is not good at, right, But it's also something that will give them critical reading skills, which is what I want to do.

Right. So Yeah, I think that I've.

Mostly been thinking about ways to keep them from relying on it because I think that often if they rely on it, they'll they'll they'll put.

Themselves in the worst position.

Yeah, when it comes to future work, they won't develop the skills that they're going to need, and the skills that we tell them and their parents they're gonna get with their college degree.

Right, it almost feels like we need more deliberacy in university education because I was not taught to write. I just did a lot of it until I got good enough grades and Daniel CHOYNDL look great mental, but I've had tons of them, and it almost feels like we need classes where it's like, Okay, no computer for this one. I'm gonna print this paper out and you're going to underline the things that are important and talk to me about Almost feels like we need to rebuild this because yes, we shouldn't be using chet GPT to half us our essays. But at the same time, human beings are lazy.

Yeah, I mean for me, I also prefer to read off the computer, but I often read PDFs because I'm terrible at keeping files, right, physical Like, you know, I'm not going to keep a giant filed for with all the papers that I've read and then written my liner notes in.

You guys can see in my office they're just piled around, like you can't see this end.

That's academia.

Yeah, but I just have piles of paper with empty coffee bugs everywhere that the camera is not facing.

But it's the terrible system so at least.

On my computer, if I'm like, oh, I read that paper like a year ago, what did I think?

I can click on it and see my own notes, and.

I do think that there's something to keeping those records and kind of actively.

Reading in that way. I don't know how I ended this without telling you how to make students do that. No, you started with the correct answer, which is don't use chat GPT. Yeah.

I actually I've got a certain amount of sympathy with like, just keep writing it till you get good at it.

But I realized as a lecturer of that camp in my official possession, and I certainly think that it's the case that certainly Glasgow has got better over the last few years about going, oh, actually you do like, we do need to give you some kind of structuring and some buttressing on here's how to write academically, here's how.

To do research.

And I think that's all to the good. It's worth saying this started happening well before chat GPT's started pissing a lover our doorsteps, so they don't get to claim that as being a benefit.

There was the whole Wikipedia panic when I was in school.

Yeah, they think about Wikipedia the right.

It's like I should to say this to my students.

WIKIPII one of the best resources.

Absolutely fine as a starting point for research, absolutely with it whatsoever. But if you're turning in and on his level, essay, I want you to go and read the fucking things it's referencing.

Yeah, that's right.

Yeah, I think these things are often great as sources. I'm worry about chat GPT is that it's not great as a source.

It's just yeah, we've.

Been saying it often gets things wrong, and it often it'll make that sources, whereas Wikipedia will never do that.

There are some famous hoaxes, but they get cool.

Yeah, Joe, have you got any positive things to say about chat gipt.

Fan?

So I know some people who have used these kinds of things productively, not in ways that our students would, but I know some mathematicians who have been using it to do sort of informal proofs and things like that. And it does still bullshit, and it bullshits very convincingly, which makes it very difficult to use for this kind of purpose.

But it can do some interesting and cool things.

Yeah, I think some people in that sort of field have found useful and also We've mentioned this before, like, if you want CHAT to be teach and write, you're a bibliography. You've got a bibliography in one style. Tell it to put something into a different one. Then it's that, And it's good for like coding the process doing certain things.

I also think, I don't know, I'm not sure how I feel about what I'm about.

To say, but perfect, Yeah.

I'm going for it.

It is a somewhat positive thing maybe for a CHAT GPT, which is that we often have students who have like really interesting ideas and well thought out arguments, but for whom English isn't their first language, and the.

Actual writing is kind of rough and you have.

To like push through reading it to get the good idea, which is often really there and quite creative and insightful. Yeah, and so I do wonder if there's a way to use it so it just smooths off the edges of this kind of thing. But I worry that if you tell students to do that, they'll just first if they can develop the language skills.

They offering really good by the year. Yeah, what are you going to say? Do well?

So I want to say, yeah, you can see me getting agitated. Yeah, I think much correct about it, like that this is a kind of possible use. But I think this and this is why I'm getting visibly agitated here. That's students either need to or feel they need to use. This speaks to a deeper issue, right to a social issue, to political issue, to an issue about how universities work. If a student is having problems with English, then there's a number of like explanations or a number of kind of responses.

Right.

The one response is that, like Glasgow is an English teaching university, if someone's English isn't good enough to be taking a degree, then possibly it shouldn't have been let in, and why have they been learned well?

Because of money?

Or alternatively, if someone's having problems with English for like whatever reason at all, there should be supported. There should be kind of tutors. They should be people who can with English. But again that would cost university money. So of course that doesn't happen.

Right, and it doesn't happen anywhere that it doesn't extent.

It would have to happen in order for this to be a general policy.

Right, Yeah, I think it could be better, But I do think that universities often have like a writing center or a tutoring center that you can.

Send, but they don't.

They don't have the sort of spread that would be needed or the staff that would be needed for this to be Instead of using chat GPT to sound the edges.

Off, yeah, I think for example twenty three with your supervisor.

My worry especially would be that this is my first year here at Glasgow.

But I probably have a good Yeah.

I think they probably have a good writing center. Universities have been in the past. I felt very confident sending students to the writing center when they have these problems. But I think James is completely right that we don't want the universities to see this as a way to get rid of the writing center. And that's one hundred percent a risk given the financial problems.

That universities are facing, and maybe they're already not. Yeah, Like given the quality of papers, they're often good.

I think another thing, as far as this is like a social problem, is that when grading, I myself tried to grade in terms of like the ideas and argument, because this is a philosophy and not the quality of the right. But not everybody does that. So I kind of think that another part of this is figuring out how we want to evaluate the students and what we want to privilege in that evaluation.

Yeah.

Sure, So then again the kind of that that becomes a problem about what people are checking for. Not let's take this arched backwards approach to marketing, which is like how fancy is your English?

Are fancy English? Is good English? Have an a? But rather we should be kind.

Of checking for different things.

Right, So again the blame blows differently in that case, but it still becomes a question.

It's not solved technological.

Yeah, almost feels like large language models are taking advantage of a certain kind of organizational failure.

What an idea?

Crazy that the tech industry manipulating parts of society.

That wait, I have a kind.

Of related tangent here, which is like what are the use cases that open ai was expecting but didn't want to emphasize Because for everybody in universities, as soon as this came out, the first thought was students are going to use this to cheat And certainly like the people in open ai went to college, right, and that's what I hear about.

Them, so they must well, sam Oman drop down, Oh yeah, I'm.

Sure he really understands the value of a secondary He's like, I'm.

Gonna write at least fucking essays.

Anyway, Maybe he was thinking I would have loved to have a computer write my essays.

I'll devote my life.

But I mean, I'm sure that they like recognize these bad use cases, right, but they're doing nothing to mitigate them as far as I can see. And like, another one that's very related is like, you know, I'm sure you've heard of this ed phishing, right, a lot of you know, corporations get attacked and get hacked not by someone cleverly figuring out a backdoor to their system, but by somebody sending.

A social engineering.

Yeah, asking for the password to somebody else. And one of the biggest barriers to that is that a lot of the people who are engaging in fishing aren't from the same country as the company they're targeting, right, so they're not able to write a convincing email or make a phone call that sounds like that person's supervisor. But with a tool like this, you could one percent write that email. Right. It's going to make it a lot easier for these kinds of illicit schemes to work.

That has been a market increase CNBC, which would just brow up, is that on two hundred and sixty five percent increase in malicious fishing emails since the launch of chat JPTA. Great style. I mean if I could have thought of that, imagine what a criminal could do.

Right.

But also, weren't the people at open AI thinking about.

That because they don't care?

Yeah?

Yeah, we will see in duastic fuck right.

Yeah.

Yeah, they were so busy thinking about what they could do they never thought about whether they should.

Yeah, this is the kind of problem with the move fast and break things mentality, like obvious.

I mean, I think it might be the only person who has raised in the US.

But we had future problems solvers, where you you know, you think about a future problem and what bad consequences there could be of some technology and how to solve them, usually through social cues. If I could do that in fifth grade.

I would expect these people to have thought through some of the bad consequences of the technology they're putting out. And you know, some of those are cheating on tests and they don't seem to have worried about that.

Yeah, And another one is fishing. They don't seem to have a worried about that.

You know, biases and algorithms, right, So yeah, yeah, comes no surprise to you. It turns out so with a lot of the facial recognition systems they were incredibly racist.

They were going back to Microsoft's connect they could not see back.

Yeah yeah, and CCTV stuff that basically just sort of nless. It was presented with a blindingly white Caucasian.

Where I don't know, right, but like.

The sort of stuff where these like language knowledge are trained on certain sets of data, and they trained on certain assumptions and like shitting shit out right, and particularly if people think that it's actually doing any kind of thinking, and if they kind of cargo cult it, we again get a kind of social problem multiplied by technology feeding back into a social problem, right. And it's the So these guys have heard me when you're about it so much, I love it or so, but I'm profoundly skeptical of technology's ability to solve anything unless we know exactly the respect in which we want to solve it and how that technology is going to be applied. You know, like sure, experiment with bringing back dinosaurs, but but like, don't tell me that it's going to save the healthcare system unless you can demonstrate it to be step by step. How that big old rex run around on Eiland Nublade is going to save anything.

They just tried to blind.

People actually bringing back dinosaurs.

That would be great.

All right, this isn't the best example, but I already had. I had Jeff Gobwoman in my head.

I had to go to the fellas. This has been such a pleasure. Don't put that in the recording.

That's right.

We won't edit it in post, and you will give us your names.

I'm Mike Hicks, but my papers are written by Michael Townsend Hicks and I'm a lecturer at the University of Glasgow.

My website is Hicks.

It will be in the podcast profile.

Don't you worry. We'll get that just plugging my luggy stuff.

My name is Joe Slater.

I'm a university lecturer in moral and political policy at Glasgow.

I'm James Joffrees.

I'm a lecturer in political theory at the University of Glasgow.

And even if I.

Wanted to or couldn't give you what websites, I don't have.

Everyone you've been listening to Better Offline, thank you so much for listening. Everyone guys, thank you for joining me.

Thanks having me here.

This is.

Thank you for listening to Better Offline. The editor and composer of the Better Offline theme song is Matasowski. You can check out more of his music and audio projects at Matasowski dot com, M A T T O S O W s ki dot com. You can email me at easy at Better offline dot com or visit Better Offline dot com to find more podcast links and of course, my newsletter. I also really recommend you go to chat dot Where's youreed dot at to visit the discord, and go to our slash Better off Line to check out our reddit. Thank you so much for listening. Better Offline is a production of cool Zone Media. For more from cool Zone Media, visit our website cool

Zonemedia dot com, or check us out on the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts.

In 1 playlist(s)

  1. Better Offline

    87 clip(s)

Better Offline

Better Offline is a weekly show exploring the tech industry’s influence and manipulation of society  
Social links
Follow podcast
Recent clips
Browse 87 clip(s)