Clean

Did AI Write This?

Published Sep 11, 2023, 8:52 PM

Figuring out if artificial intelligence wrote a block of text can be tricky. Some companies have created tools that claim to determine if text was likely the product of a human author or AI. But as we have learned, these tools aren't reliable. What makes it so difficult to tell who wrote what?

Welcome to Tech Stuff, a production from iHeartRadio. Hey there, and welcome to tech Stuff. I'm your host, Jonathan Strickland. I'm an executive producer with iHeartRadio. And how the tech are you. I'm here to tell you something. You write like a robot. But that's okay because I do too. One of the founding fathers of the United States, James Madison, wrote like a robot. Robots weren't even a thing when he was writing back in the eighteenth century, all right, so really, I guess it's more fair to say that robots write like us. And while I'm having a little bit of fun using the word robots, what I'm really talking about is generative AI. You know, stuff like chat GPT and Google Bard, that kind of thing, These AI powered chat bots right like humans. Right, That's one of the big suff features of the chatbots. One that they can understand a prompt that we give them, That they can understand what we mean when we give them a prompt, and two that they then generate a response as if it had been written by an actual person. But obviously this also creates some challenges, some issues. So you might remember that since chat GPT became publicly available last year when OpenAI opened it up and let people start playing with chat GPT, there were people in education, teachers and administrators that sort of thing, who raise the alarm about the possibility that students could use chat GPT and similar tools to auto generate essays and stuff and thus bypass school assignments. My robot wrote it for me. Beyond the education sector, there are plenty of arenas where people are worried that the less scrupulous folks out there will attempt to pass off AI generated text as their own writing, whether this is creative writing or business writing, whatever it may be. So this then leads us to the concept of AI writing detection tools, you know, some sort of tool to determine if a piece of text originated from a real human being or from that character that Haley Joel Osmon played in that film about artificial intelligence. I forget what that movie was called. Subsequent to the release of these detection tools, we started hearing reports of teachers failing students, sometimes an entire class of students, because the detection tool indicated that the real source of the works that were being turned in by the students it wasn't from the students, but from AI. Now a lot of students have actually come forward to argue that no, no, they actually wrote those pieces themselves, that they authored that work, they didn't use AI to do it, and that they are the victim of false positives, that these writing detection tools made a mistake, and as it turns out, at least some of them, and likely a lot of them were telling the truth. And we can say that because these AI writing detection tools have abysmal accuracy rates, they are worse than chance. That's how bad these tools can be. So the success rate for an AI writing detector can be so low that it has led some of the companies to shut them down, and it led to a lot of critics to just dismiss the concept of an AI writing tool entirely. In fact, there are quite a few who have argued that AI writing detection tools are essentially snake oil. That there are companies that are making what they say are reliable tools that can tell the difference between text that was written by person and text that was written by AI, but really they're just peddling a hoax or a scam, and they're trying to make money selling these tools to various organizations like schools and such, but in fact those tools don't work, or at least they don't work very well. Even open Ai, which is the company that is responsible for chat GPT, they had a tool that was meant to be a detection tool to tell whether or not something was written by AI. It was called AI Classifier, but they shut it down earlier this year. Why because its accuracy rate was twenty six percent. Twenty six percent accurate, that is bonkers. That means nearly three quarters of the time that detection tool came up with the wrong answer. Either it gave a pass to an AI generated piece, or it accused a work that a human being actually wrote, like definitively wrote, as being the product of AI. This brings us to James Madison. James Madison wrote the US Constitution, and folks have fed the US Constitution into these AI writing detection tools and received a notification that this piece was very likely written by AI, which obviously led to lots of jocularity on the Internet, as people said, I knew it. I knew that the founding fathers of the United States of America were really robots from the future sent back in time to create a ultra capitalist society that preys upon the disenfranchised or something like. There are a lot of jokes about it, but the fact is no, it's just that this writing detection tool is completely unreliable. So you certainly cannot use these kinds of tools to justify flunking an entire class of students when you know that the reliability is so low. Now, I decided to do this short episode about AI writing detection tools after reading a couple of great pieces in Ours Technico. Those of y'all who listen to my show frequently know that I often reference Ours Technica because the folks there reliably post great articles. So in this case, the author of both pieces I read was BENJ. Edwards b E and J. Edwards, And at some point I probably should reach out to them and ask if they would like to join tech stuff for an episode to talk about something like generative AI, because Edwards has done some really good work. Anyways, as we think about the issue about how this generative AI works, the underlying technology that powers generative AI, we start to see why there's this big reliability problem. Why are we having such issues with an automated detection tool? Really determining if something was written by a person or AI. And it's because the tools like chat GPT are built on top of large language models, also known as llms, And if we take a moment to really understand llms, then we start to get a handle on why these detector tools are so unreliable. So first off, let's actually talk about a precursor to large language models. This would be recurrent neural networks or r ends. Now I've talked a lot about neural networks on this show, but just as a refresher. Neural network is an attempt to create a computer system or computer model that processes information in a way that is similar to how our brains process information. So you have layers of artificial neurons, or you can think of them as nodes. These layers connect to other artificial neurons. You have multiple connections from neuron to other neurons, and you have layers that go from top to bottom. You can think of it like at the top that's where you put input and at the bottom that's where you get output. So essentially, you feed information into the model and then the information goes through a series of operations in which data passes through these different nodes, and the nodes make decisions based upon the input, and then they send output to different nodes and eventually you get the ultimate output. And sometimes that output is correct. It gives you the answer that is correct. Sometimes it's wrong. And typically what that means is that you then have to adjust how those artificial neurons are making decisions. Those neurons apply a sort of bias to input, we call it a weight, so they will favor some types of input over others in an effort to make a decision. If they didn't, then the data would never go anywhere. You would never be able to have it processed through the system. So the weighting affects how the neuron actually processes the data, where does it pass it on to. So it may say, if value is greater than X, send to node A. If value is less than x, send to node B. That could be a very basic weight. X would be the weight in that case, and maybe that would lead you to a correct outcome. So by adjusting the weighting, you can change how these neurons make decisions. And if you build a neural network for the purposes, let's give it a hypothetical. Let's say it's identifying pictures of cats. It's always my go to. And you start looking at the output and you see that it is mistakenly saying that pictures of flowers are pictures of cats. You would say, all right, these artificial neural networks, the nodes in this artificial neural network are making the wrong decisions. The waiting is wrong in these nodes. I need to go and start adjusting things so that I can start to get back to this correctly saying whether or not an image has a cat in it or doesn't. And your goal is to train this model over and over and over again until it gets better and better at this task, so that then you can just send it any raw data you like and not have to worry about checking up on it afterward because its accuracy level will be high enough to be reliable. That's your ultimate goal, But there's a whole process of learning of training that you have to go through first. Now, a recurrent neural network, it's a little more specific than just artificial neural network. Recurrent neural networks use sequential data. These networks can and do take information from earlier inputs into consideration when processing a new input, so there's a different model, the convolutional neural network CNN, not the news channel. This is the other big type of neural network where every time data goes into an input, it's like a blank slate. It's its own thing, it has nothing about That decision is based upon any past decision. It's an instance by instance kind of case. So you're starting from scratch. But with recurrent neural networks, the network can actually incorporate past inputs as part of how it processes a current input. But one issue with these types of networks, the recurrent neural networks is that they need a full sequence before they can process the information. So when we're talking about text, like if we wanted to process text through a recurrent neural network, it would need to work over the entire text before producing a result in order to understand things like context. Sometimes this approach can lead to errors because the model essentially forgets the stuff that was at the beginning of the text by the time it gets to the end, which sounds a lot like me honestly, where I will finish a book and then I'll think, like I'll have a discussion with someone about a book that we've both read and they'll be like, Oh, I like that part where in early in the book blah blah blah blah blah, and it pays off much later, and meanwhile, I'm thinking, I totally forgot that that happened earlier in the book. I remember where we ended up, but I don't remember how we got there. Recurrent neural networks can fall into the same sort of trap, and so that creates a bit of a hurdle when it comes to things like analyzing text for the purposes of building natural language systems. But I'll explain how that all started to change in twenty seventeen. First, however, we need to take a quick break to thank our sponsors. Okay, before the break, I was talking about recurrent neural networks and how those have certain limitations when it comes to the way they process data because it has to be sequential. Well, in twenty seventeen, a group of AI researchers who were working specifically over at Google were coming up with an alternative to this approach, and they published a paper, and the paper's title was Attention is All You Need, in which they suggested that you could do something differently from the recurrent neural network approach for the purposes of analyzing stuff like text. Their approach was what they called a transformer model. While you're old, RNN would analyze text essentially a character at a time, not even a word at a time, but a character at a time, and thus that's sequential, right. The sequential data is character by character. It builds this up and then analyzes the whole thing. The transformer model instead would tackle a sentence as a unit as opposed to a character or even an entire passage of text would be a single unit, and so it would analyze this to understand the context of what was being said, and that's a huge benefit you. Getting a handle on context is absolutely critical to understanding what someone means, because words can have multiple meanings, right, and without context, we can't really be sure which meaning someone intended. So here's an example. The English word late. That can mean a lot of things if you're an English speaker. So if you're talking about the time of day, if you say it's late, you usually mean it's getting close to night time. You could say it's late at night, which means it's actually close to morning time, or maybe it even is the morning because it's still dark. And so you think of it as night, but technically speaking, it's morning and you're just saying it's late at night. If you're saying somebody is late, you could either mean they are not on time for some appointment, or tragically, you could mean that this is a person who has passed away. They are late. But you need the rest of the sentence. You need that context to understand what meaning of late was actually intended. So you need that contextual vision to be able to understand the whole thing. So transformer models began to revolutionize certain types of AI applications, specifically in the realm of natural language processing and generative AI, and it's what led to the development of large language models the lms. Essentially, a large language model is just a huge transformer model. And to make a large language model, you need a lot of text to train your model, like a lot a lot. Open AI trained its large language model known as GPT, which stands for Generative pre Trained Transformer. They trained it on countless documents, millions and millions of documents found across the web. Some authors allege that the training material included copyrighted material and that the authors did not give permission for their works to be part of the information that fed into this model. That leads into its own set of problems that are a little bit beyond the scope of what I'm talking about today, but they are big problems and they're ongoing now. Stephen King argued that his works were clearly used to train up large language models. A dead giveaway is if you ask a chatbot built on top of a large language model to recite passages from specific authors works, and if it can do that like accurate, like it's really giving you an accurate representation of that text. Yeah, there's no way could have received that information without having trained on the original text at least somewhere. Now, if it's just making stuff up, that's different. That falls into the category of hallucinations, which we might touch upon again before we finish shut this episode. Anyway, the benefit of feeding so much information to a transformer model is that the transformer model, the large language model, gets pretty darn good at sussing out context. Even stuff that you would expect would trip up an AI chatbot can become a breeze. You know, you might think that slang or idioms could trip up an AI tool, but then you have to remember that these tools rely on essentially all the stuff that's on the Internet, at least all the stuff that's publicly available that's not locked behind something, and maybe even some stuff that is locked behind stuff. As it turns out, and as such, that means that these models have trained with data sets that originate from the same communities that are creating the culture that generates certain slang and idioms in the first place. So if your AI model is using the same source material where these turns of phrase and certain slang terms are are originating from, well, of course it's going to understand it because that was part of its training, so it has that grounding. It's not like me, where I am old. I don't understand slang that the kids use these days because I'm not in those communities. You wouldn't expect me to understand. I am definitely the stereotypical out of touch old dude. So when I hear people about, you know, people rizing up, I'm like, wait what? And I have to look things up. And as we all know, urban dictionary is not the most reliable of resources. It is frequently entertaining, usually in a way that is incredibly offensive, but it's not always accurate anyway. This ultimately starts to lead us to why these AI writing detection tools are not very good. The material that AI generates is built upon how we communicate. It's a built on how we write. That's how it was trained. So it's not like AI or robots, as I was facetiously saying earlier in the episode. It's not like AI has a different path toward writing than we do. The AI is not following an established set of rules that's unique to AI. Right, They're not saying, write this like artificial intelligence. So the stuff that AI produces can come across as very human and vice versa. Now, this does not mean that it is absolutely impossible for someone like a teacher to tell if something was written by AI or a student. If the teacher is actually really familiar with the writing style of that student or students in question, it's entirely possible that the teacher might notice if that writing style were to suddenly and maybe significantly change between assignments. This can be a big ask, by the way, for certain teachers, because class sizes can get huge depending on where you are, and if you're talking about an overworked English teacher who's teaching multiple classes and each class has got, you know, thirty kids in it. It can be hard to really build up a working knowledge and memory of the writing styles of every single person in every single class. But that is one way that teachers can tell. If teachers read an essay and think, wow, you know, Robert didn't write like this in the essay we did last month, this is a very different approach to writing and per perhaps that's an indicator that someone else wrote the piece, whether that was AI or maybe you know, another human being, and that can be an indication something hinky is going on. Also, I mean, obviously some people get sloppy. This happens a lot too when people just aren't paying attention as they're using AI to generate either you know, an educational assignment or business or whatever. There have been so many examples of how people have accidentally copied and pasted not just the body of the text, but stuff that's outside the body of the text, like it might even be a little disclaimer saying it was made by AI, or it could be a command like regenerate response. That's something you find in certain chat bots, and that is just what regenerate response means. It just means, hey, can you create a new AI response to the initial prompt I gave you. So I wrote a prompt, I had you generate response. I want you to create a whole new response based on that original prompt. If you have regenerate response written at in your essay, that's a dead giveaway that you copied and pasted that essay off of an AI chatbot. So there are ways that teachers can tell the difference, but they aren't. It's not as granular as saying, oh, this is clearly something that was written by artificial intelligence versus this was written by a human. It's more like this is different from what I have received before from this particular student, or this contains obvious errors that reveal that the student has used AI. Now, the AI writing detection tools are at least claiming to use a couple of strategies to try and determine if something was written by AI or a human. So they're saying, we can automate that process, and we can actually analyze a block of text and give you a determination as to whether or not that was made by AI or a human, which suggests that maybe there is some sort of fundamental difference between the way AI generates content and the way people do. But these strategies that the AI writing detection tools are built upon have fundamental flaws, and we know that because we know the tools are bad. It was bad enough for open ai to shut down its version back in June. So this isn't like just us postulating that these tools are bad. We know they're bad. We know they create things like false positives. So knowing that already they are unreliable, you then have to start asking, well, why are they unreliable? What are the things that are leading these tools to make these wrong determinations? And when we come back, I'll talk about how Bene Edwards and those OURS Technica articles really kind of digs into two main concepts that end up leading to these writing detection tools trying to make a determination and why they are fundamentally flawed. But first let's take another quick break. So before the break, I mentioned that I was going to talk about some strategies that Binge Edwards outlines in his RS Technica articles, and they fall into two categories. The first is called perplexity, and that really means how surprising or perplexing are the word choices, how creative are the sentences in a given piece of text compared to an AI training model. So the thinking behind this is that if a block of text seems to conform to the same sort of stuff that the language model would produce, then AI probably created the text. That's the idea if they're saying essentially that if the text is really similar to what AI would create, then AI probably created it. And let's think about how some tools use autocomplete to help you write a text or sentence. Using a purely hypothetical scenario to kind of get our minds wrapped around this, Let's say that you were typing into something that has autocomplete built into it, the sentence or the phrase I'm going to go for a and then whatever tool you're typing it into suggests the word walk as an autocomplete option. Well, that would be because the language model that is powering this autocomplete function has a has sampled millions of passages, millions and millions and millions of documents, and has found that the word walk has been the most common word to follow the phrase I'm going to go for a and so therefore it offers that as the suggestion, and maybe it would even offer you a few options. Maybe it would say walk, maybe it'd say swim in the UK, maybe it'd say a curry. Who knows so, but you know, it would give you maybe a couple of different options, but they would be the ones that would most likely follow that phrase based upon the training material that that large language model had used to build itself up. Right, So if you were to measure the perplexity of the sentence I'm going to go for a walk, it would be very very low, very low perplexity because it's in line with what the language model would expect. So the thought is, if a passage in general has a very low perplexity, these tools tend to suspect that the passage as a whole could have come from AI. So let's say that it had a very hyperplexity. Let's say that instead of saying I'm going to go for a walk, you said I'm going to go for a zebra or zebra if you're in the UK. Well, that doesn't want it doesn't really make any sense. But two, that would be very perplexing, right, that would not be something that the large language model would expect. And so if it has high perplexity, then the writing detection tool is more likely to say this was written by a human, because what generative chat system would have made that sentence, And he's like, no, sane robot would say I'm going to go for a zebra. Clearly some human wrote this. Now, the problem is these companies are training their large language models on enormous amounts of human generated text. And unless you're purposefully trying to be really a original in your writing, that's a kind way of saying you're being a weirdo as you're writing your sentences. Chances are a lot of the stuff you're writing is going to have a fairly low perplexity, unless you're trying to write in like the milieu of humor or absurdity. If unless you're purposely trying to do that, then chances are your perplexity is going to be pretty low too. Particularly for very structured writing like business writing or academic writing, that perplexity is going to be very low. So unless you're prone to throwing in very odd, random, weird sentences like William Shakespeare's Othello is one of the great tragedies of English theater, and also I enjoy shoving hot dogs through mail slots. Well, there's a pretty good chance that an AI detector tool is going to think that your human written, legitimate essay was in fact an AI's work, because the perplexity would likely be pretty low, again unless you're doing something really avant garde, So that there's a fundamental flaw and logic of using perplexity as one of your metrics for determining if something was written by AI versus a human. Ben Jedwards also goes on to explain that another factor that AI detection tools will take into consideration is one that's called burstiness. Perplexity and burstiness makes me feel like I've fallen into a Lewis Carroll novel. But anyway, burstiness really has to do with variability, particularly between sentences. So y'all probably have noticed I have a tendency toward really long sentences, and often with a lot of parentheticals thrown in there. Now, if I also incorporate short sentences on occasion, breaking up these very long sentences, this creates a lot more variety, a lot more dynamic elements between my sentences, because I'm switching back and forth between these very long, pontificating sentences and then short ones to make a point. Maybe in some sentences I use tons of adverbs to describe action. Maybe in the next sentence I don't use any adverbs at all. This is what creates that variability. The conventional wisdom is that AI generated work is more uniform, it's more consistent, it has less variability from sentence to sentence. Your sentence length and complexity is going to remain more or less the same throughout an entire passage. So if you're able to qualify how dynamic a writing style is, the thinking goes. You could potentially determine if a human wrote it or if an AI wrote that specific piece. If it's not very dynamic, well that leads more toward AI. But that approach depends upon a couple of things that are not always reliable. So first up, it assumes that AI generated content is going to contain you to show more consistency than the stuff that humans. Right, that's going to continue to be this very consistent approach to sentence structure. But the language models and the generative AI that are built on top of the language models are growing more sophisticated all the time. A lot of these companies that make these language models are mining platforms like x formerly known as Twitter or Reddit in order to train their language models. They're reading these sort of idiosyncratic messages that people write. Sometimes they're writing purposefully in a way that is not consistent, and it can get to be a little unpredictable. Well, if you're training your language model on these things, then over time the language models and the tools that are built on top of them begin to reflect that training material. It means that we should expect generative AI to start increasing variability in sentence because that's what we're training it on. You can't expect to train it on one thing and it generates something totally different. It's going to kind of mimic the material that was used to teach it in the first place. So that means you're going to see a reduction in the gap between how AI creates text and how humans do. But on top of that, again, for certain types of writing, human authors may take a more structured approach and they may purposefully reduce variability between sentences or unconsciously reduce variability. That means that their writing is going to start looking more like the stuff that these writing detection tools assume. Is a marker for AI generated content. If I were to write a term paper, I would probably take a more consistent, uniform approach to my writing style. That's not to suggest that would be the right choice, right, Like, I'm not saying that if you write a term paper you need to have this very consistent, uniform approach because they can get really boring to read papers that are written in a style like that. But that would probably be my inclination, like thinking in my head, I'd be I want to make sure I'm consistent, I'm academic, i am thoughtful, I'm methodical. That means that the work I would produce would have this low burstiness because I was purposefully doing it. Even if that was the wrong decision, it probably be the one that I would make because I'd be working under the mistaken belief that this is somehow more academic. So these AI writing detection tools are looking for texts that has low burstiness and low perplexity before suggesting that AI had created that particular block of text. But as we've talked about, humans right in that kind of style too, particularly for formal writing, and so you get a lot of false positives, like if you feed the US Constitution to a writing detection tool, and it says, well, Ai wrote this, Well, a lot of stuff has been written about the Constitution, including passages from the content Institution. The Constitution itself is clearly available on the web, so it's obviously part of these large language models training sets. So of course it's going to reflect what's in the training set. It was literally incorporated into it. So if you're working backward from that logic, then your conclusion, oh Ai wrote this because it reflects what the language model was trained on. Well, yeah, but that's because the language model was literally trained on the material you were analyzing. It becomes the sort of catch twenty two sort of situation. So we cannot rely on these detection tools in large part. Now, this doesn't even touch upon the challenges that non native English speakers face with their writing. When they're writing in English and these AI detection tools are used on their work, they can face disproportionate bias when it comes to these detection tools. They get a lot more false positive So you're already seeing a lot of false positives anyway, because as we've discussed, the criteria being used by these AI writing detection tools are faulty because it's making assumptions that humans are not writing in those styles when in fact they are, and that AI is writing in one specific style, when in fact, at least over time, it migrates away from that. So you got a double whammy here. Now, there are some applications of AI detection tools where it works and it makes sense, just not in writing, but for stuff like photo or video manipulation. AI detection tools can still look for telltale signs that can indicate that maybe what you're looking at has at least in some part been created by a generative AI tool, right like an image creation tool. Obviously, there are examples of this where you take one look and you know immediately that this was made by AI, because you look at it and you're like, no one has that many fingers on one hand, but there are other cases where it may not. It may be far more subtle to a human perception, but if you were to actually analyze the image deeply with a very well trained AI detection tool, it could indicate this was made by AI because of little subtle things. Maybe it's inconsistent lighting, Maybe it's a blinking pattern of a person in a video, things like that, Little things that would be hard for us to spot as human beings, but easy for a detection tool to spot. These AI detection tools make sense. They're not necessarily foolproof or flawless, but they have a better success rate than when it comes to writing, because it's just not that clear cut when we're talking about writing. This is unfortunate when teachers may rely heavily on AI writing detection tools in order to determine if their students are actually doing their own work or not. If the teachers are unaware that these detection tools are unreliable, they can make some really drastic decisions that will have a huge negative impact on their students' work and lives, and that's not really fair. Hopefully, the educators out there are themselves educating themselves to be repetitive about these tools and their unreliability, because otherwise they're going to be punishing students and they can't justify it because it's all based on a tool that has proven to be unreliable at the get go, unless, of course, we're talking about instances where someone has copy and pasted some ridiculous part of an AI generated response that just gives it away. That's a different case. Entirely obviously, But yeah, I think it's important to understand the limitations of these As we explore generative AI, and we look at the pros and the cons and we consider the impact the generative AI has on multiple segments of our lives, we also have to really think about how do we know when it's in use, and how do we know that the tools we're using to make those determinations are actually good tools. In the case of these AI writing detection tools, it looks to me like you might as well not even look at them. You are more likely than not to get an incorrect answer, because again, we train these generative tools to communicate very much the way humans do, at least in certain use cases, and those use cases typically are the ones where we're most concerned about whether or not AI was put to use in the first place. So really interesting articles over on Ours Technica. It leads to this really deep discussion about generative AI, the limitations that we have in detecting it, And obviously there are a lot of other things we could touch on. I mentioned copyright. That's a big one, because if AI can regurgitate copyrighted works with no flaws, then that can be a huge blow to authors, for example, or we talked a little bit about hallucinations. Hallucinations are when an AI tool does not have the information to be able to determine what should come next in a sentence. You have to remember when you really boil it down these AI generative tools, what they're doing is they're following a very sophisticated statistical model to determine what should come next in its answer. So you give it a prompt and it's referencing this incredibly complicated statistical model to say, all right, what should I put as a response. Some of the information involves things like the actual answers to questions, but there are cases where the AI model may be unable to identify what the answer to the question is, but it still needs to answer your query. It doesn't have the answer, so it makes it up, but following this very sophisticated statistical model so that the answer it generates appears to be valid even though it's just completely made up. This is what we call hallucinations in AI. It's when AI generates an answer in order to respond to a query, but that answer is fabricated. It's a confabulation. That's another word that some people are using rather than hallucination, and it comes across as being very much legitimate because again, these very sophisticated statistical models make it seem authoritative and knowledge. The way the sentences are structured, it doesn't come across wishy washy. It's not like maybe it's blah blah blah. It ends up being it's blah blah blah and presented in such a way that you feel like it's reliable, even though ultimately it's not. That's another issue. It's related to what we're talking about. And it's also means that as a student, or as a business writer or as a lawyer, as one person found out earlier this year, you should not rely on generative AI as your one and only source for anything AI. Generative AI has even been found to fabricate quotations from people. Obviously that's not good either. There are lots of issues here. Anyway. I hope that was some food for thought for y'all. I hope you're doing well. I will talk to you again really soon. Tech Stuff is an eye Heart Radio production. For more podcasts from iHeartRadio, visit the iHeartRadio app, Apple Podcasts, or wherever you listen to your favorite shows

In 1 playlist(s)

  1. TechStuff

    2,448 clip(s)

TechStuff

TechStuff is getting a system update. Everything you love about TechStuff now twice the bandwidth wi 
Social links
Follow podcast
Recent clips
Browse 2,445 clip(s)