A Strange War: AI v. AI Detectors

“Upload your essay into Turnitin by 11:59pm on Thursday night?
You meant start the essay at 11:57 then submit it at 11:58, am I right?” — 
Gigachad ChatGPT student.

We begin our discussion by discussing the sweet smell of plagiarism.

It wafts in the air as educators run around like headless chickens, looking here, looking there as they flip through oddly good essays with panicked expressions.

“Was this AI-generated, Bobby?!” says a hapless teacher, staring at a piece of paper that seems curiously bereft of grammatical errors, suspecting that Bobby could never have created something of this caliber.

“No teacher, I just became smart!” Bobby cries, running off into the sunset because he is sad, he is going to become a member of an emo boyband, and he doesn’t want to admit that he generated his homework with ChatGPT.

generated by Midjourney, if that wasn’t obvious

This smell casts fear and trepidation over every single part of our education system, for it threatens to break it; after all, education is special and it is meant to be sacrosanct — after all, is it not the very same system that is designed to teach humans facts and knowledge and above all, to communicate and collaborate to solve the problems of our era with intelligence, initiative, and drive?

It’s unsurprising that the world of education has flipped out over ChatGPT, because artificial intelligence opens up the very real possibility that schools may be unable to detect it.

Fun and games, right? It’s just a bunch of kids cheating on assignments with artificial intelligence? It’s not going to affect the older generation?

As it turns out, no — that’s not the case. I’ll explain why later.

But before that, let’s talk a bit about the part of our education system that AI is threatening: Essay-writing.

If students simply choose to let their work be completed by artificial intelligence and forget all else, that just means that they’ve forgone the education that they’re supposed to have received, thereby crippling them by an act of personal choice, right…?

But each of us has been a student, and if we have children, our children either will be or have been students too; there is a deep emotional connection that stretches across the entire world when it comes to this.

Therefore, when Princeton University CS student Edward Tian swooped in to offer a solution,it’s not all that surprising that the world flipped.

Enter GPTZero.

Humans deserve the truth. A noble statement and a very bold one for a plagiarism detector, but something that’s a little deeper than most of us would probably imagine.

But consider this.

Not everyone who uses AI text is cheating in the sense of doing something that they are not supposed to and thereby violating rules, therefore the word ‘plagiarism detector’ doesn’t quite or always apply here.

This algorithm, as with other algorithms that attempt to detect AI-generated text, is not just a plagiarism detector that merely serves to catch students in petty acts of cheating —it is an AI detector.

An AI Detector At Work.

So how does it work?

GPTZero assigns a likelihood that a particular text is generated by AI by using two measures:

Perplexity, and Burstiness.

Essentially, in more human language than that which was presented on GPTZero’s website, GPTZero says that…

The less random the text (its ‘perplexity’), the more likely it was generated by an AI.

The less that randomness changes throughout time (its ‘burstiness’), the more likely that the text was generated by AI.

Anyway, GPTZero gives each text a score for perplexity and burstiness, and from there, outputs a probability that given sentences of a text were generated by AI, highlighting the relevant sentences, and easily displays the result to the user.

Alright, sounds great!

Does GPTZero deserve the hype, though?

…Does this actually work?

Let’s try it with this pleasant and AI-generated text that is exactly about the importance of hype (lol).

That’s 100% AI-generated and we know that as fact.

…Would we know if we didn’t see it in the ChatGPT terminal window, though?

…Okay, let’s not think about that.

Down the hatch…

…And boom.

As we can see, GPTZero, humanity’s champion, managed to identify that the text that we had generated was written by AI.

Hurrah!!!

Or…?

I proceeded to rewrite the essay with another AI software.

…After which GPTZero essentially declared:

So nope, GPTZero can’t detect rewritten texts that were generated with AI — which it should be able to if it truly is an *AI* detector in the best sense — and which in turn suggests that the way that it’s been operationalized has yet to allow it to be the bastion protecting humanity from the incursion of robots into our lives.

It’s not that GPTZero — or even OpenAI’s own AI Text Reviewer, amongst a whole panoply of different AI detectors – are bad or poorly operationalized, by any means. Rather, it’s that the operationalization is supremely difficult because the task is punishingly hard, and that we are unlikely to have a tool that can detect AI-generated text 100% unless we perform watermarking (MIT Technology Review) and we would have to use multiple algorithms to be able to detect text, or come up with alternate measures to do so.

An Arms Race between AI Large Language Models (LLMS) and AI Detectors — and why you should care (even if you’re not a student).

As I’ve mentioned, there is an arms race at hand between AI Large Language Models (LLMs) like ChatGPT, and AI detectors like GPTZero, the consequence of which is likely that the two will compete with one another and each will make progress in its own way, progressing the direction of both technologies forward.

Personally, I think that AI detectors are fighting a losing battle against LLMs for many reasons, but let me not put the cart before the horse — it is a battle to watch, not to predict the outcome of before it’s even begun.

Implications of this strange war:

But why should you care about any of this if you’re not a student? It’s not like you’re going to be looking at essays constantly, right?

Let’s take aside the fact that you’re reading a blog post right now, and let’s also move away purely from the plagiarized essay bit that we’ve been thinking about, as we gravitate towards thinking about how ChatGPT is a language model.

It’s a good bet that you use language everywhere in your life, business, and relationships with other people in order to communicate, coordinate, and everything else.

When we go around on the Internet, it’s not always immediately evident what was AI generated, what was generated by a human or, for that matter, what was inspired by an AI and later followed through by a human.

The whole reason we need something like a plagiarism detector is that we may not even be sure that a particular piece of language (which we most often experience in the form of text) is AI-generated with our own eyes and minds, to the point that we need to literally rely upon statistical patterns in order to evaluate some thing that we are looking at directly in front of us, thereby recruiting our brains as we evaluate the entirety of an output.

The problem is…

Language doesn’t just exist as text.

Language exists as text, yes, but also as speech. Moreover, speech and text are easily convertible to one another — and we know very well what ChatGPT is doing: Generating text.

We now know that there are Text To Speech (TTS) models that generate speech from text. They’re not necessarily all great, but that’s besides the point — it presages the translation from text into voice.

Think about it.

If the voices that are generated by AI become sufficiently realistic-sounding and their intonations (VALL-E, is that you?), how might you know that these voices aren’t real unless there are severe model safeguards that impede the models from functioning as they are supposed to?

Now combine that indistinguishable voice with sophisticated ChatGPT output that can evade any AI detector and in turn may, depending on the features that end up developing, evade your own capacity to tell whether you are even interacting with a human or not.

How would that play out in the metaverse?

How would that play out in the real world, over the phone?

How would you ever know whether anyone that you’re interacting with is real or not? Whether they are sentient?

The battle between AI and AI Detectors is not just a battle over the difference between an A grade and a C grade.

It’s a battle over a future where what’s at stake is identifying what even qualifies as human.

Can you write a Master’s thesis with ChatGPT?

Since the very dawn of time, students have sought new and creative ways to pass their exams that uh, do not include just studying.

People have hidden scraps of paper inside their pens, written down answers on their forearms, transcribed ancient Chinese texts onto underwear…

“pls stop” — every teacher throughout history

Even now, there is a disturbing number of articles on WikiHow about cheating, namely 10 Ways To Cheat on a Test Using Body Parts and even 3 Ways to Cheat on a Test Using Pens or Pencils… (WikiHow, why do you have so many of these????)

…And the list goes on!

Now, it’s no secret that it’s very possible to cheat with ChatGPT and that this has thrown educators worldwide for a loop, but then I received a rather funny question earlier on the Artificial Intelligence Megathread that I started on Lowyat.net.

Well.

Could you potentially write a Masters thesis with ChatGPT?

It so happened that someone on the ChatGPT Malaysia Facebook group had asked about the same thing, so I thought ok, let’s make it happen.

Anyway, I was curious about whether it was actually possible, so I decided to give it a go.

Here’s what I asked:

Okay, so at the very least the software proposed a bunch of topics that seemed kind of plausible and interesting.

Anyway, since I’m involved in the education industry and AI – based learning is very interesting to me, I decided to ask ChatGPT to follow up on #7, as follows:

Okay, wow! I had sources too! This was getting interesting! But then…

I looked at this, and I was captivated: Was I on my way to get a Master’s degree for this man?

No, wait. Wasn’t this even better? Wasn’t this thing essentially describing the process of creating a personalized new education technology company for me???

I set out in earnest, yearning to go where no man had ever gone before!

Okay, seemed great so far! I ran out of words, though, so I asked ChatGPT to continue:

Okay, uh…

Do you see what I’m seeing here?

Rather than actually writing the thesis, ChatGPT was malingering — it was casually not doing what it was told to do, and presenting me with some nonsense summary!

That won’t do! You think just because you’re an AI assistant you get to be lazy?!

I asked it to continue, and provide results in detail.

For about five, this seemed really really plausible, so I was happy again…

For about five minutes, before my skepticism began again.

…So I checked the references, only to realize that they mostly couldn’t be found anywhere.

Okay, I was thinking to myself.

This is a wonderful software, I declared, trying to beat back the cognitive dissonance.

Surely the third page will be a little bit better? So I thought.

At this point, I realized – ChatGPT had failed.

The two methodology sections contradicted themselves, and there wasn’t a possibility of reconciliation unless I proceeded to prompt chatGPT with the specific information that it actually needed, which I decided not to because the rewarded yielded by that effort would actually be better spent writing the thesis if I actually had a clear idea of how to do so.

So, how do we answer our research question?

With a solid no.

  1. As you can see, there’s a word limit for responses, which means that you will have to re-prompt ChatGPT, which is likely going to lead it to drift from the original prompt.
  2. ChatGPT’s memory for prior responses is about 4000 tokens (words) and it will not completely remember everything that you told it before unless say, you intelligently summarize.
  3. There is no guarantee that the logic or factuality of your piece will be valid or that even any of the sources that you cite will be accessible or even relevant to what you are writing about, as you see from the questionable sources.

Sorry to those of you out there hoping that ChatGPT was going to help you get your Master’s degree, but it’s not gonna happen right now.

Even if you can though, should you? I guess that’s up to each person to decide, but what I would say is that submitting something AI generated for a degree means that you didn’t get the degree — the AI did and got certified and you did not.

Let me not moralize this or romanticize education, but approach the matter in a logical way — when this starts to happen on a large scale, if it does happen, I can imagine that companies or other institutions that used to take these degrees seriously will simply no longer take them seriously, thereby causing degrees as a whole to become about as worthless as MOOCS among prominent companies (i.e. companies that actually generate large amounts of business and have a vested interest in hiring actually talented people) and leading to what we already see, to a degree, in institutions such as tech companies and start-ups… Whereby many of these companies don’t pay the most attention to the particular degree that you received, but rather whether you are capable of demonstrating the specific skills that they are looking for and communicating your perspective in the course of an interview in which there is no opportunity to make use of AI software.

How will artificial intelligence change not just education, but also the job market at large?

We’ll be finding out, and we’re going to be in for a wild, wild ride!

I’ll have lots more to say about this in the days ahead, so if you would like to read about the intersections between AI, writing, and education, do consider dropping me a follow and I’ll see you in my next pieces!

— V

How AI Tech Will Disrupt Businesses (24th February)

Do you ever feel like you might have gotten yourself in something a little bigger than you’d imagined was possible?

Excited to announce that I’ll be speaking about AI for the “How AI Tech Will Disrupt Businesses” panel on the 24th of February! Thank you Vulcan Post for the feature and MrMoney TV x Entrepreneurs and Startups Malaysia for the invitation!

You’ll be able to meet me there directly and hear me talk about the ways AI is going to change businesses around the world alongside my fellow panelist, Richard Ker!

If you’ve not heard of Richard, the man is a legend at creating incredible infographics and marketing, and I respect both his trite observations and the value that he’s created for literally thousands of people throughout Malaysia and far beyond; the man is a true blue digital authority, If you’re looking for something specific, feel free to check out this article that he’s written on Facebook, amongst other things; the man is everywhere!

In other words, what does this mean?

It means I need to level up!

This conference is something that I’m truly honored to be a part of, and a wonderful opportunity to learn from many incredible minds that I won’t be missing by any stretch of the imagination.

As we speak, I’m preparing for with all my might at the moment even as I read and learn more about artificial intelligence, building up that reading habit again thoughtfully documented by my dear friend Sandy Clarke and that I’ll make sure to work towards in the days ahead as I build this platform.

Meanwhile, if you haven’t already, please feel free to join Artificial Intelligence Malaysia! I’ve had some pretty wild conversations in the past day or so, and it would be great to add a diversity of voices to the group especially if you’re really interested in AI and everything that it has to offer 🙂

In preparation for that, know that I’ve been reading extensively and creating lots of other content as well because I know that anything I have to make this worth your time, and will do my very best to do so.

Till we meet, then!