You might know that AI helps people write cool and smart stuff fast.
But AI isn’t always right; it can make mistakes and sometimes copy others. That’s where tools to find AI-written texts help. These tools can tell if a piece was made by AI.
OpenAI tried but didn’t succeed in making a tool that could spot ChatGPT’s writing. Instead, they worked on a new idea. They thought about marking the text like leaving a special mark, which shows through word choice. This mark helps identify the writing later.
Yet, even with this new marking idea, they haven’t shared it with everyone. Here’s why:
We’ve got some updates on OpenAI’s Text Classifier to talk about.
OpenAI decided to stop using their text classifier because it wasn’t accurate enough, Techcrunch says. You can check out how well it did in our post about accuracy here: https://originality.ai/blog/ai-content-detection-accuracy.
This field is tough, and we really feel for OpenAI’s team. Knowing they made ChatGPT, we get the huge pressure they were under to make sure everything was right.
Looking at this from the outside, it’s clear that OpenAI, being a big name in AI tech and seen as the ones who started AI-made stuff, probably had to be extra careful. Since they made ChatGPT, getting things wrong and catching stuff that isn’t actually there would have been really tough for them to handle.
Making a tool that can spot text made by ChatGPT isn’t simple at all! There are lots of hard parts to figure out, and it’s important to understand that no solution will be perfect 100% of the time. Trying to be super accurate without making mistakes and causing problems is very tricky, especially for OpenAI more than others.
We believe that both transparency and accountability are crucial in this industry, and we commend OpenAI for making the tough call to shut it down when its performance fell short of the desired standard.
You might already know that AI tools help people create fun and smart content fast. But, since AI isn’t perfect, it can sometimes make mistakes like copying others or not being good quality. This is where tools that find AI-written stuff are useful.
These tools can spot if something was made by AI, even if it doesn’t copy from anyone. This is really helpful these days because Google doesn’t like it when websites have text that doesn’t sound real or was made by a machine.
Marketers, businesses, and educators all want to know if a text is written by AI. They use tools to check.
Today, we’re looking at the OpenAI Text Classifier. OpenAI made big AI models like ChatGPT and GPT-3. We’ll see how good their tool is compared to others.
OpenAI Text Classifier is a tool that makes finding important topics in big groups of documents or chats easy. It works really well and is simple to use, here’s how:
Remember, the OpenAI Text Checker doesn’t catch everything. It needs at least 1,000 characters to work, which is about 150-250 words.
Sadly, even though it can create text, it can’t spot copied work.
Here are the good and bad points about it:
Good Points:
Bad Points:
Let’s dive deep into this. We’re going to see how OpenAI’s Text Classifier and Originality.ai stack up in spotting AI-made texts. We used Jasper, a big deal in GPT2-powered writing, to create our samples.
Now, Originality.ai is really good at spotting stuff made by GPT2, GPT3, GPT3.5, or even ChatGPT. It’ll be cool to see which one of these, between OpenAI and Originality.ai, does better at finding content made entirely by AI.
The OpenAI Text Classifier uses a unique way to guess if text was made by AI. It decides by looking at chances – from “very unlikely” (under 10% chance) to “unlikely” (10%-45%), then “unclear” (45%-90%), next “possibly” (90%-98%), and finally “likely” (more than 98%).
Originality.ai does something different. It shows numbers that tell you how much of the text AI made and if any part is copied.
Let’s compare the OpenAI Text Classifier and Originality.ai side by side.
Right from the start, the OpenAI Text Classifier seems to miss the mark on spotting that our examples were written by AI.
OpenAI has openly shared that their detection tool doesn’t hit the mark much, finding AI-written text right only about 26% of the time in their tests. They also saw mistakes where it wrongly flagged texts 9% of the time. In our own check with seven examples, the tool got it wrong twice.
On the other hand, Originality.ai did a good job, correctly spotting AI content in five out of seven tries. They only got it wrong once and were unsure another time.
We need to understand that OpenAI checks if something is made by AI by being really sure about it. But Originality.ai sees if it’s right about something being made by AI. That’s why Originality.ai says it’s really good at this.
Even though they seem similar, OpenAI might not always seem as good. Not because it’s not smart, but because it’s extra careful.
OpenAI has said their Text Classifier should not be the only thing people use to make decisions. It should help with other ways of figuring out who wrote something.
Based on what we found, Originality.ai stands out as the best option for those who need a reliable way to tell if AI made the content.
We’re diving deep now. Let’s see how the OpenAI Text Classifier and Originality.ai do in spotting AI writings. Our examples come from Jasper, a strong tool for making content.
Originality.ai is great at finding stuff written by GPT2, GPT3, GPT3.5, and even ChatGPT. It’ll be cool to see how these two compare when checking content made entirely by AI.
The OpenAI Text Classifier uses a unique system to figure out if writing was done by AI. It looks at the chances, marking them as “very unlikely” if there’s less than a 10% chance, “unlikely” for 10%-45%, not sure for 45%-90%, “possibly” if it’s 90%-98%, and “likely” if it’s more than 98% likely AI did it.
Originality.ai works a bit differently. It gives you numbers showing how much of the text AI made and how much is copied from somewhere else.
Right from the start, the OpenAI Text Classifier couldn’t really tell if our examples were made by AI.
OpenAI has openly said that their tool for spotting AI-written text isn’t very good. It could only get it right 26% of the time when they tried it out. Also, it made mistakes 9% of the time, saying something was AI-written when it wasn’t. In our own check with seven examples, the tool got it totally wrong two times.
On the other hand, Originality.ai was able to spot AI-made content in five out of seven tries. It made one mistake and was unsure once.
OpenAI judges how accurate it is by checking if it’s really sure something was made by AI. On the other hand, Originality.ai counts how many times it’s right about AI-made content, which is why they say they’re so accurate.
Even though they’re similar, OpenAI might not seem as sharp. But it’s not about being worse; it’s more about being extra careful.
OpenAI has said that their Text Classifier isn’t the main tool to decide where text comes from. It’s more like a helper alongside other ways of finding out text sources.
Because of this, Originality.ai stands out to us as the clear pick for anyone who wants a reliable and accurate program to recognize AI-made content.
We looked closely at how OpenAI’s Text Classifiers compare with Originality.ai’s ability to spot AI-written text.
We’re going to see how our AI Detector (Originality.ai) matches up against OpenAI’s AI Text Classifier. (https://platform.openai.com/ai-text-classifier)
“OpenAI labels each document with terms like very unlikely, unlikely, unclear if it is, possibly, or likely AI-generated.”
We use two categories: AI and Human. OpenAI uses five, so we include the unsure ones (unclear if it is, possibly, likely) as AI. The others (very unlikely, unlikely) mean Human.
In the pictures below, the Y-Axis shows what actually is, and the X-Axis shows AI’s guess.
We see some key differences in how well they work:
If you want to find AI-made content quickly and easily, OpenAI is a good choice. It’s simple to use and gives you clear results fast.
Another good thing is that this tool comes from the people who made ChatGPT. So, it’s likely to get even better over time. Plus, it’s free, unlike some other tools that try to catch AI tricks.
Want a fast and simple way to find out if a text was made by AI? The OpenAI Text Classifier is a great pick. It uses a special version of the GPT model to figure out if AI, especially ChatGPT, created the text.
But if you’re after something even more dependable, try Originality.ai. This tool offers a pay-as-you-go plan, so it’s easy on your wallet. Plus, it’s super accurate and has cool features that make it stand out from the rest.
Through digital leadership we empower people to leverage the opportunities in global economy
@ 2024 Netus AI.