Gold Penguin Logo with Text Challenges OpenAI's Claim That AI Detectors Don't Work came out with a challenge to OpenAI a few days after OpenAI claimed AI detection doesn't really work. Originality claims it does to a decent extent and is willing to donate to charity to prove how serious they are.
A fight between two robots in a boxing ring, as a cartoon in 4k with a lot of colors, made with Midjourney
A fight between two robots in a boxing ring, as a cartoon in 4k with a lot of colors, made with Midjourney
September 25, 2023 11:52 pm

In the ongoing debate surrounding the effectiveness of AI content detectors,, an AI detection company, recently threw some hands with OpenAI.

Their challenge: to prove whether AI detectors truly work or not. What's unique about this challenge is that it's not just about proving a point; it's for a charitable cause – take that as you wish.

OpenAI's Controversial Statement

The controversy began when OpenAI made a bold assertion that AI detectors, in their view, don't really work. This statement raised eyebrows in the AI community, as it cast doubt on the efficacy of tools designed to identify AI-generated text.

OpenAI's claim, however, was made without providing data or context to support it.'s Response: Charity, not one to shy away from a debate, responded with a challenge.

Their argument is that AI detectors do indeed work, albeit with some imperfections. They assert that the usefulness of AI detectors depends on the specific use case and that they can provide over 95% accuracy with a low false positive rate of under 5% in many scenarios.

The Challenge Details

The heart of's challenge lies in the creation of a new dataset containing both AI-generated and human-written text. This dataset will be subjected to the scrutiny of's AI detection system. Here's where the charity aspect comes into play:

  • If incorrectly identifies a piece of writing, they pledge to donate to charity.
  • Challengers who believe in OpenAI's assertion must donate for each correct prediction made by

This challenge not only adds a layer of excitement to the debate but also contributes to a charitable cause, with donations going to a mutually agreed-upon charity such as SickKids.

How AI Detectors Work and Their Limitations's statement also offers a glimpse into the inner workings of AI detectors. They explain that these tools use various detection models, such as "Bag of Word" Detectors, Zero-Shot LLM Approaches, and Fine-Tuned AI Models.

However, the statement acknowledges that their effectiveness can be limited, especially beyond newer large language model AI-generated content like GPT-4.

Emphasizing Data-Backed Claims

One of the key points in's statement is the importance of data-backed accuracy claims. They cite their own detector's performance on GPT-4 generated content, boasting an accuracy rate of over 99% with a mere 1.5% false positive rate. This is pretty bold, and some users may disagree based on their own use of the tool.

AI Detectors in Academia takes a clear stance on the use of AI detectors in academia. They recommend against using AI detectors for academic disciplinary actions, as these tools cannot provide the same level of proof as traditional plagiarism checkers. They claim their tool is built for content publishers – not schools.

The Ever-Changing Landscape of Bypassing AI Detectors

The statement also touches upon the evolving landscape of bypassing AI content detectors. What used to be effective methods for bypassing detection are no longer as potent due to improved detection techniques. It's a cat and mouse race. It's never going to end.

Understanding Detection Scores

A critical point of clarification is provided regarding detection scores. A score like 40% AI and 60% Original does not indicate the percentage of AI-generated content within a piece. Instead, it represents the detector's confidence in its prediction.

Closing Thoughts: Balancing Accuracy and Real-World Use

In essence, the challenge issued by Originality invites scrutiny and debate into the effectiveness of these so-called AI detectors. They recognize that while AI detectors aren't perfect, they can serve vital roles in many applications when used judiciously.

The debate not only raises questions about the future of AI detection but also underscores the importance of data-driven claims in the AI community.

It remains to be seen how OpenAI will respond and whether other players in the AI space will join in.

One thing is certain: the debate over AI detectors is far from settled, and the outcome of this challenge could have far-reaching implications for AI content detection in the years to come.

Stay Updated On The Latest AI Tools
Get notified with the biggest releases and tips in Artificial Intelligence delivered straight to your email. We don't sell your data.
Written by Justin Gluska
Justin is the founder of Gold Penguin, a business technology blog that helps people start, grow, and scale their business using AI. The world is changing and he believes it's best to make use of the new technology that is starting to change the world. If it can help you make more money or save you time, he'll write about it!
Notify of

1 Comment
Most Voted
Newest Oldest
Inline Feedbacks
View all comments