By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: This AI Writing Detector Shows Its Work. For Me, It's a Step in the Right Direction
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > This AI Writing Detector Shows Its Work. For Me, It's a Step in the Right Direction
News

This AI Writing Detector Shows Its Work. For Me, It's a Step in the Right Direction

News Room
Last updated: 2025/06/12 at 10:30 AM
News Room Published 12 June 2025
Share
SHARE

This article was written by an actual, flesh-and-blood human — me — but an increasing amount of the text and video content you come across online is not. It’s coming from generative AI tools, which have gotten pretty good at creating realistic-sounding text and natural-looking video. So, how do you sort out the human-made from the robotic?

The answer is more complicated than that urban legend about the overuse of em-dashes would have you believe. Lots of people write with an (over)abundance of that particular piece of punctuation, as any editor will tell you. The clues may have more to do with the phrasing and the fact that, as with any writer, large language models tend to repeat themselves.

AI Atlas

That’s the logic behind AI-detection programs. The problem is that those systems are often AI-powered themselves, and they provide few details about how they arrived at their assessments. That makes them hard to trust.

A new feature from the AI-detection company Copyleaks, called AI Logic, provides more insight into not just whether and how much of something might have been written by AI, but what evidence it’s basing that decision on. What results is something that looks a lot like a plagiarism detector, with individual passages highlighted. You can then see whether Copyleaks flagged it because it matched text on a website known to be AI-generated, or if it was a phrase that the company’s research has determined is far more likely to appear in AI-produced than human-written text.

You don’t even necessarily have to seek out a gen AI tool to produce text with one these days. Tech companies like Microsoft and Google are adding AI helpers to workplace apps, but it’s even showing up in dating apps. A survey from the Kinsey Institute and Match, which owns Tinder and Hinge, found that 26% of singles were using AI in dating, whether it’s to punch up profiles or come up with better lines. AI writing is inescapable, and there are times when you probably want to know whether a person actually wrote what you’re reading. 

This additional information from a Copyleaks-checked text marks a step forward in the search for a way to separate the AI-made from the human-written, but the important element still isn’t the software. It takes a human being to look at this data and figure out what’s a coincidence and what’s concerning.

“The idea is really to get to a point where there is no question mark, to provide as much evidence as we can,” Copyleaks CEO Alon Yamin told me.

A noble sentiment, but I also wanted to see for myself what the AI detector would detect and why.

How AI detection works

Copyleaks started out by using AI models to identify specific writing styles as a way to detect copyright infringement. When OpenAI’s ChatGPT burst on the scene in 2022, the company realized it could use the same models to detect the style of large language models. Yamin called it “AI versus AI,” in that models were trained to look for specific factors like the length of sentences, punctuation usage and specific phrases. (Disclosure: Ziff Davis, ‘s parent company, in April filed a lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.)

The problem with using AI to detect AI is that large language models are often a “black box” — they’ll produce an output that makes sense, and you know what went into training them, but they don’t show their work. Copyleaks’ AI Logic function tries to pull back the veil so people have a better sense of what in the copy they’re evaluating might actually be AI-written. 

“What’s really important is to have as much transparency around AI models [as possible], even internally,” Yamin said.

Read more: AI Essentials: 29 Ways to Make Gen AI Work for You, According to Our Experts

AI Logic uses two different approaches to identify text potentially written by an LLM. One, called AI Source Match, uses a database of AI-generated content from sources either created in-house by Copyleaks or on AI-produced sites online. This works much like a traditional plagiarism detector. “What we’ve discovered is that AI content, a lot of the time, if you ask the same question or a similar question over and over again, you’ll get similar answers or a similar version of the same answer,” Yamin said.

The other component, AI Phrases, detects terms and groups of words that Copyleaks’ research has determined are far more likely to be used by LLMs than by human writers. In one sample report, Copyleaks identified the phrase “with advancements in technology” as potentially AI-written. Copyleaks’ analysis of generated content found that the phrase appeared 125 times per million AI-written documents, compared with just six times per million documents written by people.

The question is, does it work?

Can Copyleaks spot AI content and explain why?

I ran a few documents through Copyleaks to see if AI Logic can identify what I know to be AI-created content, or if it flags human-written content as AI-written.

Example: A human-written classic

What better way to test an artificial intelligence tool than with a story about artificial intelligence? I asked Copyleaks to test a section of Isaac Asimov’s classic 1956 short story The Last Question, about a fictional artificial intelligence tasked with solving a difficult problem. Copyleaks successfully identified it as 100% matched text on the internet and 0% AI-written. 

Example: Partially AI-written

For this example, I asked ChatGPT to add two paragraphs of additional copy to a story I wrote and published earlier in the day. I ran the resulting text — my original story with the two AI-written paragraphs added at the bottom — through Copyleaks. 

Copyleaks successfully identified that 65.8% of this copy matched existing text (because it was literally an article already on the internet), but it didn’t pick up anything as being AI-generated. Those two paragraphs ChatGPT just wrote? Flew completely under the radar. 

A screenshot of Copyleak showing an article text as being written by AI.


Enlarge Image

A screenshot of Copyleak showing an article text as being written by AI.

Copyleaks thought everything in this article was written by AI, even though only a few paragraphs were.

Screenshot by Jon Reed/

I tried again, this time asking Google’s Gemini to add some copy to my existing story. Copyleaks again identified that 67.2% of the text matched what was online, but it also reported that 100% of the text may have been AI-generated. Even text I wrote was flagged, with some phrases, like “generative AI model,” described as occurring more frequently in AI-written text. 

Example: Totally AI-written

In a test of generative AI’s ability to create things that are totally out of touch with reality, I asked it to write a news story as if the Cincinnati Bengals had won the Super Bowl. (In this fictional universe, Cincinnati beat the San Francisco 49ers by a score of 31-17.) When I ran the fake story through Copyleaks, it successfully identified it as entirely AI-written. 

A screenshot of the Copyleaks interface showing why it thinks certain text was written by AI.


Enlarge Image

A screenshot of the Copyleaks interface showing why it thinks certain text was written by AI.

Copyleaks’ AI Logic quickly realized this story about the Cincinnati Bengals winning the Super Bowl was written by an AI chatbot.

Screenshot by Jon Reed/

What Copyleaks didn’t do, however, is explain why. It said no results were found in its AI Source Match or its AI Phrases, but with a note: “There is no specific phrase that indicates AI. However, other criteria suggest that this text was generated by AI.” 

I tried again, this time with a different ChatGPT-generated story about the Bengals winning the Super Bowl 27-24 over the 49ers, and Copyleaks provided a more detailed explanation. It calculated the content was 98.7% AI-created, with a handful of phrases singled out. These included some seemingly innocent terms like “made several critical” and “testament to years of.” It also included some strings of words that spread across multiple phrases or sentences, like “continues to evolve, the Bengals’ future,” which apparently occurred 317 times more frequently in the database’s AI-generated content than in human text documents. (After raising the issue with the first attempt with Copyleaks, I tried it again and got similar results to this second test.)

Just to be sure it wasn’t operating entirely on the fact that the Bengals have never won a Super Bowl, I asked ChatGPT to write an article about the Los Angeles Dodgers winning the World Series. Copyleaks found that 50.5% matched existing text online, but also reported it was 100% AI-written. 

A high-profile example

Copyleaks did some testing of its own, using a recent example of a controversial alleged use of AI. In May, the news outlet NOTUS said that a report from the Trump administration’s Make America Healthy Again Commission contained references to academic studies that did not exist. Researchers who were cited in the MAHA report told media outlets that they did not produce that work. Citations to nonexistent sources are a common result of AI hallucination, which is why it’s important to check anything an LLM cites. The Trump administration defended the report, with a spokesperson blaming “minor citation and formatting errors” and stating that the substance of the report remains unchanged. 

Copyleaks ran the report through its system, which reported finding 20.8% potential AI-written content. It found some sections around children’s mental health raised red flags in its AI Phrases database. Some phrases that occurred far more often in AI-written text included “impacts of social media on their” and “The Negative Impact of Social Media on Their Mental Health.”

Can an AI really detect AI-written text?

In my experience, the increased transparency from Copyleaks into how the tool works is a step forward for the world of AI detection, but this is still far from foolproof. There’s still a troubling risk of false positives. In my testing, sometimes words I had written just hours before (and I know AI didn’t play a role in them) could be flagged because of some of the phrasing. Still, Copyleaks was able to spot a bogus news article about a team that has never won a championship doing so. 

Yamin said the goal isn’t necessarily to be the ultimate source of truth but to provide people who need to assess whether and how AI has been used with tools to make better decisions. A human needs to be in the loop, but tools like Copyleaks can help with trust. 

“The idea in the end is to help humans in the process of evaluating content,” he said. “I think we’re in an age where content is everywhere, and it’s being produced more and more and faster than ever before. It’s getting harder to identify content that you can trust.”

Here’s my take: When using an AI detector, one way to have more confidence is to look specifically at what is being flagged as possibly AI-written. The occasional suspicious phrase may be, and likely is, innocent. After all, there are only so many different ways you can rearrange words — a compact phrase like “generative AI model” is pretty handy for us humans, same as for AI. But if it’s several whole paragraphs? That may be more troubling.

AI detectors, just like that rumor that the em dash is an AI tell, can have false positives. A tool that is still largely a black box will make mistakes, and that can be devastating for someone whose genuine writing was flagged through no fault of their own.

I asked Yamin how human writers can make sure their work isn’t caught in that trap. “Just do your thing,” he said. “Make sure you have your human touch.”

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Apple encryption row: Does law enforcement need to use Technical Capability Notices? | Computer Weekly
Next Article 3 Insights From Mavrck’s Creator Compensation Report |
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

Google is shutting down Android Instant Apps over ‘low’ usage
News
I live between Android, Windows, iPhone, and Mac, and it’s pure chaos
News
What’s next for Mac? Here’s what I predict is coming to the Mac line-up | Stuff
Gadget
Apple Music Replay finally stops acting like a separate app in iOS 26
News

You Might also Like

News

Google is shutting down Android Instant Apps over ‘low’ usage

2 Min Read
News

I live between Android, Windows, iPhone, and Mac, and it’s pure chaos

8 Min Read
News

Apple Music Replay finally stops acting like a separate app in iOS 26

3 Min Read
News

Costco slammed as policy forces shoppers to prove ID ‘more than once’

6 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?