Shy Girl, Mob Rules, and the AI Detection Problem

The publishing world got its first real AI scalp this week, and the fallout is horrendous (and sad, honestly).

On March 19, Hachette Book Group cancelled the US release of Shy Girl, a horror novel by Mia Ballard, and discontinued the UK edition. The reason? Hachette suspected large portions of the book were AI-generated.

Ballard denies writing the book with AI. She told the New York Times that an unnamed acquaintance she'd hired to edit the original self-published version had used it without her knowledge.

I genuinely don't know if that is true in the sense that, how can anyone know? Unless someone confesses. But the way this has unfolded deserves a closer look because the mob mentality surrounding it, and some arguments that are being thrown around, are just as concerning.

How we got here

Ballard self-published Shy Girl in early 2025 on Kindle Unlimited. It quickly gained traction and got thousands of ratings on Goodreads. Hachette picked it up, published a UK edition in November 2025 through its Wildfire imprint, and had a US release scheduled for April 2026 through Orbit. While not everyone liked it, which is hardly uncommon, the early reviews were largely positive.

Then the tide turned. A Reddit post in January 2026 from someone claiming to be a book editor flagged the writing as having the hallmarks of large language model output (LLMs are a specific type of generative AI model). A YouTuber published a bloated (over two and a half hour) video analysing the text. It went viral with over 1.2 million views.

And that's when things really went off.

Bad reviews and AI accusations started piling up on Goodreads. But I’d love to know how many of those one-star reviews came from people who'd actually read the book before the controversy, and how many showed up after the Reddit threads and the YouTube detective work framed it as “AI slop”? I’m not saying it was all of them, but the answer still matters. Because we all know what happens when a narrative takes hold online…

The 78 per cent figure nobody is questioning

One of the most cited pieces of 'evidence' in this whole saga is the claim that Shy Girl is 78 per cent AI-generated. That number has been repeated in headline after headline. But almost nobody is talking about where it came from or why.

Max Spero is the founder and CEO of Pangram, an AI detection software company. After the Reddit post and the Frankie's Shelf video went viral in January 2026, Spero saw the online controversy, ran the book through his own product, and posted the results publicly on X. Nobody asked him to do it as far as I can see. Not Hachette or the New York Times. So it appears as though he inserted himself and his commercial product into a delicate, trending story, then published the dramatic number.

I'm not saying the test result is necessarily wrong (though AI detectors are known to underperform). But I am saying that a CEO running his own product on a viral controversy and publicly sharing the results doesn’t sound like an independent investigation.

Why has the obvious conflict of interest gone so unquestioned? Or have I missed something? The 78 per cent figure has been treated as near-fact, when it's one output from one commercial tool wielded by someone with a direct financial stake in the outcome. That concerns me.

Enter James Frey

Some commentators are raising the name James Frey. Let’s take a quick look at his story.

Author, Frey, was hugely shamed in the industry when his 2003 memoir A Million Little Pieces was exposed as partly fabricated. An investigation was published in January 2006 and Oprah Winfrey, who'd championed the book through her book club, hauled him back on her show and told him to his face that he'd betrayed millions of readers. He was dropped by his literary agent and his publisher settled a huge class-action lawsuit.

But Frey didn't just disappear. By late 2007 he had signed a large new deal with HarperCollins. He kept on writing and then in 2023, he openly discussed in an interview having used generative AI in a manuscript he was working on called FOURSEVENTYSIX. He said he'd included a footnote telling readers he'd used AI, that the AI had been instructed to mimic his style, and that he considered every word his own. That book was never published under that title.

Then, Next to Heaven came out in June 2025. His publisher told Book of the Month that Frey had written every word. Frey himself told the New York Times he hadn't used AI on it, though he'd experimented with it on an earlier project he'd abandoned. But by July 2025, only a month after publication, he had apparently confirmed AI was involved after all.

It’s a rollicking ride but now to the critical way it is distinct to Ballard’s.

The race argument

I've seen people on Bookstagram arguing that the difference in treatment between Frey and Ballard is a race issue. Ballard is a Black woman. Frey is white. Frey kept his career. Ballard lost her deal.

I understand why people reach for that framework and there is another black author who has received some accusations of AI use in Goodreads reviews. She’s also an Indie author who has recently got a publishing deal which sounds a bit suspicious, doesn’t it? But AI is mentioned in just 46 out of nearly 10,000 reviews she’s received, and it has not affected her contract. So, while I see concerns with it (for other reasons), I don't find it a compelling argument that this is about race at this point. The reason seems simpler: money.

Frey was already a bestseller when his first scandal broke. A Million Little Pieces had sold over five million copies. He had commercial leverage, and the fact is that would make publishers more willing to take another bet on you, despite the controversy. And even then, he had still been dropped by his agent, he’s still faced the class-action lawsuit. He was publicly humiliated on national TV. The idea that he sailed through unscathed just isn't true. He clawed his way back because he had an existing readership and enough capital built within the industry already, to get another deal.

Ballard was a debut self-published author, early career without the big money making track record. She didn't have the same buffer and ultimately publishing houses have to make enough money to stay afloat so, of course, some decisions will be financial.

I imagine when Hachette weighed the risk of standing behind her versus the cost of letting her go, it seemed easier to let her go. Brutal, and I’m not saying I like it, but I think it would apply to many new authors caught in a controversy like this.

Hellish social media and comments sections

Meanwhile, and sadly, Bookstagram is burning away. There are people asking if they should rip up their copies, and piling on with a confidence the evidence doesn't support.

One comment in particular stopped me cold. A woman argued that Ballard must have used AI on Shy Girl because, she claimed, the cover of Ballard's other book, Sugar, was AI-generated. But I think this commenter was confusing what happened with the original Shy Girl cover, where Ballard, unfortunately, used an image she wasn’t entitled to on the cover (she later apologised and fixed the situation). But even if the earlier novel’s cover was AI-generated, an AI book cover does not prove an AI-written book. The leap in logic there is as frightening as it is ridiculous.

I use generative AI to make images and short videos for Instagram to promote my novels sometimes (example below). That doesn't mean my novels were written with generative AI. If that logic held up, half the authors with a Canva account would be under investigation.

Assistive vs generative: the gap nobody wants to learn about

Another claim I saw today was that behind the scenes publishers are actively asking new authors to use AI. This wouldn’t surprise me but, if true, I’d assume they’re talking about assistive AI tools. Those are the ones that do grammar checks and finding repeated words or phrases in a manuscript. That's a world away from asking ChatGPT to write your chapters.

The distinction between assistive and generative AI is one I wish more readers, writers, and commentators would take the time to understand. I also wish writing organisations and publishers would come out to state explicitly where they stand on assistive AI tools (please🙏🏻). There seems to be a quiet acceptance of them, but the vagueness around it is awkward.

I wonder if they're scared, though, and that is relatable. Scared that defining the line too clearly will invite the same mob energy currently aimed at Ballard. But then, leaving the ruling on it so vague is creating confusion, and confusion can fuel accusations based on “vibes” rather than evidence. In the meantime, I think (not all, but) a majority of writers use the spell and grammar checking abilities of assistive AI products.

The inconvenient truth about AI detection

Now, maybe Shy Girl really was written mainly by generative AI. Maybe it was the editor. Maybe it was Ballard herself. But here's the uncomfortable reality: right now, I don’t believe there’s a reliable way to prove it without a confession.

I've tested AI detectors myself. I've run fully AI-written text through them, part-AI text, and 100 per cent human-written work. The results were all over the shop. Across multiple detection tools, I got correct results along with wildly wrong ones. Every category. Every tool I could find, including some paid versions. The technology isn't there and I honestly can’t tell you if it ever will be. But anyone claiming certainty based on an AI detection program should be met with scepticism at the very least.

Given that, and that the public surely does not know all of the details, I find what happened to Ballard horrendous.

A note for the 'we told you so' crowd

If it eventually comes out, through some actual admission, that Shy Girl was predominantly AI-generated and that Ballard was responsible, I suspect some people will read this post with huffs of 'we told you so' and 'it was obvious.'

But in the broader sense? The actual charge—of Shy Girl having been written with AI—is not obvious or “sure”. And what I've said here is worth saying regardless of the eventual outcome because this story raises questions about mob justice, flawed detection tools, the concern of someone with a commercial interest testing the work, and the lack of understanding many people have about the differences between assistive versus generative AI. And most of these issues won’t disappear any time soon.

So I'm not afraid to be wrong here. I just prefer to be cautious about this rather than reckless in my accusations of an early career author.

Morgan x

(Mia, should you read this I feel for you right now. Hang tight. Regardless of what happened, this will pass. x)

 

If you’re a writer, you might be interested in this post: AI Is Already Changing How You Write — Even If You Don't Use It

Next
Next

AI Use Statement