How I taught my AI team to detect B.S
agents&me // Issue #6
From: Tom
A bench under the trees
Tuesday, late afternoon
Last Tuesday my Gatekeeper agent rejected three drafts before I even read them.
One scored 67 on a scale I invented. The scale measures bullshit.
I should back up.
---
It started with a sentence I couldn’t stop fixing
A few weeks ago my Copywriter agent delivered a draft. Technically fine. Structured well. But one paragraph made me wince:
“This game-changing approach will transform how you work, giving you the freedom to focus on what truly matters.”
You know that sentence. You’ve read it a hundred times on LinkedIn. It says absolutely nothing. No proof, no picture, no person behind it. Just air shaped like confidence.
I rewrote it. Moved on. Next day, another draft:
“It’s not just about saving time. It’s about working smarter.”
Same disease. Different symptom.
I could keep fixing these by hand. I’d been doing it for weeks, actually, the same em dashes, the same two-part template sentences, the same vague superlatives that sound important but prove nothing. Every review cycle, same patterns, same corrections.
And then I thought: wait. If I can describe exactly what’s wrong, can I teach an agent to catch it too?
So I built a BS detector
It’s a skill called `/bs`. You feed it any text, it scores it on a 1-100 “bullshit scale”:
- 1 = Authentic, specific, sounds like a real person
- 100 = Maximum bullshit, manipulative, or clearly AI-generated
The skill looks for specific patterns. Em dashes and template sentences. Vague claims without proof. Manipulative language and fake urgency. That generic “AI voice” we all recognize (and pretend we don’t use).
It doesn’t just flag problems. It quotes the exact sentences that smell off and explains why.
Now my Gatekeeper agent runs `/bs` on every piece of content before review.
The before/after that convinced me it works
Real example. Here’s a paragraph from a first draft, scored 67:
Before (Score 67):
“This game-changing approach will transform how you work, giving you the freedom to focus on what truly matters. It’s not just about saving time. It’s about working smarter and building systems that scale with your ambitions.”
The feedback: “Sounds like a LinkedIn influencer selling a course. Three empty superlatives. One em dash pattern. Zero specific numbers. Zero visual detail.”
I revised it. Same idea, different execution:
After (Score 23):
“I used to spend 4 hours reviewing AI drafts every week. Now the Gatekeeper catches the bad ones before I see them. Last Tuesday I reviewed two drafts instead of nine. I went for a walk.”
The feedback: “Sounds like Tom talking to a friend. Specific numbers (4 hours, two drafts, nine). Physical detail (went for a walk). Honest, no superlatives.”
That’s the shift. From words that sound important to words that show something.
But here’s what I didn’t expect
I built the detector as a filter. A gate. Something to catch bad output and bounce it back.
It did that. Fine.
But after about two weeks, something else happened. The first drafts started arriving cleaner.
Not perfect (I’m not sure they’ll ever be perfect, and I’m weirdly okay with that). But cleaner. Fewer em dashes. Fewer empty superlatives. More specific numbers. The Copywriter agent had internalized what gets rejected, so it stopped making those mistakes in the first place.
I didn’t train it to write better. I trained the Gatekeeper to reject worse. And the writing improved anyway.
This is the part that interests me more than the detector itself. Quality control isn’t just a filter. It’s a feedback loop. The standard got built into the workflow, and the workflow got better because the standard existed.
(I think there’s a management lesson in there somewhere, but I’m still working it out. Something about how clear rejection criteria teach faster than vague encouragement. Ask any writer who’s gotten a “nice but needs work” rejection versus a “this specific paragraph doesn’t earn its place” rejection.)
The irony, obviously
I’m using AI to catch AI. A machine judging whether other machines sound too machine-like.
I know.
💎 This week’s gem: the BS detector skill
The complete /bs skill I use to catch AI-sounding content.
What’s included:
- The full prompt with scoring rubric (1-100 scale)
- Complete checklist of patterns to detect
- Output format template
- Integration instructions for your Gatekeeper workflow
What you’ll be able to do:
- Score any text for BS before publishing
- Catch specific AI patterns automatically
- Build quality control into your workflow (not your review time)
🔒 Available to paid subscribers.
That’s it for this week.
If this was useful, forward it to someone (real human) building with AI.
Want the full BS Detector Skill? Subscribe for $15/month.
See you next week ✌️
-- Tom
(the guy whose AI now rejects his own writing for sounding too AI)
P.S. This newsletter was made almost entirely by my AI team. The BS detector scored this draft at 19. I’ll take it.
P.P.S. Want to build your own AI team? join me to the next online workshop. from zero to running in 2 hours. Details at getagents.today
P.P.P..S. I read every reply. The real me, not the AI.



This hit home. I catch myself rewriting the same empty “sounds confident but means nothing” sentences again and again. Turning that into a systematic check is such a smart move.