NOTCONTENT / training
Back to Blog
Methodology

The Eval Mindset: Stop Guessing Whether AI Output Is Good Enough

80-90% of serious AI work is evaluation, not prompting. If your team can't measure AI quality systematically, they're guessing. Here's how to stop.

Jeremy Somers
Jeremy SomersFounder, NotContent·Mar 17, 2026·5 min read

The Dirty Secret

Here's something the AI training industry doesn't tell you: 80-90% of the work in building reliable AI workflows isn't writing prompts. It's building evaluations.

Prompting gets all the attention because it's the visible part. You write a prompt, you get a response, it feels like progress. But the real question — "is this output actually good enough to use?" — is the question most teams never answer systematically.

They eyeball it. They "feel" whether it's right. They compare it to their mental model of quality and make a gut call. Sometimes the gut is good. Often it's not. And when it's not, you don't find out until the output is in front of a client.

The teams I train that get the best results are the ones that stop guessing and start measuring.

What an Eval Actually Is

An evaluation — an eval — is any systematic way of measuring whether AI output meets your standards. That sounds complicated. It's not. It ranges from dead simple to genuinely sophisticated, and the simple version works for most teams.

The simplest eval: Before you use AI for a task, define what "good" looks like. Write it down. Three to five criteria. Then evaluate the AI's output against those criteria every time. That's it. That's an eval.

For a creative brief, "good" might mean: (1) clearly states the strategic tension, (2) identifies the target audience with psychographic specificity, (3) includes a measurable objective, (4) is under 500 words. Now you can evaluate every AI-generated brief against those four criteria instead of vibing on whether it feels right.

The intermediate eval: Run the same prompt multiple times and compare outputs. AI is non-deterministic — it gives different answers each time. If you run a prompt five times and get five wildly different results, the prompt isn't reliable enough for production use. If you get five similar results that all meet your criteria, you've got a workflow you can trust.

The advanced eval: Use AI to evaluate AI. Build a second prompt that scores the output of the first prompt against your quality criteria. This sounds circular, but it works remarkably well for certain tasks. You're essentially building a quality control layer that runs automatically, flagging outputs that don't meet the bar before a human ever sees them.

Why Benchmarks Don't Matter

The AI industry loves benchmarks. "GPT-4 scores 90% on the bar exam." "Claude outperforms on graduate-level reasoning." These numbers are meaningless for your team.

Your team doesn't need an AI that passes the bar exam. They need an AI that produces creative briefs that meet your agency's standards, in your brand's voice, for your specific clients. No benchmark measures that. The only way to know if AI works for your use case is to build an eval for your use case.

This is why I tell every team I train: stop reading benchmark comparisons and start testing. Take a task your team does regularly. Define what "good" looks like. Run it through AI. Measure the output against your criteria. That 30-minute exercise will tell you more about AI's value to your team than every benchmark article combined.

The Eval Changes the Prompt

Here's what happens when teams start evaluating systematically: they get better at prompting without being taught.

When you have clear criteria for "good," you can see exactly where the AI output falls short. And when you can see where it falls short, you can adjust the prompt to address those specific gaps. The eval creates a feedback loop that improves everything upstream.

Team without evals: "The output isn't great. Let me try a different prompt." This is trial and error. It's slow, and you're not sure when you've arrived.

Team with evals: "The output scores 4/5 on strategic specificity but 2/5 on brand voice. I need to add more voice examples to the prompt and include a brand positioning statement in the context." This is systematic improvement. You know exactly what to fix and can measure whether the fix worked.

Building Evals Into Your Workflow

The teams that get the most out of AI build evaluation checkpoints into their workflow:

Pre-production eval. Before you rely on an AI workflow for real work, test it. Run it 10 times with different inputs. Score the outputs against your criteria. If 8 out of 10 meet the bar, the workflow is production-ready. If 5 out of 10 meet the bar, the prompt needs more work.

In-production eval. Every AI output gets a quick pass against the quality criteria. This doesn't mean reading every word — it means checking the criteria list. Does it meet the brief? Is the tone right? Are the facts verifiable? This takes 2-3 minutes and prevents the errors that take 2-3 days to fix.

Retrospective eval. Every month, review the AI outputs that made it to production. Which ones needed significant human editing? Which ones sailed through? This data tells you where your workflows are strong and where they need refinement.

The Quality Threshold Question

Every team reaches a point where they ask: "How good does AI output need to be for us to use it?"

The answer isn't 100%. The answer is: good enough that the human editing time is less than the time it would take to create from scratch. If AI produces a first draft that's 70% there, and your team can get it to 100% in 20 minutes, that's a win — as long as creating from scratch would have taken 2 hours.

The eval mindset helps you quantify this tradeoff instead of guessing at it. You know the AI hits 70% because you measured it. You know the editing takes 20 minutes because you tracked it. And you know the from-scratch alternative takes 2 hours because you benchmarked it.

Numbers replace vibes. That's the eval mindset.

Jeremy Somers

Jeremy Somers

Founder, NotContent

15 years as a creative director (Spotify, Nike, Pepsi, Samsung, Mercedes-Benz). Built the first AI-assisted creative agency in 2022.

See where your team stands

Take the 2-minute Readiness Scorecard and get a personalized program recommendation.

Take the Readiness Scorecard →