8/10/2025

Chasing Scores: The Real Reason AI Companies are Obsessed with Benchmarks (and Not You)

Ever get that feeling of AI whiplash? You read a headline about some new model that's shattered every record & is basically a genius in a box. Then you try to use an AI tool to do something simple, like summarize an email thread, & it completely misses the point or just makes stuff up.
If you've felt that disconnect, you're not crazy. There's a HUGE gap between how AI companies measure success & what you, the user, feel as success. And it all comes down to their obsession with one thing: benchmarks.
Here’s the thing, for years, the entire AI industry has been running on a treadmill, chasing higher & higher scores on standardized tests. It's a race to get the best numbers on leaderboards with names like GLUE, SuperGLUE, & MMLU. But as we're all starting to realize, a straight-A student on paper isn't always the person you'd want to actually work with. Turns out, the same is true for AI.
Let's break down why this happens, what the real cost is, & how things are FINALLY starting to change.

What Even Are Benchmarks & Why Do They Exist?

Okay, so let's be fair. Benchmarks aren't inherently evil. They were created for some pretty good reasons. Think of them like the SATs for AI models.
Back in the day, before these standardized tests, it was kind of the wild west. Everyone was building their own AI, but it was almost impossible to tell if one was actually better than another. It was all apples & oranges. The history of AI evaluation started with conceptual ideas like the Turing Test—could a machine trick a human into thinking it was also human? Then it moved into testing AI in very narrow domains, like playing checkers or being an "expert system" that could diagnose diseases based on a set of programmed rules.
But as AI got more complex, especially with natural language understanding (NLU), the community needed a way to measure progress on a level playing field. That’s where benchmarks like the General Language Understanding Evaluation (GLUE) came in. Introduced in 2018, GLUE was a collection of nine different tasks—like sentiment analysis & grammar checking—that could be used to spit out a single score.
For the first time, researchers had a clear target. It was a standardized way to say, "My model is better than your model because it scored higher on GLUE." And honestly? It worked. It galvanized the entire industry, leading to rapid improvements & pushing the boundaries of what was possible. It was a necessary step to get us where we are today.

The Benchmark Treadmill: When a Good Idea Goes Wrong

The problem is, the industry didn't just use benchmarks as a helpful tool. They became the only thing that mattered. And when you're ONLY focused on the final score, you start to do some weird things.

"Teaching to the Test"

Remember that kid in school who didn't really understand the subject but was amazing at memorizing just enough to ace the test? That's what started happening to AI models. This is what researchers call "overfitting" or "saturation." The models got SO good at the specific, narrow tasks on the benchmark that they started to lose the ability to generalize to situations they hadn't seen before.
The GLUE benchmark got so saturated with high-scoring models that it basically became obsolete. So, they had to create a harder test: SuperGLUE. But it's just a new version of the same game. Labs are still optimizing for the test, not for genuine understanding. As one research paper put it, benchmarks aren't just passively measuring AI; they are "deeply political, performative and generative," meaning they actively shape the AI that gets built, for better or for worse.

Real Life is Messy

The bigger issue is that these benchmarks are a sanitized, laboratory version of the real world. They don't account for the beautiful, chaotic messiness of human communication. They don't include:
  • Sarcasm & irony
  • Typos & weird grammar
  • Cultural context & slang
  • Ambiguous questions with no single right answer
  • Users changing their minds halfway through a sentence
So you get a model that can ace a reading comprehension test but completely falls apart when a customer asks, "Hey, I bought that thingy from you guys last week, the blue one, & it's making a weird noise. What do I do?" The benchmark score is 99/100, but the user experience is 0/100.

The Hype is REAL (and It's Funded)

So if benchmarks are so flawed, why is everyone still so obsessed with them? Two words: money & media.
There is "massive top-down pressure" from investors & corporate boards who are all asking their CEOs, "What's our AI strategy?" The easiest, most digestible way to show progress is to point to a leaderboard. "Our model just beat Google's on the SuperGLUE benchmark!" sounds a lot better in a press release or a board meeting than, "We're making slow but steady progress on making our chatbot slightly less annoying."
The media and analysts fuel this with things like the Gartner Hype Cycle, which tracks emerging technologies through a phase of "Inflated Expectations." AI is currently at the peak of that hype, and companies are racing to seem like they're at the forefront, often by treating AI like a "shiny object" rather than a strategic tool. This creates a powerful incentive to focus on the metric that gets you good press & your next round of funding, which is the benchmark score. User satisfaction is, unfortunately, an afterthought.

The Messy, Human Problem of "User Satisfaction"

This brings us to the other side of the coin: Why don't companies just focus on making users happy instead? Because it's REALLY, REALLY hard.
Unlike a benchmark, user satisfaction is subjective, emotional, & changes from person to person. Traditional methods for measuring it, like surveys, just aren't cut out for the complexities of AI. How do you measure a user's satisfaction with an AI's "language intelligence" or "contextual awareness"? It's not a simple 1-to-5 scale.
This is where you see the disconnect in the real world. We've all seen the stories:
  • Air Canada's chatbot famously made up a bereavement fare policy. When the customer tried to get the airline to honor it, Air Canada's legal team argued in court that the chatbot was a "separate legal entity" & was responsible for its own actions. They lost, by the way.
  • DPD, the delivery company, had its AI chatbot swear at a frustrated customer & then compose a poem about how terrible DPD was as a company.
  • A Chevy dealership's chatbot was tricked by a user into agreeing to sell a brand new Chevy Tahoe for $1.
In every one of these cases, the underlying AI model might have been a benchmark champion. But in the moment that mattered—the actual interaction with a human—it was a catastrophic failure. These aren't just funny anecdotes; they're examples of what happens when the focus is entirely on technical capability & not on user-centric design & safety. We saw it with the disastrous Willy Wonka Experience, where AI-generated marketing created a magical expectation that reality couldn't even remotely match.
The real tragedy was IBM's "Watson for Oncology." It was hyped as an AI that would help cure cancer, but it was quietly shelved after it was found to be giving unsafe & incorrect treatment recommendations. It was a powerful lesson that real-world application is infinitely more complex than a controlled environment.

The Shift to What Actually Matters: Real-World Value

Slowly but surely, the industry is waking up from its benchmark-induced hangover. As powerful AI models become commoditized—meaning anyone can access a state-of-the-art model through an API—the game is changing. The new competitive advantage isn't about having the highest score anymore. It's about how you use the AI to create a genuinely helpful & satisfying experience.
This is where the focus has to shift. Instead of just chasing benchmark scores, smart companies are realizing the IMMENSE power of training AI on their own, specific data. This is precisely the philosophy behind a new wave of tools. For example, a platform like Arsturn helps businesses build no-code AI chatbots that are trained on their specific website content, product catalogs, & support documentation.
The result? The AI doesn't just know a generic answer to "how do I request a refund?" It knows the exact refund policy for the product you bought, your purchase date, & can even start the process for you. It's the difference between an AI that's a "know-it-all" & an AI that's a genuine expert on one thing: your business. That's a complete game-changer for actual user satisfaction.

Human-in-the-Loop: The Best of Both Worlds

The most promising path forward is a concept called "Human-in-the-Loop" (HITL) evaluation. This isn't about getting rid of AI, but about creating a partnership between humans & machines.
In a HITL system, humans are involved at every stage:
  1. Training: Humans provide high-quality, nuanced data to train the model, going beyond simple labels.
  2. Evaluation: Instead of just checking if an answer is "correct," human evaluators assess it for tone, helpfulness, safety, & brand voice. They use their judgment in ways a machine can't.
  3. Correction: When the AI gets something wrong in the real world, that interaction is fed back into the system with human correction, so the AI learns from its mistakes continuously.
This creates a constant feedback loop that refines the AI over time, making it not just smarter on paper, but more helpful in practice. It requires a commitment to building a system with clear guidelines, a diverse group of human evaluators to avoid bias, & scalable feedback mechanisms.
Honestly, this is where the rubber meets the road. For any business, the ultimate benchmark isn't some abstract score on a test created in a lab. It's whether the AI is actually working. Is it reducing the number of support tickets? Is it generating more qualified leads? Is it helping customers find what they need faster, so they leave your website feeling happy & not frustrated?
This is the kind of practical, results-driven AI that solutions like Arsturn are built for. By helping businesses build meaningful connections with their audience through personalized chatbots, the focus shifts from a generic score to tangible business value. The goal becomes creating an AI that provides instant customer support & engages with website visitors 24/7 in a way that's genuinely helpful, boosting conversions & building trust. That's a benchmark worth chasing.
It's a pretty interesting shift we're seeing. The era of chasing leaderboard glory is slowly giving way to a more mature, user-centric approach to AI. Companies are realizing that the smartest AI isn't the one with the highest test score, but the one that makes people's lives just a little bit easier.
Hope this breakdown was helpful & gives you a new way to look at the next big AI announcement. Let me know what you think.

Copyright © Arsturn 2025