Voltar ao Blog
/5 min read

Everything That Mattered in AI This Year (And Everything That Didn't)

AIStrategy

2025 was deafening. Every week brought a new model launch, a new funding round, a new "this changes everything" announcement. If you followed AI news closely, you'd think the entire industry reinvented itself four times over.

Most of it was noise. Here's what actually mattered for teams building AI systems that ship to production. And what you can safely ignore heading into 2026.

What mattered

Reasoning models became production-ready

The single biggest technical shift this year. OpenAI's Thinking models, Anthropic's extended thinking, Google's Gemini 3 with deep reasoning. Models that plan before they answer. They break problems into steps, evaluate approaches, and check their own work.

This mattered because it moved the accuracy bar on complex tasks from "good enough for a demo" to "reliable enough for production." Code generation with constraints. Multi-step analysis. Planning tasks with competing requirements. Reasoning models handle these at a level that standard models simply can't.

The cost is 5 to 20x higher per request. Worth it for the hard 20% of tasks. Overkill for the easy 80%.

Open-source models caught up

Alibaba's Qwen3-Max. A trillion parameters. Open weights. Free to deploy on your own infrastructure. Two years ago, open models were a tier below proprietary ones on every benchmark. That gap closed this year. For most production tasks, open models deliver comparable quality with dramatically better economics at scale.

This matters because it gives companies a real choice. Self-host for cost control and data privacy. Use APIs for speed and convenience. The lock-in argument for proprietary models weakened significantly.

AI coding tools crossed the usefulness threshold

Claude Opus 4.5, GPT-5 Codex, Gemini Code Assist. These tools went from "interesting novelty" to "daily driver" for working engineers. Boilerplate generation, refactoring, test writing, code explanation. The mechanical parts of software engineering got automated to a genuinely useful degree.

The teams that integrated these tools into their workflows shipped faster. The ones still debating whether to adopt them fell behind.

Enterprise AI spending got serious (and so did the scrutiny)

Companies moved from AI pilot budgets to AI production budgets. Google launched Gemini Enterprise. Microsoft pushed Copilot across the 365 suite. Salesforce baked Einstein into everything.

But scrutiny followed the spending. Gartner's data on license underutilization made the rounds. CFOs started asking for ROI numbers instead of accepting "strategic AI investment" as an answer. This is healthy. The companies that measured their AI spend against actual productivity gains made better purchasing decisions.

IBM bet $11 billion on data infrastructure, not models

The Confluent acquisition was the clearest signal that the smart money understands where the real bottleneck lives. Not in model quality. In getting the right data to the model at the right time. Real-time event streaming as the foundation for AI systems that take actions, not just answer questions.

What didn't matter (as much as the hype suggested)

AI video generation

Sora launched. Runway pushed Gen-3. Every creative tool added video generation. The demos were stunning. The production applications were thin.

Video generation is still too slow, too expensive, and too inconsistent for most commercial use cases. You can't reliably generate brand-consistent video at scale. The editing workflows are primitive. The output needs heavy human curation.

This will matter eventually. In 2025, it generated more Twitter impressions than business value.

The model size wars

Every quarter brought a bigger model. More parameters. Higher benchmark scores. The marketing departments worked overtime: "our model is 2% better on MMLU."

Production teams stopped caring. The difference between the top five frontier models on most practical tasks is negligible. What matters is reliability, cost, latency, and how well the model integrates with your existing stack. The benchmarks that matter happen in your pipeline, not on a leaderboard.

AI hardware announcements

Every chip company announced an "AI-optimized" processor. Custom silicon. New architectures. Bold performance claims. Most of them won't ship meaningful volume until late 2026 or 2027.

If you're making infrastructure decisions today, the current generation of NVIDIA GPUs and cloud GPU instances are what's available. Plan with what exists, not what's been announced.

Autonomous AI agents (the marketing version)

"Fully autonomous AI agents that handle everything" was the marketing theme of the year. The reality: useful AI agents in production are heavily constrained, carefully scoped, and have humans in the loop for high-stakes decisions.

The autonomous agent vision is real. The timeline is longer than the demos suggest. Teams that shipped useful agents in 2025 did it by narrowing scope aggressively, not by pursuing full autonomy.

Heading into 2026

The pattern is clear. The useful advances in 2025 were infrastructure improvements. Better reasoning. Better open models. Better tooling for developers. Better data pipelines.

The noise was mostly about spectacle. Bigger models. Flashier demos. Bolder claims about autonomy.

If your AI strategy for 2026 focuses on infrastructure, integration, and measurable outcomes, you're building on what actually worked this year. If it focuses on chasing the newest model announcement every quarter, you'll spend a lot and ship little.

The companies that won with AI in 2025 were the ones that treated it as engineering, not magic. That's not changing in 2026.

Talvez goste de