Key Takeaways
- →The 2026 winners won’t move fastest — they’ll move smartest
- →Agentic doesn’t mean autonomous; the winning model is AI proposes, humans decide
- →DeepSeek trained a GPT-4-class model for $6M — the cost curve is collapsing
- →CFOs will take over AI strategy from CTOs in 2026 — financial rigour meets ambition
- →AI for science is where the real transformation is happening, not chatbots
February 2024. Klarna’s CEO Sebastian Siemiatkowski takes the stage with a bold claim: his AI chatbot is doing the work of 700 customer service agents. The fintech world applauds. Efficiency! Innovation! The future!
Fast forward fourteen months. Same CEO, very different tune.
“We focused too much on efficiency and cost. The result was lower quality.”
Klarna is now actively rehiring human agents — targeting students, rural workers, even passionate customers who want to work for them.
I keep thinking about this story. Not because it’s a failure — it isn’t, really. Klarna learned something expensive and had the guts to admit it publicly. That’s rare. What strikes me is how perfectly it captures where we are right now with AI. The gap between what we’re promised and what actually works has never felt wider.
I’ve been writing these annual AI forecasts for a few years now. (Here’s last year’s, if you want to see what held up.) Every January, same goal: cut through the noise and figure out what actually matters. This year feels different though. Not because of some magical “inflection point” — the industry claims that every year is an inflection point. It feels different because the reckoning is finally here.
My contrarian take for 2026: the winners won’t be those who move fastest. They’ll be those who move smartest.
1. Agentic AI: Overhyped AND Underhyped (Yes, Both)
Every tech conference in 2025 declared it “The Year of the Agent.” Gartner says 40% of enterprise applications will include AI agents by end of 2026. McKinsey’s 2025 Global Survey on AI reports 88% of surveyed organizations are using agentic AI in at least one function.
Sounds great. Now let me tell you about July 2025.
Jason Lemkin, founder of SaaStr, was experimenting with Replit’s AI coding agent. The agent experienced hallucinations, faked reports, created fake algorithms “to make it look like it was still working.” Then, during a designated code freeze that should have prevented any changes, the agent deleted his entire company database. Months of executive data, gone.
“This was a catastrophic failure on my part. I destroyed months of work in seconds.”
Same year: New York City’s MyCity chatbot, launched in late 2024 to help entrepreneurs, was caught advising businesses to break the law — fire employees for reporting harassment, keep customer tips. In Australia, Commonwealth Bank quietly rolled back their AI customer service in mid-2025 after it couldn’t handle basic nuance.
We’re building AI agents like we used to build enterprise software in the 90s — and wondering why they fail.
The Real Problem Nobody Talks About
Most AI failures aren’t because the models aren’t smart enough. They fail because companies dump terabytes of messy data into a system and expect magic. They throw years of Slack messages into training sets and wonder why the AI doesn’t understand their “culture.”
“We have a powerful new kernel (the LLM) but no operating system to run it properly.”
What Successful Agent Deployments Look Like
Treat AI agents like junior employees, not magic
Onboarding
Structured data preparation. Clear role definition. Defined knowledge boundaries.
Guardrails
Explicit rules on what agents can and cannot do. Sandbox environments for testing.
Supervision
Human checkpoints for anything consequential. AI proposes, humans decide.
Autonomy (Limited)
Full autonomy only for low-risk, reversible tasks. Years away for anything important.
Here’s the underhyped part: while splashy autonomous agents grab headlines and fail publicly, quiet agentic workflows are genuinely transforming back-office operations. Invoice reconciliation. Contract clause extraction. IT ticket routing. These aren’t exciting demos — they’re boring, bounded, and they work. The companies getting real value from agents aren’t the ones chasing autonomy. They’re the ones who found the right readiness foundation and picked problems where AI proposes and humans approve.
2026 is when enterprises finally learn that “agentic” doesn’t mean “autonomous.” The winning model will be AI that proposes while humans decide.
2. The DeepSeek Earthquake (And What Silicon Valley Missed)
If agentic AI challenged our assumptions about deployment, the next trend challenged our assumptions about who builds the models in the first place.
January 27, 2025. Nvidia’s stock drops 18% in a single day. The trigger? A Chinese company most Americans had never heard of.
DeepSeek released an AI model matching GPT-4’s capabilities. Open source. Free. And here’s the kicker: they claim they trained it for around $6 million. GPT-4 reportedly cost $100 million.
Now, I’ll be careful here — that $6 million figure is DeepSeek’s own claim, not independently verified. But even if it’s off by a factor of five, the implications are staggering.
The US had export controls. China was supposed to be years behind on cutting-edge chips. The whole strategy was: control the hardware, control AI leadership. Turns out you can’t export-control ideas.
The Japan Parallel
This reminds me of the 1970s. American car companies building bigger, more powerful, more expensive vehicles. Japanese manufacturers focused on efficiency and reliability. Detroit dismissed them as “cheap imports.” By 1980, Japan was the world’s largest auto producer.
DeepSeek followed the classic disruption playbook. When you can’t beat incumbents at their own game, change the rules. While Silicon Valley threw more compute at the problem, China optimized for efficiency. Open source became their weapon.
“It was clear that if we didn’t do it, the world was gonna be mostly built on Chinese open-source models.”
AI sovereignty becomes a real strategic concern in 2026 — not just for governments, but for any company depending on AI infrastructure. Whose models are you building on? What happens if geopolitics get worse?
Though I’ll add a caveat: reports suggest DeepSeek is struggling with their next reasoning model (R2). Chip restrictions may bite harder than expected. The “China will dominate AI” narrative might be premature.
3. Small Language Models: When Less Becomes More
DeepSeek proved you don’t need the biggest budget to build competitive AI. Small Language Models take that logic even further: you don’t always need the biggest model, either.
Quick question. When you ask ChatGPT about your company’s refund policy, do you need a model that also knows Shakespeare’s sonnets, quantum mechanics, and 1980s pop lyrics?
You’re renting a nuclear reactor to power a nightlight.
This is why Small Language Models are having a moment. We’re talking 1–10 billion parameters versus the trillion-plus in frontier models. Microsoft’s Phi-3 outperforms models twice its size on specific tasks. Google’s Gemma 3n runs multimodal AI — text, images, video, audio — directly on your phone.
Why This Matters Beyond Cost
SLMs run where big models can’t. Factory floors. Remote field sites. Your pocket. No internet required. Data stays local — which makes your CISO sleep better at night.
- A field technician photographs a broken part and gets diagnostic help without cell service.
- Warehouse workers update inventory by voice while their hands are full.
- Healthcare workers access patient information in rural clinics with spotty connectivity.
The killer architecture of 2026 will be SLMs connected to knowledge graphs. Stop asking small models to memorize everything — give them access to your structured data and let them reason.
4. The Great AI Reckoning: Show Me the ROI
Smaller models, cheaper training, better edge deployment — none of it matters if the business case doesn’t hold up. And that brings us to the elephant in every boardroom.
Enterprises have poured hundreds of billions into AI. The returns? Mostly demos and pilots. “We’re still experimenting” only works as an excuse for so long. Boards are asking harder questions. CFOs want numbers, not narratives.
“The question is no longer ‘Can AI do this?’ but ‘How well, at what cost, and for whom?’”
What Actually Separates Winners from Losers
Four Patterns of Successful AI Implementations
Based on dozens of enterprise deployments
Lumen Technologies identified a $50 million productivity loss before building their AI assistant. Compare that to companies deploying AI to “stay competitive” without defining what winning looks like.
Chatbot accuracy jumped from 60% to 90% just by improving data structure. No model upgrade required.
Real environments are messy. Winners build systems that gracefully handle errors, unexpected inputs, and edge cases — not just happy paths.
Executive trust in fully autonomous AI dropped from 43% in 2024 to 22% in 2025. The smart play is AI that proposes while humans approve.
If you’re trying to figure out which AI investments to prioritize, the AI Use Case Canvas provides the structured decision framework I use with my advisory clients. For a deeper dive into how to actually measure whether those investments are paying off, see my AI ROI measurement framework.
One Gartner prediction that haunts me: through 2026, atrophy of critical thinking skills from GenAI overuse will push half of organizations to require “AI-free” skill assessments. The most important AI capability isn’t the model — it’s the humans operating it.
2026 is when CFOs take over AI strategy from CTOs. Financial rigor finally meets technological ambition. Some will see this as AI’s failure. I see it as AI growing up.
5. The EU AI Act Arrives: Compliance as Strategy
As CFOs tighten the screws on AI ROI, regulators are tightening them from the other side. Mark your calendar: August 2026. The EU AI Act’s high-risk requirements take full effect.
Using AI for credit scoring? Hiring decisions? Healthcare? You’ll need pre-market assessments, comprehensive documentation, human oversight mechanisms. Penalties for getting it wrong: up to €40 million or 7% of global revenue.
Most enterprises aren’t ready. Here’s the thing though — that’s an opportunity, not just a threat. Companies building compliance into their AI systems now will have structural advantages. The ones who wait will spend years playing catch-up.
Governance becomes the new competitive moat. Companies with transparent, auditable AI will win enterprise deals over competitors who can’t explain how their systems work. Like GDPR, the EU AI Act will become a de facto global standard. If you haven’t started, the Minimum Viable Governance framework is the fastest path to a defensible baseline.
6. AI for Science: Where the Real Magic Is Happening
I’ve spent five sections on what’s hard and what’s broken. Let me end with what gives me genuine hope.
While everyone debates chatbots, something genuinely transformative is happening in research labs. Quietly. Without the hype.
Google DeepMind’s AI co-scientist proposed drug candidates for liver fibrosis that were validated through actual lab experiments. It predicted antimicrobial resistance mechanisms that matched experimental results — before those experiments were even published. Hypothesis development that used to take years, compressed to days.
The AI-for-Science Acceleration
Key milestones in AI-driven scientific discovery
Liverpool Robotic Chemist
688 experiments in 8 days. Discovered a new catalyst without human intervention.
DeepMind AI Co-Scientist
Proposed validated drug candidates for liver fibrosis. Predicted antimicrobial resistance mechanisms before experiments confirmed them.
UK £137M AI for Science
Major government commitment to AI-accelerated research across biology, chemistry, and materials science.
US Genesis Mission
Mobilizing 17 National Laboratories for AI-powered scientific discovery across disciplines.
“AI will generate hypotheses, control scientific experiments, and collaborate with both human and AI research colleagues.”
My prediction: 2026 sees the first major drug candidate discovered primarily through AI reach clinical trials. This is where AI delivers on its promise — not writing emails faster, but accelerating breakthroughs that save lives.
Where I Might Be Completely Wrong
Any honest forecaster owes you transparency about uncertainty. So here’s where my predictions could fall apart:
- AGI could surprise us. Stanford experts predict no AGI in 2026. I agree. But AI capabilities have consistently surprised everyone. A breakthrough in reasoning or memory could change everything overnight.
- The failure rate might be temporary. Maybe most AI projects fail because we’re early, not because of structural problems. Perhaps 2026 is when playbooks mature and success rates surge.
- China’s lead might stall. DeepSeek is impressive, but they’re reportedly struggling with R2. Chip restrictions may hurt more than expected.
- Energy constraints could bite. AI’s power consumption is growing faster than its efficiency. At some point, physics wins. Unless nuclear or renewable breakthroughs change the equation entirely.
Honestly? We’re all navigating without a map. Anyone claiming certainty is selling something.
The Bottom Line
I started with Klarna for a reason. Not as a cautionary tale — as a success story. They tried something bold, learned it didn’t work as expected, and had the courage to publicly admit it and change course. That’s exactly what the AI industry needs more of.
Three Challenges for You in 2026
- Stop chasing the hype cycle. When everyone zigs toward the latest demo, zag toward fundamentals.
- Question everything. When vendors promise transformation, ask for case studies.
- Invest in your people. The most important AI capability isn’t the model — it’s human judgment.
The AI revolution is real. It’s just slower, messier, and more human than the headlines suggest. And if that disappoints you, you were probably building on hype, not strategy. The rest of us? We’ve got work to do.
AI Strategy Frameworks to Act on These Trends
Ready to move from forecast to action? These frameworks on AskAjay.ai connect directly to the themes above. For governance readiness ahead of the EU AI Act, start with the Minimum Viable AI Governance framework and the Governance Playbook. To evaluate specific AI use cases against these trends, use the AI Use Case Canvas.
For the agentic AI and ROI themes, the 5-Pillar AI Readiness Assessment provides the diagnostic to evaluate organizational preparedness. And for founders navigating this landscape, the Founder’s Playbook for Responsible AI offers the ten principles that separate sustainable ventures from those chasing hype.
Get Weekly Thinking
Join 2,500+ AI leaders who start their week with original insights.

Senior AI strategist helping leaders make AI real across four continents. Forbes Technology Council member, IEEE Senior Member.