← All research

The Shift to Agent-First AI: What Together.ai and Fireworks.ai Model Changes Tell Us

April 23, 2026

CS
Colin Smillie

Founder, Developer, AI Researcher

AI inference platforms are deprecating mid-tier chat models and replacing them with agent-first architectures built on Mixture of Experts. The shift reflects a fundamental change in what AI systems are optimized for: not answering questions, but executing multi-step workflows with tool use, planning, and action. Platforms like Together.ai and Fireworks.ai are converging on MoE plus agent capability as the baseline for production inference.

Over the past few weeks, something important has been happening across AI infrastructure platforms like Together AI and Fireworks AI.

Models are being deprecated. Replaced. Consolidated.

At first glance, it looks like routine platform churn.

It isn't.

It's a signal.

The models disappearing

Across platforms, a consistent set of models is being removed or deprioritized:

  • DeepSeek v3.2
  • Qwen3 8B
  • Qwen3 VL 30B
  • GLM 4.x and early 5 versions
  • Llama 3.3 70B

These models share a common profile: mid-tier capability, earlier-generation reasoning, limited agent functionality.

They were strong chat models. They are weaker agent models.

And that distinction now matters more than anything else.

What's replacing them

The replacements are models like Kimi K2.6, Qwen3.6 Plus, GLM 5.1, and the GPT-OSS 20B and 120B class.

These aren't better versions of the same thing. They represent a different design philosophy:

AI is no longer optimized for answering questions. It's optimized for doing work.

The rise of agent-first architecture

Modern models are increasingly built around a simple idea: the model is not the endpoint. It's the orchestrator.

This shift shows up in three core capabilities.

Tool use

Models are expected to call APIs, execute code, and retrieve data.

Multi-step reasoning

Not “answer this question,” but break down tasks, plan steps, and iterate toward a result.

Execution capability

The output isn't just text. It's actions, workflows, and decisions.

That is what defines an agent-first model. And it's why older models are being phased out.

MoE dominance: the architecture behind the shift

At the center of this transition is one idea: Mixture of Experts (MoE).

Instead of one monolithic neural network, MoE models are built from many smaller “experts.”

The 1,000 employee analogy

Imagine a company with 1,000 employees, but only 5 to 10 people work on any given task. You still have massive bench strength and deep specialization. But you only use the most relevant team for the job at hand.

How it works

  • A router decides which experts to activate
  • Only a subset of the model runs per token
  • Different inputs may trigger different experts

Why MoE is winning

Performance per dollar. Massive total capacity, lower compute per request.

Specialization. Experts can focus on coding, reasoning, or language nuance.

Better scaling. You can grow total model size without proportionally increasing cost.

Tradeoffs

MoE is not free. It's harder to train, routing decisions are complex, and load balancing is a real engineering challenge.

A subtle but important effect

Because different experts can be selected on different runs, the same model can produce different results across runs. That introduces variability and divergence in reasoning paths. For anyone evaluating AI systems, that becomes incredibly important.

Why older models are being removed

Now the earlier list makes sense.

The deprecated models don't use MoE effectively. They lack strong routing and specialization. They struggle with long-horizon reasoning. They aren't designed for agent workflows.

They're optimized for answers. Not execution.

Platform strategy: Together vs Fireworks

These shifts also reveal how the platforms are positioning themselves.

Together AI

Broad, fast-moving model catalog. Rapid adoption of new releases. High experimentation velocity. Think of it as the discovery layer for emerging models.

Fireworks AI

Curated, performance-focused selection. Aggressive deprecation of weaker models. Standardizing around fewer, stronger options. Think of it as the production-ready inference layer.

Both are converging on the same conclusion: MoE plus agent-first models are the future.

A different path: OVHcloud

While Together and Fireworks compete on which models to run, OVHcloud is taking a different approach.

Infrastructure first

OVH focuses on GPU infrastructure, sovereign cloud environments, and private deployments.

Model-agnostic strategy

Instead of curating models, they enable Hugging Face ecosystems, custom deployments, and enterprise-controlled AI stacks.

Why this matters

OVH isn't competing on model quality. They're competing on where and how models run. That matters most for regulated industries, data sovereignty requirements, and serious enterprise AI adoption.

The real takeaway

This isn't a model refresh cycle. It's a shift in what AI is.

Then: chat-first, static responses, one-shot answers.

Now: agent-first, tool-using, multi-step, execution-driven.

And under the hood, MoE architectures are enabling massive scale, efficient compute, and specialized reasoning all at once.

What to watch next

If this trend continues, and it will:

  • Smaller models will disappear from serious workflows
  • Agent capability will become the default expectation
  • Model evaluation will shift from “accuracy” to “task completion and reliability”

The shift is already underway. Chat-era models are being quietly retired, and agent-era models are taking their place.

Frequently asked questions

What is an agent-first AI model?

An agent-first AI model is designed to execute multi-step tasks rather than produce single-turn answers. These models are built around tool use, planning, and action. They can call APIs, execute code, retrieve data, and iterate toward a result. The shift from chat-first to agent-first reflects a broader change in how AI platforms evaluate and deploy models.

What is Mixture of Experts (MoE) in AI?

Mixture of Experts is a neural network architecture where the model is composed of many smaller specialized sub-networks called experts. A routing mechanism selects which experts to activate for each input, so only a fraction of the total model runs per request. This provides large model capacity at lower compute cost per token, enabling better performance per dollar.

Why are AI inference platforms deprecating older models?

Platforms like Together.ai and Fireworks.ai are removing models that lack strong agent capabilities. Models like DeepSeek v3.2, Qwen3 8B, and Llama 3.3 70B were effective chat models but struggle with tool use, multi-step reasoning, and long-horizon execution. The replacements, such as Kimi K2.6 and Qwen3.6 Plus, are designed for agent workflows from the ground up.

What is the difference between Together.ai and Fireworks.ai?

Together.ai operates as a broad discovery layer with rapid adoption of new model releases and high experimentation velocity. Fireworks.ai takes a more curated approach, focusing on fewer production-ready models and aggressively deprecating weaker options. Both are converging on MoE and agent-first models as the standard for inference.

Can MoE models produce different results on the same input?

Yes. Because MoE architectures route inputs to different expert sub-networks, the same prompt can activate different experts across runs. This introduces variability in reasoning paths and outputs. For anyone evaluating or benchmarking AI systems, this divergence is an important factor to account for in testing methodology.