top of page

Salesforce Agentic Work Units: Outcomes Beat Activity

Tokens measure noise. Agentic Work Units measure tasks. Neither measures revenue. The metric that matters is whether your agents have the context to produce one.

 

What Salesforce Just Said About AI Value

On February 25, 2026, Salesforce introduced the Agentic Work Unit (AWU) — a platform-level metric defined as one discrete task accomplished by an AI agent. Salesforce reported 2.4 billion AWUs to date, with 771 million in Q4 alone, up 57% quarter-over-quarter.


The framing matters more than the number. For two years, the AI industry has measured progress in tokens — how much language a model can produce. Salesforce is now publicly arguing that tokens measure how much an AI talks, not the work it actually completes. That shift is the right one.


It is also the shift Spotlight.ai built the platform around from day one. The only question that has ever mattered in revenue intelligence is whether AI moved a deal. Not whether it generated a summary. Not whether it called an API. Whether it produced an outcome a CRO can defend in a board meeting.


AWUs Are Better Than Tokens. They Are Not Outcomes.

An AWU counts work performed. An outcome measures whether that work changed the business. Those are different things.


A populated CRM field is an AWU. A qualified deal is an outcome. A meeting summary is an AWU. A coached rep who closes the deal is an outcome. A flow that triggers an email is an AWU. Pipeline that converts is an outcome.


Salesforce acknowledges this elastic relationship. Their own framing notes that the goal is a high inference-to-work ratio — using input tokens to produce concise, high-value output. The unspoken extension of that argument: the work itself has to be high-value. A million AWUs that do not move a single deal is a million units of compute spend with no revenue return.

📊 Sellers spend only 25% of their time actively selling. The rest is consumed by administrative work including CRM data entry, research, and follow-ups. AI can double selling time by automating these tasks — but only if the automation produces outcomes reps can act on.

— Bain & Company, 2025

Why Agents Produce Busy Work Without Context

An agent without context is a worker without a job description. It can execute. It cannot produce value. The pattern repeats across every AI deployment we have seen in enterprise sales:


No Playbook Awareness

A generic agent can summarize a discovery call. It cannot tell you whether the rep skipped Decision Criteria, whether the Champion stake test passed, or whether the deal matches the win patterns from your last twenty closed-won opportunities. Without playbook context, the summary is content. Not intelligence.


No Deal History

An agent that does not know what your team won last quarter cannot evaluate the current pipeline against it. It treats every deal as net new. It misses the comparable deal in stage four that slipped because the Economic Buyer never engaged. It cannot warn you that this deal looks the same.


No Schema for Qualification

MEDDPICC is a forecasting framework, not a checklist. An agent that does not understand element interdependencies — that a Champion without a confirmed Economic Buyer is fragile, that Pain without Metrics is unquantified urgency — will populate fields without validating them. The CRM looks complete. The pipeline still misses.


No Sense of What Good Looks Like

Every revenue org has a definition of a healthy deal. That definition is built from your historical wins, your ICP, your pricing thresholds, your competitive losses. An agent without that grounding executes against a generic template. Generic templates produce generic pipeline.


Context Has a Name: The Knowledge Graph

Spotlight.ai is built on a Knowledge Graph that holds 40 million signals across enterprise sales motions — buyer interactions, deal histories, MEDDPICC evidence, win and loss patterns, account intelligence. It is the layer underneath every Spotlight agent.

The Knowledge Graph is not a vector store. It is a structured map of how revenue actually moves through your organization. Who the real Champions are. Which Decision Criteria show up in won deals and disappear in lost ones. What evidence quality looks like for a deal in your specific industry. How long the Paper Process actually takes for procurement at companies like the one your rep is selling into.


When a Spotlight agent qualifies a deal, runs an inspection, generates a business case, or coaches a rep, it draws on this graph. That is why the output is an outcome and not an AWU. The agent is not executing a task in isolation. It is executing against a model of what success looks like for your business.


📊 75% of B2B sales organizations will augment traditional sales playbooks with AI-guided selling solutions by 2025 — driven by the pressure to transform massive amounts of unstructured data into actionable next steps.

— Gartner, Sales Predicts

MCP Is the On-Ramp. The Brain Is Already Here.

Model Context Protocol gives any AI agent a standard way to plug into external context. The protocol matters. The context that flows through it matters more.


Most enterprise teams now building agents on top of Salesforce, Snowflake, or Slack are solving the same hard problem: where does the agent get its sales context? Without a structured source of truth, every team rebuilds the same brain — partially, inconsistently, with worse data.


Spotlight.ai has spent five years building that brain. The Knowledge Graph is MCP-ready. Your internal agents, your custom workflows, your AI initiatives — they can consume Spotlight context as a service. You do not need to recreate the qualification model, the deal scoring logic, or the playbook awareness layer. It exists. Connect to it.


What Outcome-Grade AI Looks Like in Revenue

The difference between AWUs and outcomes is the difference between activity reports and revenue impact. Here is what outcome-grade AI produces, measured against real Spotlight customers:


Pipeline Conversion Lift

A 300-user customer moved conversion from 7.8% to 12.5% over twelve months — a 4.8 point gain. The AI did not generate more activity. It produced better qualified pipeline at the top of the funnel.


Win Rate Multiplication on Qualified Deals

Tulip, a manufacturing platform customer, recorded a 3.3x win-rate improvement on deals that passed Spotlight qualification. The metric is not how many opportunities the agent touched. It is what happened to the ones it qualified.


Direct Revenue Impact

A Fortune Cyber 60 security leader attributed $14.4 million in direct revenue impact and 3.8x pipeline conversion to autonomous deal execution. That is not an AWU count. That is a CFO-defensible number.


Hours Returned to Selling

The same 300-user deployment recovered 4,530 workdays in a year. Hours saved is the rare AWU-style number that converts cleanly to outcomes — but only because Spotlight ties saved hours to specific qualification, inspection, and content-generation tasks the rep no longer has to do.


Activity Metrics vs Outcome Metrics: A Comparison

The taxonomy of what we measure in AI is moving fast. Tokens were the v1 metric. AWUs are the v2. Outcomes are what the next generation of revenue leaders will be held to.

Metric Type

What It Measures

What It Misses

Tokens Processed

How much the AI talks

Whether anything useful happened

Agentic Work Units (AWUs)

Discrete tasks completed by an agent

Whether the task produced revenue, time saved, or risk reduced

Outcome Metrics

Pipeline conversion, win rate, ACV, hours saved

Nothing — these are the metric the CFO actually cares about

A revenue org that grades AI on AWUs will reward platforms that generate the most activity. A revenue org that grades AI on outcomes will reward platforms that move the metrics on the board deck.


How to Evaluate AI Sales Tools by Outcome, Not Activity

When a vendor reports an activity number, ask the follow-up. The follow-up is always the same: what changed in the business?

●      Ask for pipeline conversion delta. Not how many calls the AI summarized — how the conversion rate moved on summarized opportunities versus the baseline.

●      Ask for win rate on qualified deals. If the AI qualified pipeline, the qualified deals should win at a higher rate than the unqualified ones.

●      Ask for ACV impact. Better qualification surfaces larger deals. If the average ACV did not move, qualification did not improve.

●      Ask for forecast accuracy. An agent producing real intelligence makes the forecast more predictable. Variance to forecast should narrow.

●      Ask what context the agent uses. If the answer is the call recording and a generic LLM, the agent is producing AWUs. If the answer is a structured Knowledge Graph of your sales motion, the agent is producing outcomes.


Stop Counting Activity. Start Counting Outcomes.

Salesforce moved the conversation in the right direction. Activity is not the answer. Tasks completed are not the answer. The answer is whether your agents have the context to produce business results — and whether you can prove it with metrics that matter to revenue.


Spotlight.ai gives your AI initiatives the brain they need. Knowledge Graph as a service. MCP-ready. Outcome-grade.

Salesforce Just Validated What We Built Spotlight Around: Outcomes Beat Activity

 

FAQs About Agentic Work Units and Outcome-Based AI


What is an Agentic Work Unit (AWU)?

An AWU is a metric introduced by Salesforce in February 2026, defined as one discrete task accomplished by an AI agent — a prompt processed, a reasoning chain completed, or a tool invoked. It replaces token consumption as a measure of AI activity but does not measure business outcomes.


Why are AWUs not enough to measure AI value in sales?

AWUs count tasks completed. They do not measure whether those tasks moved a deal, qualified a pipeline opportunity, or returned revenue. A million AWUs that do not change a sales metric is activity without outcome.


What metrics should sales leaders use to evaluate AI tools?

Pipeline conversion rate, win rate on qualified deals, ACV expansion, forecast accuracy variance, and hours returned to selling. These are the metrics a CFO can defend and a CRO can be held to.


How does the Spotlight.ai Knowledge Graph differ from a generic LLM?

A generic LLM has no understanding of your playbook, your historical deals, your qualification standard, or your competitive dynamics. The Spotlight Knowledge Graph holds 40 million structured signals from enterprise sales motions, giving agents the context to produce outcomes specific to your business.


Can I use the Spotlight Knowledge Graph with my own AI agents?

Yes. The Knowledge Graph is built to be MCP-ready, so internal agents, custom workflows, and third-party AI initiatives can consume Spotlight context as a service rather than recreating qualification and deal-scoring logic from scratch.


Does measuring outcomes mean ignoring activity metrics entirely?

No. Activity metrics like AWUs are useful for capacity planning and infrastructure cost. They should not be confused with business value. Outcome metrics are what tie AI investment to revenue.

Comments


bottom of page