CM × AI  Field Guide
A Practitioner’s Field Guide Edition I · MMXXVI
Retail · CPG · Merchandising · Artificial Intelligence

Category
Management &
Artificial
Intelligence

The discipline, its history and methods across every industry — and how to run it with intelligent agents.

An end-to-end field guide
From the strategic business unit to the agentic category team.
Fourteen chapters · a prompt library · a simulation playbook.
14Chapters
13Prompts
10+Industries
↓  Scroll to begin  ↓
Abstract

For four decades, category management has been the operating system of modern retail — the discipline that decides what sits on the shelf, at what price, in what space, and why. Artificial intelligence is now rewriting how that operating system runs.

This field guide does two things. Part I explains the discipline itself: where category management came from, the canonical methods that still define it, how it is practised — very differently — in groceries, pharmacies, fashion, electronics, marketplaces, convenience retail, B2B distribution and beyond, and how it relates to the wider family of commercial disciplines it is so often confused with. It also covers the discipline’s lesser-known twin in procurement, which shares the name but inverts the logic.

Part II is a practical playbook for doing category management with AI: a multi-agent operating model, the context you must engineer before any model can help you, a library of ready-to-use prompts for each step of the process, the role of simulation and scenario labs, and the governance and antitrust guardrails that keep the whole thing legal and trustworthy. It closes with a maturity model and a 90-day roadmap.

A note on intellectual honesty runs through Part II: the gap between what AI can do for category management and what it reliably does today is wide. Vendor and consultant performance figures are presented as claims, attributed, and qualified. Where a capability is a pilot or a roadmap item rather than a proven deployment, this guide says so plainly.

Part One

The Discipline

Before you can automate category management, you have to understand it. Seven chapters on what it is, where it came from, and how it is really practised.

DefinitionsHistoryThe 8-Step Process Category RolesShopper InsightTen IndustriesProcurement’s Twin
Part I · Chapter One
01

What Category Management Is

A category is not a shelf. It is a business — and category management is the discipline of running it like one.

Walk into any supermarket and you are standing inside a portfolio of businesses. The coffee aisle is one business. Oral care is another. Chilled ready-meals, pet food, laundry, salty snacks — each is a distinct enterprise with its own customers, its own competitors, its own economics, and its own reason to exist inside the store. Category management is the discipline of recognising that fact and acting on it: of treating each product category as a strategic business unit with its own role, strategy, targets and accountability, rather than as an undifferentiated stretch of shelving to be filled.

That single shift — from managing brands or SKUs in isolation to managing the category as a coherent whole — is the intellectual core of the discipline. It changes the question a buyer asks. Not “how many cases of this brand can I sell?” but “what is this category for, who is it for, and how do I grow the whole of it?”

DefinitionsFour ways to say the same thing

There is no single canonical definition of category management, but the major ones converge. It is worth seeing them side by side, because each emphasises a different facet of the discipline.

Definition — Nielsen, 1992

“A process that involves managing product categories as business units and customising them on a store-by-store basis to satisfy customer needs.” — the most-cited operational definition; note the emphasis on localisation.

Definition — The Partnering Group / ECR

“A distributor/supplier process of managing categories as strategic business units, producing enhanced business results by focusing on delivering consumer value.” — the canonical industry definition; note “distributor/supplier”: collaboration is built in.

Definition — Institute of Grocery Distribution (UK)

“The strategic management of product groups through trade partnerships which aims to maximise sales and profit by satisfying consumer and shopper needs.” — note the explicit split between the consumer and the shopper.

Pull the four definitions apart and the same four ideas appear in each. First, the category — not the brand or the SKU — is the unit of management. Second, it is run as a strategic business unit: with a role, a P&L logic, targets, and someone accountable for them. Third, every decision is anchored in shopper and consumer value, not in internal convenience. Fourth, it is executed collaboratively, between a retailer and one or more of its suppliers. Hold those four ideas; the rest of this guide is commentary on them.

First principlesThe consumer and the shopper are not the same person

One distinction deserves to be settled immediately, because the entire discipline — and every AI prompt later in this guide — depends on it. The consumer is the person who uses or eats the product. The shopper is the person standing in the aisle, or scrolling the app, making the purchase decision. They are frequently different people — the parent buying for the child, the partner doing the weekly shop — and even when they are the same person, they are in a different mindset. Category management lives at the moment of the shopping decision. It is shopper-facing by construction. Marketing builds the brand in the consumer’s head; category management wins the few seconds in front of the shelf.

Why it existsThe problem category management solves

Before category management, retail buying was organised around suppliers and brands. A buyer negotiated with Brand A, then with Brand B, and the “category” was simply whatever those separate negotiations happened to produce. Three pathologies followed. Assortment sprawled, because every supplier pushed every line and nobody owned the total. Shelf logic was incoherent, because the layout reflected supplier deals rather than how shoppers chose. And nobody was accountable for the category’s overall health — growth, margin, shopper satisfaction — because no one was looking at the category as a whole.

Category management is the cure for all three. It installs an owner, a definition, a role, a strategy and a scorecard for each category — and it does so collaboratively, so the supplier’s deep category expertise is harnessed rather than merely resisted.

The category is the smallest unit of retail you can actually have a strategy about.
The working premise of the discipline

Two disciplines, one nameA warning before you go further

The phrase “category management” names two distinct professional disciplines. This guide is overwhelmingly about the first — retail / FMCG category management, the sell-side discipline born in groceries. But there is a second, equally real discipline called procurement category management: the buy-side practice of grouping an organisation’s purchasing spend into categories and sourcing each strategically. Same name, opposite end of the value chain. Chapter 7 is devoted to it. For now, simply know that the ambiguity exists, so you are never confused by it.

Figure 1.1 — The two disciplines that share the name
 Retail / FMCG category managementProcurement category management
Side of the dealSell side — the retailer’s offer to shoppersBuy side — the organisation’s purchasing
LogicDemand-led: categories defined by how shoppers shopSupply-led: categories defined by supply markets
GoalGrow category sales, margin & shopper satisfactionCut total cost of ownership, manage supply risk
Core frameworkThe 8-step process; category rolesThe Kraljic matrix; 7-step strategic sourcing
Covered inChapters 1–6 & all of Part IIChapter 7

What good looks likeThe signatures of a well-managed category

You can recognise a well-managed category without seeing a single spreadsheet. It has a clear definition that matches how shoppers think, not how the buying desk is organised. It has been assigned an explicit role in the store. Its assortment, pricing, shelf layout and promotions are mutually coherent and visibly serve that role. It is measured against a scorecard that someone owns. And it is reviewed on a cycle, so it adapts. The chapters that follow give you each of those instruments in turn — and Part II hands the cycle to a team of AI agents.

Chapter 1 in one line

Category management treats each product category as a business with its own role, strategy and scorecard — managed collaboratively, anchored in shopper value, and run on a continuous cycle.

Part I · Chapter Two
02

A Short History, 1969–Today

Category management was not invented in a boardroom. It was made possible by a barcode and forced into being by a crisis.

Disciplines have origin myths; category management has an origin mechanism. It could not exist until retail became data-rich, and retail did not become data-rich until products could be scanned. The story therefore begins not with a theorist but with an engineer, and a 12-digit code.

The enablerThe barcode makes the category measurable

In 1973 the IBM engineer George Laurer developed the Universal Product Code — the 12-digit numeric barcode. On 26 June 1974, a pack of Wrigley’s Juicy Fruit gum was scanned at a Marsh Supermarket in Troy, Ohio: the first commercial UPC scan. Through the 1980s, chains rolled out scanning nationally. The consequence was profound and is easy to undervalue. For the first time, a retailer knew — at the level of the individual item, store and day — exactly what had sold. Retail had become a measured environment. You cannot manage a category as a business unit until you can see its numbers; the barcode is what let you see them.

The originatorBrian Harris and the category as a business unit

The conceptual leap belongs to Brian F. Harris, an Australian-born marketing academic at the University of Southern California. In the 1980s Harris built the Apollo Space Management System — early software that calculated optimal shelf-space allocation, and the direct ancestor of every planogramming tool in use today. From that work grew a bigger idea: that a retailer should treat each category the way a corporation treats a division — with a role, a strategy, a P&L and an owner. In 1990 Harris founded The Partnering Group, the consultancy that would codify the method. The exact birth year of “category management” is genuinely fuzzy: the concept dates to the late 1980s, while the formal term and the codified eight-step model were popularised through the mid-1990s. It is most honest to describe it as a discipline developed across roughly 1989–1997.

The proofProcter & Gamble and Walmart

Two corporate moves turned theory into practice. First, in 1987, Procter & Gamble abandoned its 56-year-old brand-management structure — the very system Neil McElroy had invented at P&G in 1931 — and reorganised into category business units, each with its own sales, finance and product-development functions. The largest CPG manufacturer in the world had just declared that the category, not the brand, was the unit that mattered. Second, beginning in 1987–88, P&G and Walmart built the era’s defining collaboration: continuous replenishment, shared data, and ultimately Walmart’s Retail Link system, which gave suppliers item-store-day POS data. Walmart–P&G became the textbook proof that retailer and supplier sharing data could grow a category for both. It is still cited that way.

The crisisECR forces the method onto the industry

In the early 1990s, US grocers felt mortally threatened by mass merchandisers, warehouse clubs and drugstores piling into food. Trade associations commissioned a study, and in 1993 Kurt Salmon Associates published Efficient Consumer Response, claiming the grocery supply chain could be re-engineered to remove a figure widely quoted as up to $30 billion in cost. ECR launched the same year as an industry-wide movement, built on four pillars: efficient assortment, efficient replenishment, efficient promotion and efficient product introduction. Category management became the demand-side strategic process within ECR. ECR Europe followed in 1994; the 8-Step Approach to Category Management was launched under the ECR banner in 1996. A method that began as one academic’s framework was now an industry standard.

1969–73

US grocers form a committee for a uniform product code; IBM’s George Laurer develops the UPC barcode (1973).

26 June 1974

First UPC scan — a pack of gum, Marsh Supermarket, Troy, Ohio. Retail becomes measurable.

1979

IRI founded by John Malec and Gerry Eskin, pioneering scanner-based market tracking (InfoScan, BehaviorScan).

1987

P&G replaces brand management with category business units; begins the data-sharing partnership with Walmart.

Late 1980s

Brian Harris develops the category-as-business-unit concept; the Apollo space-management system precedes modern planogramming.

1990

Harris founds The Partnering Group, which will codify the method.

1992

Nielsen publishes its definition of category management — categories as business units, customised by store.

1993–94

ECR launches in the US (1993, after the Kurt Salmon report) and Europe (1994). Category management becomes its demand-side process.

1995

Tesco launches Clubcard; dunnhumby (founded 1989) turns loyalty data into category insight — the dawn of shopper-centric CM.

1996

The 8-Step Approach to Category Management is launched under the ECR banner — the method becomes an industry standard.

2000s

Loyalty data, store clustering and shopper insight push CM from data-driven toward shopper-centric; the FTC examines category-captain antitrust risk.

2010s

E-commerce creates the digital shelf; the category must now be managed across store, web and app at once.

2020s

Retail media networks monetise the shelf; machine learning industrialises forecasting and assortment; agentic AI arrives on the roadmap.

The measurersNielsen, IRI and the syndicated-data industry

A discipline that runs on data needs data suppliers. Two firms came to dominate. AC Nielsen evolved from store audits into scanner-based measurement; Information Resources, Inc. (IRI), founded in Chicago in 1979, pioneered scanner panels and the InfoScan national tracking service, forcing the whole industry onto syndicated scanner data. That data — standardised, cross-retailer, comparable — became the analytical fuel of category assessments and scorecards. The lineage still matters: Nielsen’s retail arm is today NielsenIQ; IRI merged with The NPD Group in 2022 to form Circana. When Chapter 10 discusses the “context” you must feed an AI, syndicated panel data from these firms is one of its load-bearing inputs.

The critiqueWhy the classic model started to crack

By the 2000s the eight-step process was under fire from its own practitioners. It was slow and resource-heavy — one widely cited industry survey found only around 9% of supplier companies used the full eight-step process; most used trimmed-down versions. It had drifted toward being retailer-centric, a lever for extracting slotting fees and captain payments rather than a genuine win-win. And it was criticised as insufficiently connected to the actual shopper. An FMI / Deloitte study reportedly found essentially every retail and CPG respondent believed the prescribed process needed to change. The response was two-fold: streamlined proprietary processes (Nielsen, for instance, promotes a five-step version), and a decisive turn toward shopper-centric category management — the subject of Chapter 4.

A note on dates

Sources genuinely disagree on when category management “began” — the concept (late 1980s), the term, and the codified 8-step model (mid-1990s) emerged at different times. This guide treats it as a discipline developed across 1989–1997 rather than asserting a single founding year. The $30 billion ECR savings figure is the 1993 Kurt Salmon report’s claim; the “9% use the full process” statistic is a single, undated survey figure. Treat both as cited claims.

Chapter 2 in one line

The barcode made the category measurable; P&G and Walmart proved collaboration worked; the ECR crisis made the eight-step method an industry standard — and its own weight then pushed the discipline toward the shopper.

Part I · Chapter Three
03

The Methods: Process, Roles & Tactics

One process, four roles, four levers and a scorecard. The classic toolkit — and the one your AI agents will run.

If Chapter 1 gave you the philosophy and Chapter 2 the history, this chapter gives you the machinery. Almost everything an AI agent does in Part II of this guide is an automation, an acceleration or an augmentation of one of the four instruments described here: the eight-step process, the category-role framework, the category-captain model, and the four tactical levers. Learn them as a practitioner would; you cannot supervise an agent doing work you do not understand.

The processThe eight-step category management cycle

Developed by The Partnering Group and codified through ECR, the eight-step process is the canonical reference model. It is a cycle, not a project: step 8 loops back to step 1. Few organisations run all eight steps in full — most use streamlined versions — but every streamlined version is a compression of these eight, so this is the model to know.

1
Category Definition

Decide which products constitute the category and its sub-segments — defined through the shopper’s eyes (what they see as substitutable), not by the buying desk’s structure.

2
Category Role

Assign the category’s strategic role in the total store — Destination, Routine, Occasional/Seasonal or Convenience. The role dictates how much price, space and attention it earns.

3
Category Assessment

Diagnose the category by consumer, market, retailer and supplier — sales, margin, GMROI, share, penetration, availability — to find the biggest gaps and opportunities.

4
Category Scorecard

Set the measurable targets and KPIs the category must hit, by role, over a defined horizon — and name who owns them.

5
Category Strategy

Choose the strategy archetypes to close the gaps — traffic-building, transaction-building, profit-generating, cash-generating, excitement-creating, image-enhancing, turf-defending.

6
Category Tactics

Translate strategy into specifics across the four levers: assortment, pricing, shelf/space and promotion.

7
Plan Implementation

Operationalise it — accountabilities, timelines, pricing-system loads, planogram deployment, replenishment rules, store and supplier briefing.

8
Category Review

Measure actuals against the scorecard; run post-promotion analysis; periodically re-examine role and strategy — then loop back to step 1.

The rolesFour jobs a category can hold

Step 2 deserves its own discussion, because it is where scarce resources — space, capital, price investment, management attention — are deliberately allocated rather than spread evenly. A retailer assigns each category exactly one of four roles, based on how important the category is to the target shopper, how much it differentiates the retailer, the competitive dynamics, and its financial contribution.

Destination

The retailer wants to be the provider; the category defines the store and drives store choice. Deepest assortment, most aggressive pricing, premium space. A small minority of categories.

Routine / Preferred

Everyday-need categories where the retailer aims to be consistently preferred. The workhorse — usually the largest share of categories, sales and profit. Balanced assortment, competitive pricing.

Occasional / Seasonal

Bought periodically or tied to seasons and events. A secondary profit role, but a powerful tool to differentiate the retailer during a specific window.

Convenience

Completes the basket and supports one-stop shopping. Managed primarily for margin — shoppers are less price-sensitive here. Tighter range, secondary placement.

The discipline of roles is that everything else follows from them. A Destination category warrants deep assortment, aggressive known-value-item pricing, premium space and a strong promotional calendar. A Convenience category warrants the opposite on every lever. A category plan whose tactics contradict its assigned role is, by definition, a broken plan — and, as Chapter 13 notes, “does this plan match its role?” is exactly the kind of check an AI governance agent can run automatically.

The collaborationThe category captain — and its dangers

Retailers rarely have the analytical firepower to manage every category deeply in-house, so they borrow it. A category captain is a single supplier — almost always the category’s leading manufacturer — that a retailer designates to provide deep analysis and recommendations for the whole category, including the products of the captain’s own competitors. In exchange for privileged data access, the captain invests its shopper insight and category expertise. A second supplier is often named category validator to challenge the captain’s recommendations; a neutral, brand-agnostic category advisor may also be used.

The arrangement is genuinely useful and genuinely dangerous. The danger is structural: you have asked a self-interested competitor to advise on how its rivals are treated on the shelf. Antitrust scholars and the US FTC have identified two principal harms — competitive exclusion (the captain quietly disadvantages rivals’ facings or recommends their delisting) and facilitated collusion (one captain serving most retailers in a market becomes a common information conduit). The accepted safeguards: captains should advise, not decide; retailers retain final authority; validators review; and the captain is never told how rivals price or promote. Hold this thought — Chapter 13 shows that an AI category captain sharpens every one of these concerns.

Documented example

The UK Groceries Code Adjudicator’s 2015–16 investigation into Tesco found suppliers had paid “large sums of money in exchange for category captaincy or participation in a price review” — a concrete instance of the captain role drifting into a payment-extraction mechanism rather than a value-creation one.

The leversThe four tactical levers

Step 6 executes strategy through four coordinated levers — the category-management adaptation of the marketing mix. Their golden rule is internal coherence: the four must reinforce each other and the category role.

Figure 3.1 — The four tactical levers
LeverWhat it decidesKey concepts
AssortmentWhich SKUs the category carries — breadth and depthRange optimisation, good-better-best architecture, private-label role, SKU rationalisation, innovation slotting
PricingThe category’s price architecture and postureKnown-Value-Items (KVIs), price ladders, private-label gaps, EDLP vs. Hi-Lo
Space & ShelfWhere and how each product is placedPlanograms, facings, adjacencies, eye-level placement, blocking to the decision tree
PromotionThe promotional planCadence, depth, mechanics (feature, display, multibuy, coupon), promo ROI, post-event analysis

The scorecardHow a category is measured

Steps 3 and 4 run on numbers. A category scorecard mixes growth, profit, productivity and shopper metrics. The most important to know:

  • GMROI — Gross Margin Return on Inventory Investment. Gross margin divided by average inventory cost; the single best measure of how hard the category’s capital works. A GMROI of 4 means $4 of margin per $1 of inventory.
  • Sales per linear metre / per facing / per square foot — space productivity; the basis of every planogram decision.
  • Penetration — the share of baskets (or households) that contain the category. The clearest measure of how broadly a category recruits shoppers.
  • Basket / attachment analysis — what is bought alongside what; items per basket; cross-sell rates.
  • Market share & category growth — the category’s and brand’s share of a defined market, and its movement over time.
  • Inventory turns, sell-through, on-shelf availability, price index, promo ROI, markdown rate — the operating vital signs.
Chapter 3 in one line

The eight-step cycle runs the category; four roles allocate resources to it; four coherent levers execute it; a scorecard measures it; and a category captain — usefully but dangerously — helps a retailer do all of it.

Part I · Chapter Four
04

Shopper-Centric & Demand-Based CM

The major evolution of the discipline: stop starting with the category’s sales data, and start with the shopper’s mind.

The classic eight-step process has a quiet bias. It begins with the category — its definition, its data, its performance — and only later asks about the shopper. Shopper-centric category management inverts that. It does not throw away the eight steps; it re-anchors them, so that the shopper’s needs, missions and decision logic drive category definition, assortment, space and promotion from the first step rather than the fourth. This chapter explains the three instruments that make the re-anchoring possible — and they matter doubly here, because they are exactly the instruments an AI is best placed to build for you.

The instrumentThe Consumer Decision Tree

The Consumer Decision Tree (CDT) — increasingly called the Shopper Decision Tree, because it is the shopper, not the end consumer, who chooses at the shelf — is a graphical map of the hierarchy of attributes a shopper applies when choosing within a category. It answers a precise question: in what order do shoppers narrow their choice?

Consider hot beverages. One shopper’s tree might branch first on form (ground / beans / pods / instant), then on caffeine (regular / decaf), then on roast, then on brand, then on pack size. The order of those branches is the whole point. It tells you the shelf should be blocked by form first — because a shopper who has decided “pods” will not even look at the ground-coffee bay. It tells you which SKUs are genuine substitutes (same leaf of the tree) and which are not. And it tells you, when you rationalise the range, which items you can cut without leaving a need-state uncovered.

Definition — the three uses of a decision tree

Assortment: keep at least one strong item on every leaf; cut duplicates within a leaf. Shelf layout: the planogram should physically mirror the tree, branch by branch. Segmentation: the tree is the category’s sub-structure — its segments and sub-segments.

A terminology note worth carrying forward: vendors use “Consumer Decision Tree,” “Shopper Decision Tree” and “Category Decision Tree” more or less interchangeably. They mean the same artefact. When Chapter 11 gives you a prompt to generate one with an AI, it does not matter which name you use — it matters that you specify the attributes and the data the tree should be built from.

The instrumentShopper missions

The same shopper behaves differently on different trips. A shopping mission is defined by the motivation behind the trip, its context, and the behaviours that follow. The common missions:

  • Big stock-up — the planned weekly or fortnightly shop; large basket, price-conscious, list-driven.
  • Top-up / fill-in — a quick mid-week replenishment of a few items.
  • Immediate consumption / food-to-go — bought to be used now; convenience and speed dominate.
  • Special occasion — a dinner, a holiday, an event; higher willingness to pay, more exploration.
  • Emergency / distress — the forgotten essential; the shopper is captive and price-insensitive.

Missions explain why the same category plays a different role on different trips, and why store format, assortment depth and layout must flex. They are also the bridge to Chapter 12’s synthetic personas: a well-specified mission is precisely the brief from which an AI can simulate a shopper.

The instrumentThe path to purchase

The path to purchase is the full, multi-channel journey a shopper takes from a need or trigger, through awareness, consideration and decision, to the purchase and beyond. In a single-channel world it was a short walk down an aisle. Today it threads through search, social, reviews, apps, click-and-collect and the physical shelf — and category management must influence the category’s assortment, content and visibility at every touchpoint, not just the last one. Mapping the path reveals where the category is won and lost; it is the reason the digital shelf of Chapter 6 is now as strategically important as the physical one.

The category does not start at the shelf. It starts in the shopper’s head, three days before the trip.
The premise of shopper-centric CM

Classic vs. shopper-centricWhat actually changes

Figure 4.1 — Re-anchoring the eight steps on the shopper
StepClassic starting pointShopper-centric starting point
DefinitionBuying-desk product structureThe shopper decision tree
RoleFinancial contribution to the storeImportance to the target shopper & their missions
AssessmentPOS sales & share performanceShopper penetration, frequency, loyalty, switching
TacticsOptimise sales per metreServe each need-state & mission on every channel

Who drove itLoyalty data and the insight firms

Shopper-centric category management did not arrive as a theory; it arrived as a capability, the moment retailers could see households rather than just transactions. Tesco’s Clubcard (1995) and the firm dunnhumby turned loyalty data into household-level purchase histories — who buys, how often, what they switch to, what they never buy. Kroger’s 84.51° does the same in the US. That data is what lets a decision tree be estimated rather than guessed, a mission be sized rather than asserted. It is also, not coincidentally, the richest single input you will feed an AI in Chapter 10.

Chapter 4 in one line

Shopper-centric CM re-anchors the eight-step process on the shopper using three instruments — the decision tree, the mission, and the path to purchase — each of which is something an AI can now help you build from data rather than intuition.

Part I · Chapter Five
05

The Related-Disciplines Map

Category management is surrounded by a family of disciplines it is constantly confused with. Knowing the boundaries is half the expertise.

Ask ten retail professionals to draw the line between category management, trade marketing, shopper marketing and revenue growth management, and you will get ten different maps. The terms overlap, the org charts vary by company, and the vocabulary drifts. But the disciplines are distinct, and a category manager who cannot place them cannot collaborate cleanly with the teams that own them. This chapter is the map.

Push vs. pullTrade marketing and shopper marketing

These two are the easiest pair to separate, because they pull in opposite directions. Trade marketing is B2B marketing aimed at the intermediaries — retailers, distributors, wholesalers. It is a push strategy: pushing product through the channel via sales programmes and trade promotions. Shopper marketing is aimed at the shopper at or near the point of sale. It is a pull strategy: drawing shoppers to the product through in-store activation, displays and digital touchpoints.

Category management sits between them. It supplies the data-backed category rationale that makes a trade-marketing pitch credible to a retail buyer, and it shares its insight base — decision trees, missions, path to purchase — with shopper marketing. Roughly: category management decides what goes where on the shelf and why; shopper marketing decides how to activate shoppers around it; trade marketing decides how to sell the whole programme into the channel.

The collaboration wrappersJBP, ECR and CPFR

Joint Business Planning (JBP) is the commercial and governance wrapper around category plans — a multi-year, cross-functional agreement between a retailer and a supplier that turns category strategy into committed shared targets, investments and promotional calendars. If category management produces the plan, JBP is the handshake that funds and commits it.

Efficient Consumer Response (ECR), as Chapter 2 explained, is the parent movement — category management was its demand-side process. Its supply-side counterpart is CPFR (Collaborative Planning, Forecasting and Replenishment), pioneered by Walmart and Warner-Lambert in 1995 and formalised by the VICS association in 1998. CPFR has retailer and manufacturer share forecasts and jointly resolve exceptions. The clean way to hold it: category management and shopper marketing create demand; CPFR and replenishment fulfil it; JBP commits the money behind both.

The monetisation layerRevenue Growth Management

Revenue Growth Management (RGM) — also called Net Revenue Management — is a data-driven discipline for optimising net revenue and margin, not just volume, across five levers: pricing, promotions, Price Pack Architecture (PPA), trade terms, and channel mix. Price Pack Architecture is the structuring of the portfolio across price points and pack sizes to match consumer need-states. RGM effectively absorbs and systematises the pricing and promotion levers of category management and connects them to portfolio and pack decisions. Category management defines the category framework; RGM optimises how money is made within it.

The execution layerMerchandising, assortment & space planning

Retail merchandising is the umbrella discipline of what is sold and how it is assorted, priced, displayed and replenished. Within it, assortment planning (what to carry) and space planning / planogramming (how to display it) are the executional core of category management. Category management provides the strategic intent of steps 1–5; merchandising executes the tactics of steps 6–7. In apparel, as Chapter 6 shows, this same execution layer is called buying & merchandising and is governed by the Open-To-Buy budget.

The new neighbourRetail media networks

The newest discipline on the map — and the one most actively reshaping category management — is the retail media network (RMN): a retailer selling advertising inventory to brands, monetising its own first-party shopper data. US retail-media ad spend has been projected in the region of $60–70 billion a year. RMNs matter to category management for two reasons. First, they draw on exactly the category-level purchase intelligence that category teams produce. Second, and more disruptively, they invert the historical relationship: the retailer, traditionally the discerning buyer to whom the supplier sold a category-management service, now also sells media back to that same supplier. Sponsored placements increasingly shape the digital shelf the way slotting fees once shaped the physical one — reopening, in a new form, every fairness and antitrust question of the category-captain model.

The internal competitorPrivate label

Private label — the retailer’s own brands — is not a separate discipline so much as a category-management problem turned inward. Modern private label uses a tiered, “laddered” architecture: an economy tier for price-driven shoppers, a standard mid-tier for the mainstream, and a premium tier (often branded away from the retailer’s name). Each tier is managed with category-management foundations, as a sub-brand with a target segment. The hard part is that private label competes with the very national brands a category captain represents — so growing it well means modelling cannibalisation carefully and accepting that, in the hard-discount model of Chapter 6, private label simply is the category.

Figure 5.1 — The related-disciplines map
DisciplineOne-line definitionRelationship to category management
Trade marketingB2B “push” marketing to channel intermediariesUses CM’s category rationale to sell into retailers
Shopper marketing“Pull” marketing to the shopper near the point of saleShares CM’s insight base; activates around the shelf
Joint Business PlanningMulti-year retailer–supplier growth agreementThe commercial wrapper that commits the category plan
ECR / CPFRSupply-chain collaboration & joint forecastingECR is CM’s parent; CPFR is its fulfilment counterpart
Revenue Growth ManagementOptimising net revenue across 5 commercial leversSystematises CM’s pricing & promotion levers
Merchandising / space planningAssortment, layout & replenishment executionThe execution layer for CM steps 6–7
Retail media networksRetailers selling ads on first-party shopper dataMonetises CM’s insight; inverts the captain relationship
Private labelRetailer-owned tiered brandsA category-management problem turned inward
Chapter 5 in one line

Category management creates the category plan; trade and shopper marketing push and pull it; JBP commits it; CPFR fulfils it; RGM monetises it; merchandising executes it; retail media is quietly rewriting the rules underneath all of it.

Part I · Chapter Six
06

Category Management Across Industries

The same discipline, ten very different jobs. Groceries invented category management — everyone else bent it to their own shape.

Category management is often taught as if it were a single, universal method. It is not. The eight-step process and the four roles are a common grammar, but each industry speaks a different dialect — because each faces a different shopper, a different shelf, a different clock. A planogram is meaningless to a fashion buyer; an Open-To-Buy budget is meaningless to a grocer; the “shelf” itself dissolves into a search ranking the moment you move online. This chapter tours ten verticals. Read it as a reminder that, in Part II, any AI you deploy must be told which dialect it is working in.

Vertical 01Grocery & supermarkets — the birthplace

Grocery invented category management and remains its most mature practitioner. A supermarket is organised into several hundred categories, each run as a P&L, each with a retailer category manager collaborating with a supplier captain. The signatures are high SKU counts, perishability, frequent promotion and deep loyalty-data integration. Tesco pioneered loyalty-driven category management through Clubcard and dunnhumby; Kroger built the same capability in the US through 84.51°; Walmart’s Retail Link gave suppliers the POS data that makes captaincy work.

Hard discount, though, breaks the model entirely. Aldi and Lidl carry well under 2,000 SKUs against a supermarket’s 30,000–50,000, and private label can be 70–90% of the assortment. There is no retailer-versus-manufacturer collaboration to speak of, because the discounter is the brand. Category management becomes ruthless in-house range curation, sourcing and product development — one category is often one or two SKUs, and the captain role is internalised.

Vertical 02FMCG / CPG manufacturers — the other side of the table

For a manufacturer, category management is the primary lever for influencing a shelf it does not own. Large CPG firms run dedicated category leadership teams that build multi-year “category visions” — growth roadmaps they bring to retailers as captain or validator. The captain provides the retailer with marketplace expertise, shopper insight, planogram recommendations and space-to-sales analysis. Procter & Gamble is a default captain candidate across many of its categories; Unilever, Nestlé, Mars, Coca-Cola and PepsiCo all run category-leadership functions. The 2025–26 wave of CPG portfolio reshaping — Mars acquiring Kellanova, Unilever spinning off ice cream, Nestlé exiting water — is itself category-portfolio management played at the corporate level.

Vertical 03Pharmacy & drugstore — two stores in one

A drugstore is really two stores. The dispensary handles regulated prescription medicine and is managed more like a procurement formulary. The front of store — OTC, health & wellness, beauty, personal care — is classic retail category management, leveraging pharmaceutical credibility to sell adjacent categories. Front-of-store profitability has been a chronic struggle: Walgreens has refocused on categories of leadership such as women’s health and a large owned-brand programme; CVS has tested small-format stores. In Germany, dm and Rossmann show how far the own-brand model can run — dm’s Balea and alverde, Rossmann’s Isana and enerBiO — with third-party quality validation used as a category-trust strategy.

Vertical 04DIY & home improvement — the line review

Home improvement organises category management around the line review: a periodic, formal re-evaluation of every vendor and the merchandising mix within a department, with vendors competing for placement on POS data, margin and supplier performance. The defining complication is two shoppers in one aisle — the DIY consumer and the Pro/trade customer — with different missions and basket economics. Home Depot and Lowe’s in the US, and B&Q, OBI and Bauhaus in Europe, all run this line-review-driven structure; fill rate, inventory turns and sales per square foot dominate the scorecard.

Vertical 05Fashion & apparel — buying & merchandising

Apparel barely uses the words “category management.” Its parallel discipline is buying & merchandising, structured as department → class → subclass and governed financially by the Open-To-Buy — a budget for what a buyer may purchase in a period. Style and seasonality dominate over the stable “category-as-business-unit” idea; the scorecard is sell-through rate, markdown percentage and weeks of cover, not long-run category share. Zara is the instructive extreme: it does not forecast finished-goods quantities at all — it forecasts fabric and uses quick response to design-to-shelf in weeks, replacing fixed category plans with continuous dynamic assortment.

Vertical 06Consumer electronics — thin margins, vendor theatre

Electronics categories — TVs, laptops, mobile, white goods, gaming, accessories — are managed amid rapid obsolescence and price deflation. Thin hardware margins push category management toward attachment: warranties, accessories, services and financing. Vendors co-invest heavily — funded displays, in-store specialists, shop-in-shop arrangements that blur captaincy. Best Buy pioneered buy-online-pick-up-in-store as a way for physical retail to compete with pure-play e-commerce; Europe’s MediaMarktSaturn is explicitly repositioning “from a product retailer to a solution-oriented omnichannel platform.”

Vertical 07E-commerce & marketplaces — the shelf dissolves

Online, the “shelf” becomes the digital shelf — search results, ranking, browse pages, recommendations — and category management splits in two. In first-party (1P) retail, a vendor manager owns the wholesale relationship: terms, purchase orders, assortment. In third-party (3P) marketplace, sellers sell direct and the retailer’s category team curates, sets policy and shapes discoverability rather than buying stock. Shelf position becomes search ranking and content quality; assortment can be near-infinite; category management merges with SEO, retail-media bidding and pricing algorithms. The KPIs change accordingly — share of search, glance views, buy-box win rate, content completeness, review rating.

Vertical 08Convenience & fuel — the forecourt takes over

A convenience store’s tiny footprint forces extreme SKU discipline and a focus on high-velocity, high-margin immediate-consumption categories. The strategic shift of the decade: fuel no longer drives growth — food does. Prepared and fresh food has become the new destination category, with the channel converging on quick-service restaurants. 7-Eleven is building larger, food-focused stores, running a delivery app with a subscription tier, and launching its own retail media network — convenience retail acquiring the full modern category toolkit.

Vertical 09B2B distribution & wholesale — category management as a service

Distributors apply category management to manage vast multi-manufacturer SKU portfolios — and increasingly sell it as a service to their business customers. Sysco, the world’s largest broadline foodservice distributor, markets “category management and insights” to restaurants and has launched a third-party marketplace, bringing e-commerce category logic into B2B. AutoZone manages 70-plus distinct categories with dedicated category-manager roles. The “shopper” here is a professional buyer, and assortment breadth is the value proposition.

Vertical 10The long tail — automotive, telecom, banking, QSR, healthcare

  • Automotive parts: the unique problem is vehicle-application fitment — the same category managed against thousands of make/model/year combinations.
  • Telecom: handsets are often a subsidised acquisition category whose role is to drive the high-margin connectivity and services categories; portfolios now extend into mobile financial services.
  • Banking: “categories” are product lines — accounts, mortgages, cards, insurance — managed as portfolios; KPIs are products-per-customer and lifetime value, not shelf space.
  • Hospitality / QSR: the menu is the assortment; menu engineering classifies items by popularity and profitability into stars, plowhorses, puzzles and dogs.
  • Healthcare: procurement-side category management dominated by Group Purchasing Organisations that aggregate buying power across medical-supply categories.

Cross-cuttingThree forces that bend every vertical

Scale and market maturity. Large retailers run category management in-house with dedicated teams; small retailers receive it — planograms and advice — from suppliers, and are correspondingly more influenced by the captain. Developed markets are dominated by organised modern trade; in many emerging markets, traditional trade — kiosks, corner shops, open markets — can be the majority of consumption, and category management shifts from planograms to distributor- and rep-led execution.

The omnichannel problem. A category must now be managed across store, web and app at once — and a winning physical planogram is not a winning search-ranking strategy. Best practice fuses POS, market research, online search and e-commerce sales into one shopper view.

Retail media. As Chapter 5 noted, RMNs are turning paid placement into a determinant of the digital shelf across every vertical that has one.

Figure 6.1 — Ten verticals at a glance
VerticalThe “shelf” is…Distinctive challengeLead KPI
GroceryPhysical aisle, planogramSKU sprawl, perishability, discounter threatGMROI, sales/metre
CPG manufacturerSomeone else’s shelfInfluence without ownership; captaincyCategory share, distribution
PharmacyFront-of-store fixturesTwo stores in one; weak front-of-store marginFront-of-store margin, own-brand share
DIYBays, rackingDIY vs. Pro shopper in one aisleFill rate, sales/sq ft
FashionFloor, rail, seasonShort lifecycles; trend velocitySell-through, markdown %
ElectronicsFixtures + shop-in-shopObsolescence; thin hardware marginAttach rate, services mix
E-commerceSearch ranking & PDPEndless aisle; shelf = algorithmShare of search, buy-box win
ConvenienceA few square metresTiny space; food displacing fuelFoodservice penetration
B2B distributionCatalogue & warehouseMulti-manufacturer breadth; B2B buyerLines per order, fill rate
QSRThe menu boardOperational complexity per itemMenu-item profit & popularity
Chapter 6 in one line

The eight-step grammar is universal; the dialect is not — and any AI agent must be told whether its “shelf” is a planogram, a search ranking, a season or a menu board.

Part I · Chapter Seven
07

The Other Category Management

Procurement’s category management shares the name and almost nothing else. It is worth a chapter, because the confusion is constant — and because its tools are genuinely useful.

Search “category management” and roughly half the results will be about retail and roughly half about procurement. They are different disciplines. Retail category management, the subject of every other chapter in this guide, is a sell-side practice: how a retailer shapes its offer to shoppers. Procurement category management is a buy-side practice: how any organisation — a manufacturer, a hospital, an airline, a bank — organises its purchasing spend. Same two words, opposite ends of the value chain. This chapter gives procurement category management its due, both to dissolve the confusion permanently and because its core frameworks are sharp enough to be worth borrowing.

The definitionWhat procurement category management is

In procurement, category management means organising the buying team’s resources to face outward onto the supply markets that serve the organisation. Spend is grouped into coherent categories — IT hardware, logistics, marketing services, raw polymer, temporary labour — and each is given a tailored sourcing strategy. The objective is not sales growth but the optimisation of total cost of ownership, supply risk and value. It is continuous, long-term and governance-led, owned by procurement rather than by merchandising.

The frameworkThe Kraljic matrix

If retail category management has the eight-step process, procurement category management has the Kraljic matrix — created by Peter Kraljic, then of McKinsey, in the 1983 Harvard Business Review article “Purchasing Must Become Supply Management.” It segments every purchase on two axes — profit impact and supply risk — into four quadrants, each with its own strategy.

Routine items

Low impact, low risk. Strategy: simplify and automate — minimise the cost of buying them at all.

Leverage items

High impact, low risk. Strategy: exploit buying power — competitive tendering, consolidation.

Bottleneck items

Low impact, high risk. Strategy: secure supply — manage risk, find alternative sources.

Strategic items

High impact, high risk. Strategy: deep, collaborative partnerships with key suppliers.

The Kraljic matrix is to procurement what the category-role framework is to retail: a way of deciding, deliberately, that not every category deserves the same treatment. It also underpins supplier segmentation — managing strategic, leverage, bottleneck and routine suppliers with different intensity rather than treating all suppliers alike.

The processThe 7-step strategic sourcing cycle

Where retail has eight steps, procurement category management most often runs a seven-step strategic sourcing process:

  1. Project assessment — baseline current spend, supplier relationships and stakeholder expectations.
  2. Category profile — analyse spend patterns, the supplier landscape, market dynamics and risk.
  3. Sourcing strategy — define the approach, using Kraljic positioning.
  4. Supplier selection criteria — weighted evaluation criteria aligned to business needs.
  5. Supplier engagement — run RFI / RFP / RFQ to gather proposals.
  6. Negotiation — structured, data-led negotiation.
  7. Implement agreements — transition contracts to operations, with monitoring and compliance.

Spend analysis — consolidating, cleaning and categorising all supplier spend into a single view — is the foundational input to the whole cycle. The Chartered Institute of Procurement & Supply (CIPS), the global professional body, publishes its own multi-phase category-management model spanning initiation, analysis, strategy and implementation. (CIPS also maintains a separate, longer procurement-and-supply cycle; the two should not be conflated.)

Side by sideThe two disciplines compared

Figure 7.1 — Retail vs. procurement category management
 Retail / FMCG CMProcurement CM
OriginGrocery retail; Brian Harris, 1989–97Procurement; Kraljic, HBR 1983
DirectionDemand-led (sell side)Supply-led (buy side)
The “category” is…A group of products shoppers see as relatedA group of spend items served by one supply market
GoalGrow sales, margin, shopper satisfactionCut total cost of ownership; manage supply risk
Core framework8-step process; 4 category rolesKraljic matrix; 7-step sourcing
Key dataPOS, loyalty, panel dataSpend analysis, supplier & market data
Owned byMerchandising / category teamsProcurement / purchasing

Why it is in this guideThe borrowable ideas

Two things from procurement category management are worth carrying back. First, the Kraljic logic — segment by impact and risk, then differentiate treatment — is a clean mental model that applies to suppliers, to SKUs, and even to which categories deserve AI investment first (Chapter 14 uses exactly this kind of prioritisation). Second, procurement was an early and aggressive adopter of AI agents: “category agents” that monitor spend, flag savings opportunities and draft sourcing recommendations are already a live use case, and consultancies cite material efficiency gains from them. The agentic operating model of Chapter 9 has a cousin in the procurement world — and the two are converging.

Don’t conflate them

When a vendor, a consultant or a job description says “category management,” establish first which discipline they mean. An “AI for category management” tool built for procurement spend analysis will not run a retail planogram — and vice versa. The rest of this guide means the retail discipline unless it explicitly says otherwise.

Chapter 7 in one line

Procurement category management is the buy-side twin of the retail discipline — same name, inverted logic, Kraljic instead of the eight-step process — and its segmentation thinking and head start on AI agents are both worth borrowing.

Part Two

Category Management with AI

The playbook. A multi-agent operating model, the context to engineer, a library of working prompts, a simulation lab — and the guardrails that keep it all legal.

State of the ArtAgent ArchitectureContext Packs 13 PromptsSimulationAntitrust90-Day Roadmap
Part II · Chapter Eight
08

The State of AI in Category Management

An honest map of what works, what is emerging, and what is still a slide deck. Read this before you believe a demo.

Before a single prompt, a sober chapter. The distance between what AI can do for category management and what it reliably does today is wide, and a guide that pretended otherwise would set you up to fail. So this chapter does three things: it states the credibility problem plainly, it sorts every major capability into proven, emerging or hype, and it explains why most of the value on offer is not, in fact, generative AI at all.

The reality checkTwo numbers to keep on the desk

Two findings should anchor your expectations. On the demand side of belief: a 2025–26 McKinsey survey found that 71% of merchants said AI merchandising tools had so far had limited or no effect on their business, and 61% said their organisation was not at all, or only slightly, prepared to scale AI across merchandising — the reasons cited being fragmented systems, data too messy to generate useful recommendations, and uneven adoption. On the supply side of hype: Gartner has predicted that more than 40% of agentic-AI projects will be cancelled by the end of 2027 on cost, unclear value or weak risk controls, and has coined the term “agent washing” for the rebranding of ordinary chatbots and automation as “agents.”

Neither number says AI does not work for category management. They say something more useful: the constraint is rarely the model — it is the data foundation and the operating model around it. That is precisely why Chapters 9, 10 and 14 of this guide spend more time on architecture, context and roadmap than on models.

71%
of merchants say AI merchandising tools have had limited or no effect so far
Source: McKinsey survey, 2025–26
61%
say their organisation is not ready to scale AI across merchandising
Source: McKinsey survey, 2025–26
>40%
of agentic-AI projects predicted to be cancelled by end of 2027
Source: Gartner, June 2025

The honest mapProven, emerging, hype

Here is the whole landscape, sorted. The badges are deliberate: build your near-term plan on the green rows, pilot the amber, and treat the red as a direction of travel rather than a purchase.

Figure 8.1 — AI capabilities in category management, by maturity
CapabilityStatusWhat to know
AI demand forecastingProvenThe most mature use case; deployed at scale. Consultancy figures cite 20–50% forecast-error reduction — but accuracy varies hugely by category.
Computer-vision shelf monitoringProvenShare-of-shelf, planogram compliance and out-of-stock detection from store photos. Deployed at scale by specialist vendors.
Assortment & planogram optimisationProvenAI recommends SKUs by store cluster and auto-generates planograms — but a human still owns the final layout.
Price, promo & markdown optimisationProvenElasticity and promo-lift models are well established; markdown optimisation is mature in fashion.
Cannibalisation & halo analysisProvenStandard in major planning platforms; identifies which products rise or fall when a SKU or promo changes.
GenAI copilots & conversational analyticsEmergingDeployed, but in a narrow assist role — drafting reviews, summarising insight, “chat with your data.” Not autonomous decision-making.
Reinforcement learning for dynamic pricingEmergingStrong in e-commerce; weaker in physical retail. Carries a specific collusion risk (see Chapter 13).
Agentic / multi-agent category managementEmergingHeavy analyst and vendor attention; real pilots and “early capabilities.” No independently verified at-scale deployment yet.
Synthetic shopper personas & agent-based simsEmergingFast and cheap, but validity is unproven. A complement to real research, not a replacement.
“Fully autonomous category management”HypeDoes not exist at scale today. End-to-end autonomy is a marketing phrase, not a deployed product.

The surpriseMost of the value is not generative AI

The single most useful thing to understand about AI in category management is that the value leaders — forecasting, optimisation, computer vision — run on classical machine learning and operations research, not large language models. Gradient-boosted trees forecast demand. Optimisation engines allocate shelf space. Deep-learning vision models read shelves. These have been quietly deployed for years.

Generative AI and LLMs sit on top of that stack as an interface and productivity layer: they draft the category review, explain a forecast in plain language, summarise a thousand product reviews, and answer questions about the data in conversation. That is genuinely valuable — it is most of what Chapter 11’s prompt library exploits — but it is augmentation, not the engine. The mental model to carry into the rest of Part II: classical ML and optimisation do the calculating; the LLM does the reasoning, drafting and conversing; the agent layer of Chapter 9 orchestrates between them.

Where genAI genuinely helps todayThe deployed copilot use cases

  • Category review automation — drafting performance summaries and recommendation narratives, removing the manual report-assembly that consumes a category manager’s week.
  • Conversational analytics — natural-language question-and-answer over category data, so insight does not wait on a BI ticket.
  • Unstructured-data synthesis — turning reviews, social posts and survey verbatims into category themes.
  • Forecast explanation — putting a plain-language “why” on a number a classical model produced.

Note the pattern: every one of these is language work wrapped around a calculation someone else did. That is the genAI sweet spot in this discipline, and the prompt library is built to exploit exactly it.

How to read a vendor figure

The performance numbers in this field — “20–50% better forecasts,” “5% sales lift,” “4× faster planograms” — are overwhelmingly vendor- or consultant-sourced, and consultancies frequently cite results from their own products. Treat them as claims: attribute them, ask which category and which baseline they refer to, and discount accordingly. Real results vary enormously — a stable grocery staple forecasts far better than a fashion newness item.

Chapter 8 in one line

Forecasting, optimisation and computer vision are proven and run on classical ML; genAI is a real but narrow language layer on top; agentic category management is an emerging pilot, not a finished product — and the binding constraint is almost always data, not models.

Part II · Chapter Nine
09

The Agentic Operating Model

Don’t build one AI that does category management. Build a team of specialists that mirrors the eight-step process — with a human still holding the pen.

The instinct, when people imagine “AI category management,” is to picture a single, all-knowing model that ingests everything and emits a category plan. That instinct is wrong, and following it produces a system you can neither trust nor debug. The discipline already tells you the better architecture. Category management is not one task; it is eight steps, four levers and a review loop. So the AI should not be one agent — it should be a team of specialised agents, each owning a part of the process, coordinated by an orchestrator, and supervised by a human. This chapter specifies that team.

First, definitionsAssistant, agent, multi-agent system

Three words are used loosely; this guide uses them precisely.

Assistant / copilot

Responds to a prompt with text. No tools, no memory of its own goals, no autonomy. You do the thinking; it drafts. This is the bulk of what Chapter 11’s prompts drive.

Agent

Pursues a goal over multiple steps. It can call tools (query data, run a model, write a file), observe results, and decide its next action within guardrails.

Multi-agent system

Several agents, each specialised, coordinated by an orchestrator that decomposes a goal, routes sub-tasks, and assembles the result.

Most organisations should start with assistants — the prompt library is usable today with nothing more — and graduate to agents and a multi-agent system as their data foundation and governance mature. Chapter 14 sequences that journey.

The architectureA nine-agent category team

The reference architecture below maps one agent to each major job in the eight-step process, plus three cross-cutting agents — shopper insight, simulation and governance — that serve all the others. An orchestrator sits above them; a human category manager sits above the orchestrator. Below everything sits the data and tool layer of Chapter 10.

Figure 9.1 — Reference architecture for an agentic category team
Human Category ManagerSets goals · approves plans · owns the decision & the accountability
↓  goals, constraints, approvals  ↑
Orchestrator AgentDecomposes the goal · routes tasks to specialists · sequences the eight steps · assembles the category plan
↓  ↓  ↓  ↓  ↓
Definition AgentBuilds the category definition & the shopper decision tree (Step 1)
Assessment AgentRuns the diagnostic across consumer, market, retailer, supplier (Step 3)
Strategy AgentProposes role, scorecard targets & strategy archetypes (Steps 2,4,5)
Assortment AgentRange optimisation & SKU rationalisation (Step 6)
Pricing AgentPrice architecture, KVIs, elasticity (Step 6)
Space AgentPlanogram logic & shelf adjacency (Step 6)
Promotion AgentPromo calendar, depth & mechanics (Step 6)
Shopper-Insight AgentCross-cutting: missions, personas, unstructured-data synthesis
Simulation AgentCross-cutting: what-if & scenario runs before commitment (Ch. 12)
Governance AgentCross-cutting: guardrails, antitrust & role-coherence checks (Ch. 13)
Data & Tool Layer  —  POS · loyalty · panel · supply chain · competitor · external signals · forecasting & optimisation models · planogram tools

The rosterWhat each agent owns — and what tools it needs

An agent is only as good as the tools it can call. The table below pairs each agent with its job and the tools it must be given access to. “Tool” here means anything the agent can invoke — a database query, a trained model, a piece of planogram software, a web search.

Figure 9.2 — The agent roster and its tools
AgentOwnsNeeds tools for
OrchestratorGoal decomposition, routing, sequencing, plan assemblyCalling other agents; reading the context pack; writing the plan
DefinitionCategory scope & the shopper decision treeQuerying POS/panel data; substitution & switching analysis
AssessmentThe four-lens diagnostic & gap identificationQuerying sales/margin/share data; benchmarking; trend detection
StrategyRole assignment, scorecard targets, strategy archetypesReading the assessment; financial modelling; the role rubric
AssortmentSKU add/keep/cut decisionsAssortment-optimisation model; cannibalisation/halo model
PricingPrice ladders, KVI selection, gap managementPrice-elasticity model; competitor price feed
SpacePlanogram logic, facings, adjacencySpace-planning software; the decision tree; sales-per-facing data
PromotionPromo calendar, depth, mechanicsPromo-lift / causal model; post-event analysis
Shopper-InsightMissions, personas, voice-of-shopper synthesisLoyalty data; review/social text; survey verbatims
SimulationWhat-if scenarios before any plan is committedDemand simulator; the forecasting & elasticity models
GovernanceGuardrails, antitrust, role-coherence, escalationThe rule set; read access to every other agent’s output

The control principleHuman-in-the-loop, by design

Notice what the human at the top of Figure 9.1 does, and what they do not do. They set goals, supply constraints, and approve or reject plans. They do not draft the assessment or tune the planogram. This is the human-in-the-loop (HITL) pattern: the agents do the work; the human owns the decision. A lighter variant, human-on-the-loop (HOTL), has the human monitor a running system and intervene only by exception — appropriate for low-stakes, reversible actions once trust is earned.

The rule of thumb for where to place the human: the more consequential, less reversible and more externally visible the action, the more it requires explicit human approval before execution. Generating a draft assessment is safe to automate fully. Re-pricing a Destination category’s known-value items, or delisting a competitor’s SKU, is not — those cross the Governance Agent’s desk and then the human’s. Chapter 13 makes this an explicit decision rights table.

How a cycle runsOne category review, agent by agent

  1. The human sets the goal: “Run the Q3 review for the coffee category; grow margin without losing penetration.”
  2. The Orchestrator loads the context pack (Chapter 10) and sequences the steps.
  3. The Definition and Shopper-Insight agents confirm the decision tree and the live shopper missions.
  4. The Assessment agent produces the diagnostic and ranks the gaps.
  5. The Strategy agent confirms the role and proposes scorecard targets and a strategy.
  6. The Assortment, Pricing, Space and Promotion agents draft coherent tactics.
  7. The Simulation agent runs the proposed plan against scenarios and reports the expected range of outcomes.
  8. The Governance agent checks coherence, antitrust exposure and guardrail compliance, and flags anything for escalation.
  9. The Orchestrator assembles the plan; the human reviews, edits and approves.

Every numbered step above corresponds to a prompt in Chapter 11. The architecture and the prompt library are the same thing, seen twice.

A pragmatic note on building this

You do not need eleven separate deployed agents to start. A single capable model, given the right system prompt and switched between “roles,” reproduces most of this value on day one — the architecture is a logical design before it is a technical one. Vendors such as the major planning platforms are adding genuine agentic capability, but as Chapter 8 noted, these remain early. Treat Figure 9.1 as the destination and Chapter 14 as the route.

Chapter 9 in one line

Mirror the discipline in the architecture: a specialised agent per step, an orchestrator to coordinate them, cross-cutting insight, simulation and governance agents — and a human who sets the goal and owns the decision.

Part II · Chapter Ten
10

Context Engineering

A model with no context is a confident stranger. Before any prompt earns its keep, you must assemble the category’s context pack.

Here is the most common way an AI category-management initiative fails. Someone opens a chat window, types “optimise my coffee assortment,” receives a fluent, generic, slightly wrong answer, and concludes the technology is not ready. The technology was ready. The context was missing. An LLM with no context about your category is a confident stranger who has never seen your stores — and it will answer anyway. Context engineering is the discipline of assembling, structuring and supplying everything the model needs to stop being a stranger. It is the single highest-leverage activity in Part II, and McKinsey’s finding that “data too messy to generate useful recommendations” is a top failure cause is, at root, a context-engineering failure.

The principleThree kinds of context

Useful context comes in three layers, and an effective system supplies all three.

Static context

Things that rarely change: the category definition, the retailer’s strategy and positioning, the role rubric, the brand architecture, the rules of engagement. Assemble once; revise quarterly.

Dynamic context

The current numbers: POS sales, share, prices, inventory, promo results, availability. Refreshed on a cadence — weekly or faster.

Retrieved context

Pulled on demand for a specific question: a competitor’s latest planogram, a weather outlook, a relevant past category review, a regulatory note.

The deliverableThe Category Context Pack

The practical output of context engineering is an artefact this guide calls the Category Context Pack — a single, structured, version-controlled document (or a small set of them) that travels with the category and is supplied to every agent and every prompt. Building it is mostly a writing and data-assembly job, and it is mostly a one-time cost per category. The template below is the recommended structure; treat the headings as mandatory and the depth as proportional to the category’s role.

TemplateThe Category Context Pack — structure
# CATEGORY CONTEXT PACK
# Version: __ | Owner: __ | Last refreshed: __ | Refresh cadence: __

## 1. IDENTITY & SCOPE  [static]
- Category name and one-line definition
- Sub-segments / the shopper decision tree (attributes, in order)
- What is explicitly IN and OUT of scope
- Retailer banner, store formats, channels covered (store / web / app)
- Geography and market(s)

## 2. STRATEGIC FRAME  [static]
- Retailer positioning and target shopper(s)
- Assigned category role (Destination / Routine / Occasional / Convenience) + why
- Category vision / 3-year ambition
- The role rubric: what each role implies for each lever

## 3. SHOPPER CONTEXT  [static + dynamic]
- Shopper segments and their share of category sales
- Shopping missions in which this category is bought
- Key need-states and how well each is currently served
- Known purchase barriers and triggers

## 4. PERFORMANCE  [dynamic — refresh weekly]
- Sales, units, margin $, margin %, growth vs. prior year
- Category & brand market share and trend
- Penetration, frequency, basket size, attachment
- GMROI, inventory turns, on-shelf availability
- Scorecard: targets vs. actuals

## 5. THE FOUR LEVERS — CURRENT STATE  [dynamic]
- Assortment: SKU count, range list, productivity per SKU, private-label share
- Pricing: price ladders, KVI list, index vs. competitors, PL price gaps
- Space: current planogram, facings, space-to-sales ratio
- Promotion: recent promo calendar, depths, mechanics, measured ROI

## 6. COMPETITIVE & EXTERNAL CONTEXT  [retrieved]
- Key competitors’ assortment, pricing and known activity
- Relevant macro signals: inflation, weather, seasonality, events
- Regulatory or compliance constraints on this category

## 7. RULES OF ENGAGEMENT  [static — critical]
- Hard constraints (min/max SKUs, contractual ranging, legal listings)
- Pricing guardrails (floors, ceilings, index limits)
- Antitrust boundaries (see Chapter 13)
- Decision rights: what AI may draft vs. what needs human approval

## 8. DATA DICTIONARY & PROVENANCE  [static]
- Each data source, its owner, freshness, and known quality limits
- Definitions of every metric used above (no ambiguous terms)
- Known gaps the model must NOT guess to fill

The inputsThe data that feeds the pack

Section 4 onward of the pack is only as good as the data behind it. The load-bearing sources, and what each uniquely provides:

Figure 10.1 — The data inputs of category management
SourceWhat only it can tell you
POS / scanner dataExactly what sold, where, when, at what price — the spine of every assessment
Loyalty dataHousehold-level behaviour — who buys, repeats, switches; the basis of segmentation
Syndicated panel dataCross-retailer market share and competitive context POS cannot see
Supply-chain dataInventory, lead times, fill rates — whether the plan is executable
Competitor dataRivals’ pricing, assortment and share of shelf
External signalsWeather, macro indicators, events — demand drivers outside the category
Unstructured dataReviews, social and survey text — the “why” behind the numbers
Product master dataClean attributes and identifiers — the unglamorous foundation everything else joins on

The disciplineFive rules of context hygiene

  • State provenance and freshness. Every number should carry where it came from and when. A model that knows a figure is three months old can hedge; one that does not will assert.
  • Declare the gaps. Section 8 of the pack explicitly lists what is not known. An LLM’s most dangerous habit is filling silence with plausible invention; naming the gap is the cheapest defence against hallucination.
  • Define every term. “Penetration,” “share” and “margin” each have several meanings. The data dictionary removes the ambiguity so the model and the human mean the same thing.
  • Right-size the pack. More context is not always better — an overstuffed pack buries the signal. Match depth to the category’s role; a Convenience category does not need a Destination category’s dossier.
  • Version it. The pack is a living document under change control, not a one-off export. Stale context is worse than no context, because it looks current.
The uncomfortable truth

Most of the work of “doing category management with AI” is not prompting or agent-building — it is this. Assembling clean, defined, provenance-tagged context is unglamorous and it is the majority of the effort. It is also why the 71% in Chapter 8 saw little impact: they bought the model and skipped the pack. Do not skip the pack.

Chapter 10 in one line

Before any prompt, build the Category Context Pack — a structured, versioned, provenance-tagged document of identity, strategy, shopper, performance, levers, competition, rules and data dictionary — because a model’s output can never be better than the context it was given.

Part II · Chapter Eleven
11

The Prompt Library

Thirteen prompts, one per job in the process. Copy them, fill the placeholders with your Context Pack, and you have an agentic category team you can run today.

This is the working core of the guide. Each prompt below corresponds to an agent from Chapter 9 and a step from the eight-step process. They are written to be used immediately — with a capable model and nothing more than a chat window — and to scale into the multi-agent architecture unchanged, because an agent’s system prompt and a copilot’s instruction are the same artefact.

How to use this library

  • Placeholders appear as {{LIKE_THIS}}. Replace every one before running — most are filled directly from the Category Context Pack of Chapter 10.
  • Always attach the Context Pack. These prompts assume it is supplied. Without it you will get the confident stranger of Chapter 10.
  • Run Prompt 1 first, once per session — it is the system prompt that sets behaviour for everything after.
  • Keep a human in the loop. Every prompt ends by asking the model to flag assumptions and uncertainties. Read those flags; they are the point.

01 · The Orchestrator — system prompt

Set this once at the start of a session, or as the system prompt of the Orchestrator Agent. It establishes role, method, tone and guardrails for every prompt that follows.

SystemCategory Strategist — Orchestrator
You are a senior category management strategist advising a
{{RETAILER_TYPE}} operating in {{MARKET}}. You support —
you do not replace — a human category manager who owns every
final decision.

METHOD. You work the canonical 8-step category management
process: definition, role, assessment, scorecard, strategy,
tactics (assortment / pricing / space / promotion),
implementation, review. You re-anchor every step on the
shopper, not on internal structure.

OPERATING RULES.
1. Use only the supplied Category Context Pack and data. If a
   fact is not in context, say "NOT IN CONTEXT" — never invent
   numbers, shares, prices or shopper claims.
2. Separate fact, inference and assumption. Label each. End
   every output with an "Assumptions & uncertainties" list and
   a "What data would sharpen this" list.
3. Keep every recommendation coherent with the category's
   assigned ROLE. Flag any tactic that contradicts the role.
4. Never recommend actions that disadvantage a named competitor
   on the basis of that competitor's confidential data, and
   never propose coordinating price or promotion with other
   retailers. Route anything near these lines to the Governance
   step. (See the antitrust guardrails.)
5. Quantify with ranges and state confidence. Prefer "likely
   +2 to +5% units, moderate confidence" over false precision.
6. Be concise and decision-useful. Lead with the recommendation,
   then the reasoning, then the evidence.

OUTPUT. Structured markdown: a one-paragraph summary, then
sections, then the two closing lists. Write for a busy
category manager, not for a model.

Acknowledge this briefing in one line, then wait for the
Category Context Pack.

02 · Category Definition & Decision Tree — Step 1

The foundational “definition” prompt. It produces the category scope and the shopper decision tree — the artefact every later prompt depends on. Run it against POS, panel and, ideally, switching data.

DefinitionDefine the category & build the decision tree
TASK. Define the {{CATEGORY}} category and build its shopper
decision tree.

INPUTS (attached): Category Context Pack sections 1, 3 and 5;
{{POS_DATA}}; {{PANEL_OR_SWITCHING_DATA}}; current SKU list.

DO THIS.
1. CATEGORY SCOPE. Propose what is IN and OUT of the category,
   defined by what shoppers treat as substitutable — not by the
   buying desk. State the boundary cases and your call on each.
2. DECISION TREE. Infer the hierarchy of attributes shoppers
   apply when choosing — e.g. form, then variant, then brand,
   then pack size. Give the ORDER of branches and the evidence
   for that order (switching patterns, basket data, attribute
   substitution). Render the tree as an indented list.
3. SEGMENTATION. Turn the tree into the category's sub-segments
   and sub-segments-of-segments. Size each segment by share of
   category sales where data allows.
4. NEED-STATES. Name the distinct shopper need-states the tree
   reveals, and note which are well served vs. under-served by
   the current range.
5. IMPLICATIONS. State, in three bullets, what this tree implies
   for shelf blocking and for assortment.

CONSTRAINTS. Where the data cannot support a branch ordering,
say so and offer the two most likely alternatives rather than
guessing one. Mark every inference as inference.

OUTPUT. (a) Scope statement; (b) decision tree as indented
list; (c) segmentation table with sizes; (d) need-state list;
(e) implications; (f) assumptions & uncertainties; (g) what
data would sharpen this.

03 · Category Assessment — Step 3

The diagnostic. It reads the four classic lenses — consumer, market, retailer, supplier — and, crucially, ranks the opportunities rather than merely listing them.

AssessmentRun the four-lens category diagnostic
TASK. Produce a category assessment for {{CATEGORY}} covering
the period {{PERIOD}}.

INPUTS (attached): Category Context Pack sections 4, 5 and 6;
performance data; competitor data.

ASSESS THROUGH FOUR LENSES.
- CONSUMER/SHOPPER: penetration, frequency, basket, loyalty,
  switching; which need-states win and lose.
- MARKET: category & segment growth, share, price trends,
  competitor moves.
- RETAILER: sales, margin $, margin %, GMROI, turns, on-shelf
  availability, space-to-sales, vs. targets.
- SUPPLIER: brand performance, innovation, private-label
  dynamics.

THEN.
1. For each lens, give 3-5 findings. Each finding = a number, a
   trend, and a plain-language "so what".
2. Identify the GAPS: where performance trails its potential,
   its role, or the competitive benchmark.
3. RANK the gaps into a prioritised opportunity list. Score
   each on (a) size of prize and (b) ease of action — High /
   Medium / Low — and explain the ranking.
4. State the single most important thing this category should
   fix, and why.

CONSTRAINTS. Use only supplied data. Where a benchmark is
missing, say so. Distinguish a real trend from noise; do not
over-read a single period.

OUTPUT. Four lens sections; prioritised opportunity table;
the headline issue; assumptions & uncertainties; what data
would sharpen this.

04 · Category Role Assignment — Step 2

Assigns — or pressure-tests — the category’s role. Best run as a challenge to a role you have proposed, so the model argues both sides.

RoleAssign & stress-test the category role
TASK. Recommend the strategic ROLE for {{CATEGORY}} from:
Destination, Routine/Preferred, Occasional/Seasonal,
Convenience. {{IF_APPLICABLE: We currently treat it as
{{CURRENT_ROLE}} — challenge that.}}

INPUTS (attached): Context Pack sections 2 and 4; the
assessment.

EVALUATE AGAINST FOUR CRITERIA.
1. Importance to the TARGET shopper and their missions.
2. Contribution to the retailer's DIFFERENTIATION & positioning.
3. Competitive dynamics — can we credibly lead here?
4. Financial contribution — sales, margin, traffic, basket.

DO THIS.
1. Score the category on each criterion (High/Med/Low) with a
   one-line justification drawn from the data.
2. Recommend ONE role. Make the argument.
3. Argue the strongest case AGAINST your recommendation, then
   say why you still hold it.
4. State what the role implies for each of the four levers —
   assortment depth, pricing aggressiveness, space priority,
   promotional intensity.
5. Name what would have to change for the role to change.

OUTPUT. Scorecard; recommended role + rationale; the counter-
argument; lever implications; assumptions & uncertainties.

05 · Category Scorecard — Step 4

Turns role and ambition into measurable, owned targets — the contract the rest of the plan is judged against.

ScorecardBuild the category scorecard
TASK. Build a {{HORIZON e.g. 12-month}} scorecard for
{{CATEGORY}}, whose assigned role is {{ROLE}}.

INPUTS (attached): Context Pack sections 2 and 4; the
assessment; the category vision.

DO THIS.
1. Select 6-9 KPIs appropriate to the ROLE. A Destination
   category weights growth, penetration and price image; a
   Convenience category weights margin and GMROI. Justify each
   choice in one line.
2. For each KPI give: current value, proposed target, the
   stretch-vs-realistic logic, and the lever(s) that move it.
3. Make the targets internally consistent — flag if a margin
   target and a price-image target conflict.
4. Propose a review cadence and name the owner role for each
   KPI.
5. State the 2-3 leading indicators to watch as early warnings.

CONSTRAINTS. Anchor every target to a current value from the
data. If no baseline exists, mark the KPI "baseline required"
rather than inventing a target.

OUTPUT. Scorecard table (KPI | current | target | logic |
levers | owner | cadence); leading indicators; consistency
flags; assumptions & uncertainties.

06 · Category Strategy — Step 5

Chooses the strategy archetypes that close the prioritised gaps and hit the scorecard — the bridge from diagnosis to tactics.

StrategyGenerate the category strategy
TASK. Devise the strategy for {{CATEGORY}} to close the
prioritised gaps and deliver the scorecard.

INPUTS (attached): the assessment & opportunity ranking; the
role; the scorecard; Context Pack section 2.

DO THIS.
1. From the classic archetypes — traffic-building,
   transaction-building, profit-generating, cash-generating,
   excitement-creating, image-enhancing, turf-defending —
   select the 2-3 that fit this category's role and gaps.
   Justify the selection; name the archetypes you rejected.
2. For each chosen archetype, state the strategic intent in one
   sentence and the primary lever it will pull.
3. Show the line of sight: gap -> strategy -> the scorecard KPI
   it moves. Every prioritised gap must be addressed by a
   strategy, or explicitly parked with a reason.
4. Identify the main TENSION in the strategy (e.g. growth vs.
   margin) and how you propose to hold it.
5. Brief, in 2-3 lines each, what this implies for assortment,
   pricing, space and promotion — as a handover to Step 6.

OUTPUT. Selected archetypes + rationale; gap-to-strategy-to-
KPI table; the central tension; lever handover briefs;
assumptions & uncertainties.

07 · Assortment & SKU Rationalisation — Step 6

The first tactical lever. Produces an add / keep / cut recommendation anchored to the decision tree, so no need-state is stranded.

AssortmentOptimise the range & rationalise SKUs
TASK. Recommend the assortment for {{CATEGORY}} in
{{STORE_CLUSTER_OR_CHANNEL}}: which SKUs to ADD, KEEP and CUT.

INPUTS (attached): the decision tree & segmentation; current
SKU list with sales, margin, units, productivity; Context Pack
section 5; the strategy; cannibalisation/halo data if available.

DO THIS.
1. Map every current SKU onto its leaf of the decision tree.
   Flag any leaf (need-state) with no strong item — a coverage
   gap — and any leaf with redundant duplicates.
2. CUT candidates: low productivity AND duplicated on its leaf
   AND low loyalty/penetration. For each, estimate the share of
   its volume that will transfer to other SKUs vs. be lost to
   the category, and state the confidence.
3. ADD candidates: unserved or under-served need-states, and
   credible innovation. Justify each against a shopper need.
4. KEEP: confirm the core, including low-volume SKUs that
   uniquely hold a need-state or a shopper segment.
5. Net effect: projected change in SKU count, sales and margin,
   as a range, and the key risk.

CONSTRAINTS. Respect the hard constraints in Context Pack
section 7 (min/max SKUs, contractual ranging, legal listings).
Never cut the only item on a leaf. Treat private label and
national brands by the same evidence standard.

OUTPUT. SKU table (SKU | leaf | action | rationale | volume-
transfer estimate); coverage-gap list; net effect with range;
risks; assumptions & uncertainties.

08 · Pricing Architecture — Step 6

The second lever. Builds coherent price ladders, selects known-value items and manages private-label gaps — within explicit guardrails.

PricingDesign the category price architecture
TASK. Design the price architecture for {{CATEGORY}}.

INPUTS (attached): SKU list with current prices, costs,
margins, elasticities if available; competitor price index;
the role; the strategy; Context Pack sections 5 and 7.

DO THIS.
1. PRICE LADDER. Lay out the price tiers across segments and
   good-better-best. Flag gaps that are too wide (a hole in the
   ladder) or too narrow (tiers that do not feel distinct).
2. KVIs. Recommend the known-value items whose price defines
   shopper price perception for this category, and the index
   at which each should sit vs. competitors. Tie aggressiveness
   to the ROLE.
3. PRIVATE-LABEL GAPS. Recommend the price gap between private
   label and the comparable national brand for each tier, and
   the cannibalisation risk of each gap.
4. CHANGES. List proposed price changes; for each, estimate the
   volume and margin effect as a RANGE, using elasticity where
   supplied and stating where you are extrapolating.
5. Flag any change that conflicts with the category's price-
   image target.

CONSTRAINTS. Respect every pricing guardrail in Context Pack
section 7 — floors, ceilings, index limits. Do NOT propose
prices coordinated with other retailers or set using rivals'
confidential data. Where elasticity is unknown, say so and
give a wider range.

OUTPUT. Price-ladder view; KVI table with target indices;
private-label gap recommendations; price-change list with
ranged effects; guardrail & image flags; assumptions &
uncertainties.

09 · Space & Planogram Logic — Step 6

The third lever. Produces the logic of the planogram — blocking, facings, adjacencies — as a brief a space-planning tool or merchandiser then executes.

SpaceGenerate the planogram logic
TASK. Produce the shelf-layout logic for {{CATEGORY}} in a
{{FIXTURE / BAY SIZE}} fixture.

INPUTS (attached): the decision tree; the recommended
assortment; sales and margin per SKU; sales-per-facing data;
current planogram; Context Pack section 5.

DO THIS.
1. BLOCKING. Specify how the shelf is blocked — which attribute
   blocks first, second, third — and make it MIRROR the
   decision tree, so a shopper finds their branch fast.
2. FACINGS. Recommend facings per SKU, allocating space toward
   sales and margin contribution while keeping minimum facings
   for availability on slow but need-state-critical items.
3. VERTICAL/HORIZONTAL POSITION. Place high-demand and high-
   margin items at eye level; state the rationale for each
   notable placement.
4. ADJACENCIES. Recommend cross-category and within-category
   adjacencies that support basket-building, with the shopper
   logic for each.
5. Identify the 3 changes vs. the current planogram with the
   biggest expected effect, and quantify the expected effect as
   a range.

CONSTRAINTS. This is layout LOGIC, not a finished planogram —
a space-planning tool or merchandiser executes it. Respect
fixture limits and any minimum-facing rules. State where a
recommendation needs in-store testing before rollout.

OUTPUT. Blocking scheme; facing table; placement rationale;
adjacency recommendations; top-3 high-impact changes with
ranged effects; assumptions & uncertainties.

10 · Promotion Plan — Step 6

The fourth lever. Builds a promotional calendar judged on incremental, not headline, lift — and links each event to a strategy.

PromotionBuild the promotional plan
TASK. Build a {{PERIOD}} promotional plan for {{CATEGORY}}.

INPUTS (attached): past promo events with measured lift, depth,
mechanic and ROI; the strategy; the scorecard; seasonality;
Context Pack sections 4 and 6.

DO THIS.
1. Review past events. Distinguish INCREMENTAL lift from pull-
   forward and from cannibalised volume within the category.
   Name which past mechanics genuinely paid back.
2. Propose the calendar: events, timing, SKUs, depth and
   mechanic (feature, display, multibuy, coupon, loyalty offer).
   Tie each event to a strategy archetype and a scorecard KPI.
3. For each event, estimate incremental units, incremental
   margin and ROI as a RANGE, and state the cannibalisation and
   pull-forward assumptions behind it.
4. Balance the calendar against the ROLE — a Convenience
   category should not be promoted like a Destination one.
5. Flag events whose realistic ROI is negative or unknown, and
   recommend test-and-learn instead of full rollout for those.

CONSTRAINTS. Headline sales lift is not success — judge on
incremental margin. Do not coordinate timing or depth with
other retailers. Where lift data is thin, recommend a test.

OUTPUT. Promo calendar table (event | timing | SKUs | depth |
mechanic | strategy | KPI | ranged ROI); the events to test
rather than roll out; assumptions & uncertainties.

11 · Simulation Scenario Design

Designs the what-if scenarios that pressure-test a plan before it is committed — the brief the Simulation Agent of Chapter 12 then runs.

SimulationDesign the scenario set
TASK. Design a scenario set to stress-test the proposed plan
for {{CATEGORY}} before it is committed.

INPUTS (attached): the draft category plan (assortment,
pricing, space, promotion); the scorecard; Context Pack
sections 4 and 6.

DO THIS.
1. BASE CASE. State the plan's expected outcome on each
   scorecard KPI, with the assumptions it depends on.
2. SENSITIVITY. Identify the 4-6 assumptions the outcome is
   most sensitive to (e.g. cut-SKU volume transfer rate, KVI
   elasticity, promo incrementality, competitor response).
3. SCENARIOS. Design a set covering: an optimistic case, a
   pessimistic case, a competitor-reaction case, a demand-shock
   case (weather/macro), and an execution-shortfall case. For
   each, specify exactly which input variables move and by how
   much.
4. For each scenario, state what to measure and the decision it
   informs ("if the pessimistic case still clears margin
   target, proceed; if not, hold").
5. Define the KILL CRITERIA: the scenario outcomes that should
   stop the plan.

CONSTRAINTS. A simulation is only as good as its assumptions —
make every assumption explicit and testable. Do not present
simulated numbers as forecasts; they are conditional outcomes.

OUTPUT. Base case; ranked sensitivity list; scenario
specifications table; what each scenario decides; kill
criteria; assumptions & uncertainties.

12 · Synthetic Shopper Persona

Generates a structured shopper persona for qualitative pressure-testing. Use with the caution of Chapter 12: a complement to real research, never a replacement.

InsightSynthesise a shopper persona & react
TASK. Construct a synthetic shopper persona for {{CATEGORY}}
and have it react to the draft plan.

INPUTS (attached): shopper segment data; missions; need-states;
the draft plan element to test, e.g. {{NEW_RANGE_OR_LAYOUT}}.

DO THIS.
1. Build the persona from the REAL segment data supplied:
   demographics, the missions they shop this category on, their
   decision-tree path, their barriers and triggers. Ground every
   trait in the data — cite which segment statistic supports it.
2. In the persona's voice, walk their path to purchase for this
   category: trigger, channel, in-aisle decision, choice.
3. React to the proposed change: what they would notice, value,
   ignore or be frustrated by, and why.
4. Surface 3 risks or unmet needs the plan may have missed.

CONSTRAINTS — READ CAREFULLY. This persona is a structured
hypothesis generator, NOT evidence. It cannot tell you what
real shoppers will do. Use it to find questions, not answers.
Do not let any number it "reports" enter a forecast. Flag every
trait that is extrapolated beyond the supplied data. Validate
all findings with real shopper research before acting.

OUTPUT. Persona profile (data-grounded); the narrated path;
the reaction; 3 risks/unmet needs; an explicit list of which
claims need real-research validation.

13 · Governance & Red-Team Review

The final gate. Run this against any completed plan before it reaches a human approver — it is the prompt behind the Governance Agent of Chapter 13.

GovernanceRed-team the category plan
TASK. Adversarially review the completed category plan for
{{CATEGORY}} before it goes to a human approver. Assume the
plan is flawed; your job is to find how.

INPUTS (attached): the full plan (all 8 steps); the Context
Pack including section 7 (rules of engagement).

CHECK, AND REPORT A PASS/FLAG/FAIL ON EACH.
1. ROLE COHERENCE. Does every tactic match the assigned role?
   Flag any contradiction (e.g. a Convenience category given
   Destination-level promotion).
2. INTERNAL COHERENCE. Do assortment, pricing, space and
   promotion reinforce each other, or fight? Do the tactics
   actually deliver the scorecard?
3. ANTITRUST. Does anything (a) rely on a competitor's
   confidential data, (b) recommend disadvantaging a named rival
   beyond fair category logic, (c) imply coordinating price or
   promotion with other retailers, or (d) read as self-
   preferencing if the analysis came from a category captain?
   Treat any hit here as FAIL — escalate to legal.
4. EVIDENCE. Are claims backed by supplied data? List every
   number that is asserted without provenance, and every place
   the model may have filled a gap by invention.
5. CONSTRAINTS. Are all hard constraints in Context Pack
   section 7 respected?
6. RISK & REVERSIBILITY. Which actions are hard to reverse or
   externally visible? Confirm each has explicit human sign-off.
7. EQUITY. Could the plan systematically disadvantage smaller
   suppliers or particular store types without a sound reason?

OUTPUT. A checklist with PASS / FLAG / FAIL and a note per
item; the must-fix list before approval; the items to escalate
to legal or to a human; an overall recommendation: APPROVE /
REVISE / HALT.
Chapter 11 in one line

Thirteen prompts — one orchestrator system prompt and twelve task prompts mapped to the eight-step process — turn a capable model into an agentic category team; their power comes from the Context Pack you attach and the assumption flags you read.

Part II · Chapter Twelve
12

Simulation & Scenario Labs

A category plan is a bet. Simulation lets you place the bet a hundred times before you place it once.

Every category plan is a forecast in disguise. “Cut these twelve SKUs” assumes shoppers will switch rather than walk; “raise this known-value item 4%” assumes an elasticity; “run this multibuy” assumes an incremental lift. Traditionally those assumptions were tested the expensive way: in the actual stores, with the actual quarter, and you found out you were wrong by missing the number. Simulation moves the test before the commitment. It is the part of the AI playbook that the question “are there options for simulation runs?” points straight at — and the answer is a clear yes, across five distinct techniques of very different maturity.

Why simulateThe three jobs of a simulation

Simulation does three things a static plan cannot. It quantifies uncertainty — turning a single projected number into a range of outcomes with probabilities. It stress-tests assumptions — showing which inputs the plan’s success actually hinges on, so attention goes to the ones that matter. And it de-risks the irreversible — letting you discover that a planogram reset or a delisting fails in silico, where failure costs a compute run rather than a quarter.

The optionsFive simulation techniques

“Simulation” is not one thing. Five techniques are in use, and they sit at very different points on the proven-to-emerging scale of Chapter 8. Choose by the question you are asking.

Figure 12.1 — Five simulation techniques for category management
TechniqueStatusAnswers the question…
What-if / scenario modellingProven“If I change assortment / price / space / promo, what happens to the scorecard?” The everyday workhorse, built into major planning platforms.
Monte Carlo demand simulationProven“Given uncertain demand, what is the range of outcomes and the chance I miss target?” Standard for demand and safety-stock.
Virtual / VR shelf testingEmerging“Will shoppers notice and navigate this new layout?” Real human subjects in a virtual store, often with eye-tracking.
Store / category digital twinEmerging“How does the whole category behave as a system under this plan?” A live virtual replica; deployed in pockets.
Agent-based shopper modellingResearch“What emerges when thousands of simulated shoppers each follow their own decision rules?” Largely academic; not yet a standard commercial tool.

Technique 1–2The two you should use now

What-if scenario modelling is the one to start with, because it is proven and already sits in the planning tools most retailers own. You change an input — pull a SKU, move a price, reallocate facings — and the tool re-runs the demand and financial model to show the effect on the scorecard. Run before a plan is submitted, it converts category management “from intuition to financial simulation,” in the words the vendors use. The Simulation Agent of Chapter 9 is, at its simplest, an agent that drives this capability through the scenario set that Prompt 11 designed.

Monte Carlo demand simulation adds the dimension what-if modelling lacks: probability. Instead of one demand number, you specify demand as a distribution, draw from it thousands of times, and read out the full spread of outcomes — the median, the bad tail, the probability of missing the margin target. For any plan whose success is genuinely uncertain — a new range, an aggressive price move — a Monte Carlo run is the honest way to present it to an approver: not “this delivers +3%” but “this delivers between −1% and +6%, with a 70% chance of clearing target.”

Technique 3–4The two to pilot

Virtual and VR shelf testing is different in kind: it puts real human shoppers into a photorealistic virtual store and watches them shop — navigation, hesitation, what they pick — often with eye-tracking that yields an attention heatmap of the shelf. It is the cheapest, fastest way to test whether a new planogram is actually shoppable before a single store is reset; vendors cite testing several times faster and cheaper than physical mock-ups. Treat it as a pilot-grade capability with real value for high-stakes resets.

A store or category digital twin is a live virtual replica of the category — its SKUs, its shelf, its demand — that a plan can be run against as a system rather than a spreadsheet. It is genuinely emerging: deployed in pockets, promising, not yet routine. Pilot it where the category is large and complex enough to justify the modelling cost.

Technique 5The one to watch, not buy

Agent-based shopper modelling — and its cousin, the LLM-generated synthetic shopper persona of Prompt 12 — is the frontier. The idea is seductive: populate a simulation with thousands of autonomous simulated shoppers, each with their own decision tree and budget, and watch category-level behaviour emerge. Academic frameworks exist for generating synthetic transaction data this way. But the validity is unproven, and the failure mode is specific and dangerous: a synthetic shopper is fluent and confident, and it is tempting to treat its “answer” as data. It is not data. It is a structured hypothesis. Use it to generate questions for real research, never to replace it.

The cardinal rule of simulation

A simulation is only ever as good as its assumptions, and a simulated number is a conditional outcome, not a forecast. The danger is laundering: an assumption goes in soft and fuzzy, runs through a model, and comes out as a hard-looking number on a slide. Defend against it the way Prompt 11 does — make every assumption explicit and testable, present outcomes as ranges, and never let a synthetic-persona “result” enter a real forecast.

In practiceThe simulation loop

Simulation is not a one-off; it is a loop that wraps the plan before the plan reaches a human:

  1. Design the scenario set — Prompt 11 turns a draft plan into a base case, a sensitivity list and a set of named scenarios with explicit kill criteria.
  2. Run the scenarios — what-if for the directional question, Monte Carlo for the probabilistic one, virtual shelf testing for the navigability question.
  3. Read the results as ranges and probabilities, against the kill criteria.
  4. Revise the plan where a scenario breaks it — or halt if a kill criterion fires.
  5. Hand the surviving plan, with its outcome ranges attached, to the Governance Agent and then the human approver.

Done this way, simulation is not a fancy add-on. It is the difference between presenting an approver a number and presenting them a bet they can actually price.

Chapter 12 in one line

There are five simulation options — start with proven what-if and Monte Carlo, pilot virtual shelf testing and digital twins, and watch agent-based modelling — and use all of them to turn a plan from a single number into a priced, stress-tested, kill-criteria-bounded bet.

Part II · Chapter Thirteen
13

Governance, Risk & Antitrust

An AI that helps you manage a category can also, unsupervised, break competition law on your behalf. This chapter is the brakes.

Category management was always a discipline with a legal edge. Chapter 3 explained why: the category-captain model asks a self-interested competitor to advise on its rivals’ shelf treatment, and regulators have watched it for two decades. Add AI — faster, more autonomous, harder to interrogate — and every one of those old concerns sharpens. This chapter covers the five risks that matter and the governance that contains them. It is not optional reading. An agentic category system without this chapter is a liability with a user interface.

Risk 1The AI category captain

The traditional category-captain concern is that a leading supplier, advising on the whole category, quietly favours itself — competitive exclusion — or becomes a conduit for coordination. An AI category captain amplifies this in three ways. It operates at a scale and speed no human captain could, so a subtle bias compounds fast. It is harder to interrogate — “the model recommended it” is not an audit trail. And it can ingest and act on more data, including data it should never have touched. The mitigations are the classic ones, made firmer: an AI captain advises, never decides; the retailer retains and exercises final authority; a validator — human or a separately-instructed model — reviews; and the captain’s AI is provably walled off from rivals’ confidential data.

Risk 2Algorithmic price collusion

This is the risk regulators are most actively pursuing, and it has a specific technical root. In a much-cited 2021 simulation, several independent pricing algorithms using reinforcement learning — with no instruction to collude and no communication — spontaneously learned collusive, supra-competitive pricing, simply because mutual restraint was the strategy that maximised each agent’s reward. Regulators have taken the lesson: US enforcers treat algorithmic price-fixing as ordinary price-fixing, and a landmark case over shared pricing software has put the whole field on notice. For category management the rule is blunt: a pricing agent must never be allowed to set prices using rivals’ confidential or signalled data, and never optimise toward an objective that rewards tacit coordination. Prompts 1, 8 and 13 encode exactly this.

Risk 3Hallucination & invented evidence

An LLM’s defining failure mode is the confident, fluent, wrong answer — and in category management the wrong answer wears the costume of a precise number. “Private label is losing 3.2 points of share” reads identically whether it came from the data or from the model’s imagination. Surveys of enterprise LLM use trace a large share of hallucinations to data problems — missing context, training bias. The defences run through this whole guide: the Context Pack’s explicit gap list (Chapter 10), the system prompt’s “NOT IN CONTEXT” rule (Prompt 1), the assumption-and-uncertainty list every prompt demands, and the Governance prompt’s hunt for unsourced numbers (Prompt 13).

Risk 4Data bias & supplier equity

A model trained on biased history reproduces and amplifies the bias. In category management this shows up as systematically uneven output across store types, shopper demographics or regions — and, pointedly, as a structural tilt against smaller suppliers, whose thinner data trail makes them easier for an optimisation to quietly de-range. A plan that cuts small suppliers because they are small, rather than because they fail a fair category test, is both an equity problem and a competition-law exposure. Prompt 13’s equity check exists for this; so does the rule, in Prompt 7, to judge private label and national brands by one evidence standard.

Risk 5Opacity & the un-auditable decision

A recommendation no one can explain is a recommendation no one can defend — to a regulator, to a supplier, to your own board. Research on AI adoption is consistent: people will not, and should not, act on black-box outputs. The answer is explainability by construction: every agent states its reasoning and its evidence (Prompt 1), every plan carries its assumption list, and the whole pipeline keeps an audit trail — what data went in, which agent recommended what, which human approved it, when. If you cannot reconstruct a category decision six months later, you do not have governance; you have luck.

The controlA decision-rights table

The single most important governance instrument is also the simplest: an explicit table of who — or what — may decide what. The principle, from Chapter 9: the more consequential, less reversible and more externally visible an action, the more it demands human approval before execution.

Figure 13.1 — A category-management decision-rights table
ActionAI may…Human approval
Draft assessment, decision tree, scorecardProduce in fullReview only
Recommend role & strategyRecommend with rationaleRequired before adoption
Assortment add / keepRecommendRequired before execution
SKU delisting (esp. a competitor’s)Recommend & justify onlyRequired — named approver
Price changes within set guardrailsRecommend; auto-execute only if explicitly authorised & loggedRequired for anything outside guardrails
KVI re-pricing on a Destination categoryRecommend onlyRequired — senior approver
Promotional calendarDraft in fullRequired before commitment
Anything flagged by the Governance AgentHalt & escalateRequired — legal where antitrust

The controlSix guardrails to put in place first

  • The Governance Agent runs last and can halt. No plan reaches a human approver without passing Prompt 13; a FAIL stops the plan.
  • Confidential-data walls are technical, not promised. A captain’s AI is architecturally prevented from accessing rivals’ confidential data — not merely instructed to ignore it.
  • No price coordination, ever. Pricing agents are barred from rivals’ signalled data and from objectives that reward coordination; this is checked, not assumed.
  • Every plan carries its assumptions and provenance. No number without a source; no recommendation without its reasoning.
  • The full pipeline is logged. Inputs, agent outputs, approvals and timestamps are retained as an audit trail.
  • A human owns every decision. The AI is accountable to no regulator; a named person is. Decision rights are written down and honoured.
Consult counsel — this is not legal advice

Competition law differs by jurisdiction and is moving fast, particularly around algorithmic pricing and AI. The direct legal analysis of AI category captains specifically is still thin — it is an emerging area, not settled law. This chapter is a practitioner’s risk map, not legal advice. Before deploying any agent that touches pricing, assortment across competing brands, or competitor data, involve qualified competition counsel.

Chapter 13 in one line

AI sharpens five risks — the AI captain, algorithmic collusion, hallucination, data bias and opacity — and the containment is concrete: a Governance Agent that can halt, technical data walls, a written decision-rights table, a full audit trail, and a named human who owns every decision.

Part II · Chapter Fourteen
14

The Implementation Roadmap

A maturity model to find your starting point, and a 90-day plan to take the first real step. Begin small, begin with data, begin now.

Everything in Part II converges here. You have a state-of-the-art map (Chapter 8), an architecture (9), a context discipline (10), a prompt library (11), a simulation lab (12) and a governance frame (13). What remains is sequencing — because the organisations in Chapter 8’s 71% did not fail for lack of ambition; they failed by skipping straight to the ambitious part. This chapter gives you a maturity model to locate yourself honestly and a 90-day plan to make the first move count.

Where you standA five-stage maturity model

No single industry-standard maturity model for AI in category management exists; the ladder below adapts general AI-maturity thinking to this discipline. Locate yourself honestly — most organisations are at Stage 1 or 2, and naming that is the start of progress, not an admission of failure.

Stage 1

Manual. Category management runs on spreadsheets and syndicated reports. AI, if present, is an individual experimenting in a chat window. No Context Pack, no governance.

Stage 2

Assisted. Category managers use AI copilots for drafting, summarising and analysis. Context Packs exist for key categories. Value is real but bounded by the human’s available time.

Stage 3

Augmented. Classical ML — forecasting, assortment, planogram optimisation — is deployed and trusted. Copilots are routine. The data foundation is unified enough to be reliable.

Stage 4

Agentic. Specialised agents run defined steps of the process under an orchestrator. A Governance Agent gates every plan. Humans set goals and approve; agents do the work.

Stage 5

Orchestrated. The continuous category cycle runs as a supervised multi-agent system across the portfolio. Humans manage by exception and strategy. A frontier few reach — and none fully, today.

How to climbCrawl, walk, run

The ladder maps cleanly onto a sequencing rule. Crawl (Stages 1–2): adopt the proven, low-risk wins — AI demand forecasting, computer-vision shelf monitoring — and put copilots in category managers’ hands using the prompt library. Walk (Stage 3): deploy AI assortment and planogram optimisation, and — the real work — unify the data foundation so recommendations can be trusted. Run (Stages 4–5): stand up the agentic operating model, one agent and one category at a time, behind the governance frame. The cardinal error is running before you can walk: agents on a messy data foundation simply automate the production of wrong answers.

The constraint is never the model. It is the data foundation and the operating model around it.
The lesson of Chapter 8, restated as a plan

The first moveA 90-day plan

Ambition fails on vagueness, so here is a concrete first quarter. It assumes you are at Stage 1 or 2 and want a credible, low-risk, value-proving step toward Stage 3–4. Pick one category — important enough to matter, contained enough to finish — and run this.

Days 1–30 — Foundation

  • Choose the pilot category and name a human owner accountable for the pilot.
  • Build the Category Context Pack for it — the full Chapter 10 template. This is most of the month’s work and the point of it.
  • Audit the data behind the pack: provenance, freshness, gaps. Fix what is cheap to fix; document the rest in section 8.
  • Write the rules of engagement — constraints, guardrails, decision rights — and have competition counsel review the antitrust boundaries.
  • Define success: the two or three metrics by which the pilot will be judged in day 90.

Days 31–60 — Assisted run

  • Run the prompt library end to end on the pilot category — Prompts 1 through 13 — as copilots, with the category manager driving.
  • Compare the AI-assisted outputs against how the category was last reviewed manually. Where do they agree? Where does the AI surface something missed — or get something wrong?
  • Run the Governance prompt (13) on the resulting plan and act on its flags.
  • Capture the friction: every place the model lacked context, every prompt that needed rewording. This feedback improves the pack and the prompts.

Days 61–90 — Simulate, decide, scale-plan

  • Run the simulation loop (Chapter 12) on the plan — what-if and Monte Carlo at minimum — and present outcomes as ranges with kill criteria.
  • Take the plan through real human approval using the decision-rights table. Execute what is approved.
  • Measure against the day-1 success metrics. Be honest; a clear negative result is a real result.
  • Write the scale decision: which categories next, which agent from Chapter 9 to stand up first, what the data foundation needs before Stage 4.
Figure 14.1 — The 90-day plan at a glance
PhaseFocusKey deliverable
Days 1–30FoundationA complete Context Pack, a data audit, written rules of engagement
Days 31–60Assisted runThe prompt library run end-to-end; a governed draft plan
Days 61–90Simulate & decideA simulated, approved, executed plan; a measured result; a scale decision

The five rulesWhat separates the 29% from the 71%

  1. Data before models. The Context Pack is the project. Skip it and you join the 71%.
  2. Proven before emerging. Bank the forecasting and copilot wins before piloting agents.
  3. One category before the portfolio. Prove it small, learn, then scale — never the reverse.
  4. Governance from day one. The decision-rights table and the Governance Agent are not a later add-on.
  5. A human owns it. Every plan, every decision, every quarter — named, accountable, in the loop.
Chapter 14 in one line

Find your stage honestly, climb crawl-walk-run, and make the first move a 90-day, single-category pilot that builds the Context Pack, runs the prompt library, simulates the plan and measures the result — because data, sequence and a named human owner are what separate success from the 71%.

Closing — the argument of this guide, in one paragraph

Category management has, for forty years, been the discipline of running each product category as a business — with a role, a strategy, a scorecard and a cycle. AI does not change that definition; it changes the tempo and reach of the work. A team of specialised agents, fed a well-engineered Context Pack, driven by a disciplined prompt library, pressure-tested in simulation and bounded by hard governance, lets a category manager run the eight-step cycle more often, across more categories, with the diagnosis done and the options costed before they pick up the pen. The technology is real, the limits are real, and the order of operations is everything. Build the data foundation, start with one category, keep a human accountable — and the agentic category team stops being a slide and starts being how the work gets done.

Appendix A
A

Glossary of Terms

Sixty working definitions — the shared vocabulary of category management and of its AI.

A — Category management & retail terms
TermDefinition
AssortmentThe set of SKUs a category carries — its breadth and depth. One of the four tactical levers.
Attachment / attach rateThe rate at which a complementary product is bought alongside a primary one.
Basket analysisStudy of what is bought together in a transaction; yields items-per-basket and cross-sell rates.
Buying & merchandisingApparel’s parallel to category management, structured department–class–subclass and governed by Open-To-Buy.
CannibalisationThe loss of sales from existing SKUs caused by introducing or promoting another. Paired with the halo effect.
CategoryA group of products that shoppers perceive as related or substitutable; the unit of management.
Category captainA leading supplier designated to provide whole-category analysis and recommendations to a retailer.
Category definitionStep 1: deciding which products are in and out of a category, through the shopper’s eyes.
Category managementManaging each product category as a strategic business unit with its own role, strategy and scorecard.
Category roleThe strategic job assigned to a category — Destination, Routine, Occasional/Seasonal or Convenience.
Consumer Decision Tree (CDT)A map of the hierarchy of attributes a shopper applies when choosing within a category. Also Shopper/Category Decision Tree.
Convenience (role)A category that completes the basket and supports one-stop shopping; managed for margin.
CPFRCollaborative Planning, Forecasting & Replenishment — the supply-side counterpart to category management.
Destination (role)A category that defines the store and drives store choice; the deepest resource commitment.
Digital shelfThe online equivalent of the physical shelf — search results, ranking, browse pages, recommendations.
ECREfficient Consumer Response — the 1990s movement from which category management emerged as the demand-side process.
EDLP / Hi-LoEvery Day Low Price vs. high–low promotional pricing — two opposing pricing postures.
FacingA single product unit visible at the front of a shelf; the basic unit of space allocation.
GMROIGross Margin Return on Inventory Investment — gross margin ÷ average inventory cost.
Halo effectThe lift in other products’ sales caused by introducing or promoting a SKU. The opposite of cannibalisation.
Hard discountA retail model (Aldi, Lidl) with very few SKUs and mostly private label; inverts the classic CM model.
Joint Business Planning (JBP)A multi-year retailer–supplier agreement that commits funding and targets behind category plans.
KPIKey Performance Indicator — a measure on the category scorecard.
Kraljic matrixThe 1983 procurement framework segmenting spend by profit impact × supply risk.
KVIKnown-Value Item — a product whose price defines shopper price perception for the category.
Line reviewThe DIY/home-improvement form of the category review — a periodic re-evaluation of every vendor.
Loyalty dataHousehold-level purchase data from loyalty programmes; the basis of shopper segmentation.
Open-To-Buy (OTB)An apparel buyer’s budget for what may be purchased in a period.
Panel dataPurchase data from a representative recruited household panel; reveals cross-retailer behaviour.
Path to purchaseThe full multi-channel journey from a need to a purchase and beyond.
PenetrationThe share of baskets or households that contain a category or product.
PlanogramA visual schematic specifying where and how many facings of each product appear on a shelf.
POS dataPoint-of-sale (scanner) data — item-store-day transaction records; the spine of category analysis.
Price ladderThe structured set of price tiers across a category’s segments and good-better-best architecture.
Private labelRetailer-owned brands, typically in economy, standard and premium tiers.
Procurement category managementThe buy-side discipline of grouping purchasing spend into categories and sourcing each strategically.
Range reviewA periodic re-evaluation of which SKUs a category carries — repeat, replace, retire.
Retail media network (RMN)A retailer selling advertising on its own first-party shopper data.
Revenue Growth Management (RGM)Optimising net revenue and margin across pricing, promotion, pack architecture, trade terms and channel mix.
Routine (role)An everyday-need category where the retailer aims to be consistently preferred; the workhorse role.
ScorecardThe set of targets and KPIs against which a category is measured (Step 4).
Shopper vs. consumerThe shopper makes the purchase; the consumer uses the product. Often different people.
Shopper marketing“Pull” marketing aimed at the shopper at or near the point of sale.
Shopper missionThe motivation, context and behaviour defining a shopping trip — stock-up, top-up, food-to-go, etc.
SKUStock Keeping Unit — an individual, distinctly identified product item.
SKU rationalisationRemoving underperforming or duplicative SKUs to raise category productivity.
Slotting feeA payment from a supplier to a retailer for shelf space; the practice CM partly displaced.
Space planningAllocating shelf space and designing planograms; the execution layer of the space lever.
Syndicated dataStandardised cross-retailer data sold by measurement firms (NielsenIQ, Circana); enables benchmarking.
Trade marketing“Push” B2B marketing aimed at channel intermediaries — retailers, distributors, wholesalers.
UPCUniversal Product Code — the 12-digit barcode that made retail measurable.
Validator / advisorA second supplier, or a neutral party, that reviews and challenges a category captain’s recommendations.
Vendor managerThe first-party e-commerce role owning a brand’s wholesale relationship, terms and assortment.
B — AI & agentic terms
TermDefinition
AgentAn AI system that pursues a goal over multiple steps, calling tools and deciding its next action within guardrails.
Agentic AIAI built around autonomous, goal-directed agents rather than single prompt-response exchanges.
Assistant / copilotAn AI that responds to prompts with text but has no autonomy or tools; the lightest form of AI help.
Audit trailA retained record of inputs, agent outputs and human approvals — the basis of AI governance.
Category Context PackThis guide’s term for the structured, versioned context document supplied to every agent and prompt.
Computer visionAI that reads images — used for shelf audits, share-of-shelf and out-of-stock detection.
Context engineeringThe discipline of assembling and supplying everything a model needs to answer well.
Demand forecastingPredicting future demand; the most mature, widely deployed AI use case in category management.
Digital twinA live virtual replica of a store or category that plans can be simulated against.
Generative AI / LLMAI that generates language; in CM, a productivity layer for drafting, summarising and conversing.
Governance AgentA cross-cutting agent that checks plans for coherence, antitrust exposure and guardrail compliance.
HallucinationA confident, fluent, false output from a language model.
Human-in-the-loop (HITL)A control pattern where a human approves each consequential decision; HOTL monitors by exception.
Monte Carlo simulationDrawing repeatedly from input distributions to produce a probabilistic range of outcomes.
Multi-agent systemSeveral specialised agents coordinated by an orchestrator.
OrchestratorThe agent that decomposes a goal, routes sub-tasks to specialists and assembles the result.
PromptThe instruction given to a model; in an agentic system, also an agent’s system instruction.
Reinforcement learningML that learns by reward; used for dynamic pricing — and a documented source of collusion risk.
Synthetic shopper personaAn LLM-generated simulated shopper; a hypothesis generator, never a substitute for real research.
What-if modellingRe-running a demand and financial model under changed inputs to see the effect on the scorecard.
Appendix B
B

Sources & Further Reading

The research base, organised by theme. Performance figures attributed to vendors and consultancies are claims, not established fact.

This guide synthesises industry, analyst, regulatory, academic and vendor sources. The list below is representative rather than exhaustive; URLs were valid at the time of research in 2026. Several primary regulatory documents (FTC statements, The Partnering Group materials) were referenced via secondary summaries where the originals were access-restricted — for an authoritative citation, consult the primary documents directly.

Foundations, history & methods (Chapters 1–4)

  1. Wikipedia — “Category management”; “Brian F. Harris”; “Efficient Consumer Response”; “IRI (company)”; “Growth–share matrix.”
  2. ECR Community — ecr-community.org, category management overview; ECR CatMan Network, ecrcatmannetwork.com (history of the 8-step approach).
  3. The Partnering Group — thepartneringgroup.com (Brian Harris biography; Manufacturer Category Vision Program; Omni-channel Category Management; Joint Business Planning).
  4. Umbrex — “Category Management 8-Step Process” and “Category Role Framework.”
  5. NielsenIQ — “Exploring category management processes, steps and business benefits”; “Shopper-centric category management”; “Conquering the retail shelf.”
  6. RELEX Solutions — “Consumer Decision Trees”; “SKU rationalization”; “Cannibalization & halo effects.”
  7. IBM — “The UPC”; GS1 — “50 Years of GS1.”
  8. Consumer Brands Association — the 1993 Kurt Salmon Associates ECR report; FundingUniverse — history of Information Resources, Inc.
  9. dunnhumby — “Shopper-centric category management”; “Smarter category management.”

Related disciplines & procurement (Chapters 5 & 7)

  1. CIPS — cips.org Intelligence Hub: “Category Management Cycle”; “Differences with Strategic Sourcing”; “Procurement & Supply Cycle.”
  2. Kraljic, P. — “Purchasing Must Become Supply Management,” Harvard Business Review, Sept–Oct 1983; Wikipedia — “Kraljic matrix.”
  3. Art of Procurement — “Learn the Kraljic Matrix”; “The 7-Step Strategic Sourcing Process.”
  4. McKinsey — “Revenue Growth Management: The Next Horizon”; Bain — Revenue Growth Management practice.
  5. NC State SCRC — “Introduction to CPFR”; Toby Desforges — “Trade vs Shopper Marketing.”
  6. ScienceDirect & eMarketer — retail media network definitions and market sizing; McKinsey — “Turning private labels into powerhouse brands.”

Industry verticals (Chapter 6)

  1. IESE — “Hard discounters’ secrets” and “The fast-fashion business model (Zara)”; Bain — “Battling grocery’s hard discounters.”
  2. Progressive Grocer — “How to be a category captain in today’s CPG world”; Crisp — “How category captains help brands and retailers.”
  3. Bloomberg & Supermarket News — Walgreens / CVS pharmacy strategy; DataPulse — “dm vs. Rossmann.”
  4. Retail Customer Experience & Construction Marketing Association — Home Depot / Lowe’s line reviews.
  5. Salsify — “1P vs 3P Amazon”; Profitero — “Winning on the digital shelf”; Wikipedia — “Best Buy,” “MediaMarkt,” “Sysco,” “AutoZone.”
  6. McKinsey — “Fuel retail in the age of new mobility”; CStore Dive — 7-Eleven new-format stores.
  7. FTC — “A Second Look at Category Management” (2004); American Antitrust Institute — Gundlach & Loff, “Competitive Exclusion in Category Captain Arrangements”; UK Groceries Code Adjudicator — Tesco investigation report, 2016.

AI, agents, simulation & governance (Chapters 8–14)

  1. McKinsey — “Merchants unleashed: How agentic AI transforms retail merchandising”; “From dashboards to decisions”; “The future of category management: AI category agents”; Merchant AI Accelerator.
  2. Gartner — “Over 40% of agentic AI projects will be cancelled by 2027” (press release, June 2025); “40% of enterprise apps will feature task-specific AI agents by 2026”; Hype Cycle for Retail Technologies 2025; AI Maturity Model toolkit.
  3. BCG — “How AI agents are transforming consumer goods”; Bain — “Agentic AI poised to disrupt retail”; Kantar — “The role of AI in revolutionizing category management.”
  4. RELEX — IDC MarketScape: Worldwide Retail AI-driven Assortment Planning 2025; agentic capability announcements. Blue Yonder, o9 Solutions, SymphonyAI, Aera Technology, ToolsGroup, Crisp — product documentation and announcements.
  5. NielsenIQ — “The rise of synthetic respondents”; Circana — Liquid AI; Harvard Business Review — “The AI tools transforming market research” (Nov 2025).
  6. Databricks & Target engineering — on-shelf-availability / out-of-stock modelling; InContext Solutions — digital twins and virtual store testing.
  7. arXiv — “Dynamic Retail Pricing via Q-Learning”; “RetailSynth” synthetic transaction data; surveys of LLM-agent hallucination.
  8. National Law Review — “AI antitrust landscape 2025”; FTC — materials on AI, algorithmic pricing and collusion; US DOJ — algorithmic price-fixing enforcement and the RealPage case.
  9. ScienceDirect — “A framework for human–AI collaborative decision making in intelligent retail environments” (2025).
On the figures in this guide

Quantified claims — forecast-error reductions, sales lifts, time savings, market sizes — are overwhelmingly vendor- or consultant-sourced, and consultancies frequently cite results from their own products. They appear in this guide as attributed claims, not established fact. The McKinsey 71% / 61% survey figures and the Gartner agentic-project and “agent washing” predictions are cited as those organisations’ published findings. Origin dates for category management, and the “9% use the full 8-step process” statistic, are presented with their inherent uncertainty noted in the relevant chapters.

End of Appendix B. — This concludes the field guide. Use your browser’s Save as PDF (or the button in the header) to produce a print edition.