StrategyFebruary 4, 2026

Be a deliberate follower... but sometimes following means losing

Caution isn't the problem. Inaction without testing is.

When it's OK to be a follower (and when it's not)

Following can be smart if you are buying time - and you know exactly what you're doing with that time. But if you simply hope to "do it later", you lose a compounding advantage.

  • OK to follow: you have a clear risk (regulation, data, security) and you run controlled tests.
  • Not OK to follow: a competitor can respond faster, serve cheaper, or produce more offers per day using AI.
  • Most dangerous: sales + customer support. Speed is directly tied to revenue.

Rule: if AI touches the customer expectation directly (speed, personalization, 24/7), following is risky.

A "minimum experimentation" framework (2 weeks / 3 use cases)

You don't need a 6-month program. You need a 2-week sprint with real work and a real metric.

  • Pick 3 use cases: (1) customer support, (2) sales offers, (3) internal automation.
  • Keep each use case narrow: one input -> one decision -> one output.
  • Use a "human checks" rule: AI drafts, a human approves.
  • Log everything: question, answer, edits, time saved, errors.

After two weeks you either have a number or you don't. If you don't, you don't scale - you change the use case.

KPIs: what proves AI is creating value

The most honest KPI isn't "did we use AI" - it's "did the process change".

  • Time: response/offer cycle time (e.g., 24h -> 2h).
  • Volume: how many cases/offers the team can handle per day.
  • Quality: fewer mistakes, fewer back-and-forth emails.
  • Money: conversion, average deal size, retention, service cost per case.

If KPIs don't move, AI is just an extra layer. Redesign the workflow - don't just add "another tool".