
Why Mandating AI Tools Backfires: Lessons from Amazon and Spotify
- 4 minutes - Feb 26, 2026
- #ai#leadership#adoption#culture#amazon
Two stories dominated the AI-and-work conversation in early 2026. Amazon told its engineers that 80% had to use AI for coding at least weekly—and that the approved tool was Kiro, Amazon’s in-house assistant, with “no plan to support additional third-party AI development tools.” Around the same time, Spotify’s CEO said the company’s best engineers hadn’t written code by hand since December; they generate code with AI and supervise it. Both were framed as the future. Both also illustrate why mandating AI tools is a bad way to get real performance benefits, especially for teams that are already skeptical or struggling to see gains.
What Happened at Amazon
Amazon’s memo was clear: hit an 80% weekly-usage target and do it with Kiro. External tools like Claude Code were not on the approved list. The pushback was swift. Roughly 1,500 employees publicly backed Claude Code in internal channels, arguing it performed better for many tasks. AWS sales engineers pointed out the awkwardness of being barred from using Claude internally while selling it to customers via Bedrock. So you had a top-down mandate, a single-tool requirement, and a large group of experienced engineers saying “this isn’t the right tool for the job.”
The lesson isn’t that Kiro is bad. It’s that mandating one tool and one usage level ignores task fit and preference. When you do that, you get compliance metrics (e.g. “70% have used Kiro once”) instead of genuine productivity gains, and you burn trust. For teams that already “don’t see benefits,” a mandate reinforces the idea that leadership cares more about adoption numbers than outcomes.
What Spotify’s Story Actually Shows
Spotify’s CEO described senior engineers who no longer write code—they generate and supervise. That can be a valid evolution of the role. But it’s not a universal template. Some engineers report “AI fatigue”: reviewing and fixing AI output takes more effort than writing the code themselves. So the same headline—“best engineers don’t write code”—can mean “we’ve found a better workflow” for some and “we’re now doing more review work for the same result” for others.
Treating the Spotify model as a mandate (“you should stop writing code too”) would backfire for teams where AI doesn’t yet deliver net benefit. The takeaway is that outcome and experience vary by person and task. Leadership that ignores that and pushes one-size-fits-all adoption will struggle to get real performance benefits and will frustrate the people who don’t fit the mold.
Why Mandates Undermine the Goal
When the goal is “see performance benefits from AI in our workflows,” mandates tend to:
- Optimize for the wrong thing. You get usage (and maybe 80% weekly use) instead of better cycle time, quality, or satisfaction. Teams that are already skeptical will treat the mandate as a box to check, not as a reason to change how they work.
- Remove psychological safety. If people are required to use a tool they don’t trust or that slows them down, they can’t honestly report “this isn’t helping.” So you lose the feedback you need to fix task fit, verification cost, or tool choice.
- Lock in a single tool. Amazon’s “no third-party” approach forces everyone through one pipeline. When that tool is a poor fit for certain tasks (as many engineers argued), you’re forcing inefficiency and calling it adoption.
- Ignore the METR-style result. If experienced developers are often slower with AI on complex tasks, mandating “use AI more” can literally make your best people slower. Mandates don’t fix that; they hide it behind usage metrics.
What to Do Instead
- Set outcome goals, not usage goals. “We want cycle time down and quality up” is compatible with “use AI where it helps and skip it where it doesn’t.” “80% of you must use AI weekly” is not.
- Allow tool choice where it matters. Different tasks and people benefit from different tools. Standardize where it reduces risk (e.g. security, data), but don’t force one AI assistant for everything if evidence and preference point elsewhere.
- Protect the ability to say “this doesn’t help.” Make it safe to report that AI slowed someone down or that they prefer not to use it for certain work. Use that feedback to improve workflows and tooling instead of punishing it.
- Pilot and measure. Run time-bound experiments with clear outcome metrics. “This quarter we’re trying AI-heavy workflows for docs and tests; here’s the before/after.” Let the data drive scaling or rollback, not the mandate.
Amazon and Spotify will keep making headlines. The lesson for the rest of us: mandating AI tools is a fast way to get adoption numbers and a slow way to get real performance benefits. For teams struggling to see benefits, the fix is better task fit, better measurement, and more choice—not more pressure to use a single tool, more often.


