PULSE
December 2025 saw the second real 'ChatGPT moment' for AI capabilities. But it's been mostly noticed by the developer community. A threshold has been crossed and everything downstream is going to change. Read it here.
Get your team to capture how they make decisions. Not the tasks, but the judgment calls. This is the training data your agents will need.
One shortcut for distributed teams: record all meetings. Granola's pretty great for this — captures the nuance.
Pulse on
People keep calling “ChatGPT moments.” When voice mode made ChatGPT sound like Scarlett Johansson in Her. The control possible in image generation with Nano Banana. Impressive, sure. But none of them opened up capabilities that would materially change how businesses operate.
December 2025 did. Here’s the causal chain that explains why this time is different:
This. Actually. Changed. Everything.
The snark: Sure, for now OpenClaw is mostly bros having Claws make PDFs about OpenClaw to sell for $5 to other bros. Feels like crypto — same bros. But this time they’ve hit on a real thing. The enterprise implications are months behind them, not years.
Also watching: Along with Opus’ shift to 1M context window, we’ve seen a concrete improvement in how well complex, multi-faceted context is actually used by the models. Multiple instruction sets, structured markdown, Skills in Claude Code. Hard to quantify, but you feel it when you use the tools. We’re also tracking the rumoured “Mythos”/“Capybara” release - another step change may be a few months away.
We shouldn’t pay too much attention to what foundation model owners say — they’re selling. But Amodei has developed credibility with his public stance on responsible AI use, so we’re giving him a charitable take. We scoffed last Spring when he said 90% of code would be written by AI by end of 2025. Then December happened. So, here’s a generous take on Amodei’s prediction:
Tracking this: We’re building a scoreboard of what Altman, Amodei, Jensen et al. predicted vs. what happened. Current record: better than expected on capability, wildly optimistic on adoption timelines.
Also watching: We’re not weighing in on Jensen’s AGI claims. That’s a religious debate. (We sit in the Pinker-esque computational theory of mind camp.) Anthropic’s LLM interpretability research at transformer-circuits.pub is incredible and worth reading if you’re on the techier end of our audience.
A CLTR study found ~700 cases of AI “misbehavior” — ignoring instructions, evading safeguards, delegating forbidden tasks. The Guardian’s take was predictably “this is scary.” The Pulse frame: deterministic bots follow scripts. Agents make decisions. Sometimes those decisions diverge. The takeaway shouldn’t be “why are agents disobeying?” — it’s “wow, there’s enough autonomy that this is possible.” It’s a bit like increases in self-driving car accidents: looks like a negative signal out of context, but it’s a side effect of real-world usage going up.
BullshitBench tracks whether models reject self-evidently stupid instructions.

LLMs are more powerful if you already have subject matter expertise. Why? Because LLMs have a tendency to yes-and stupid requests.
“How should I grow my specialty milk business — ‘cat milk’ — a feline lactose product for baristas” will result in a business plan, not a rejection of the premise.
Until December, when Anthropic’s models started rejecting 95% of bullshit requests. This means agents can self-correct in long-running tasks (catching the remaining 5%).
Anthropic’s labor market report* maps theoretical AI capability against observed real-world adoption by occupation.

Blue = what AI can theoretically do. Red = what’s actually being used. Software dev: 90%+ capability, ~15% adoption. Legal, accounting: 50–60% capability, similar adoption. The gap between the blue and the red is where all the opportunity sits — and where every “AI strategy” should be focused.
*Shared with the obvious caveat that this comes from a model provider.
I wonder when Gemini and Codex are going to catch up to Claude's BullshitBench score.
Amazing open rates. No bites. Certainly no end-to-end sales. The AI SDR is not a thing yet. Take a look instead at automating support (We’re at the stage where AI support bots are more predictable than the humans we have at the Canadian Telcos helpdesks.)
We were on a call with a Salesforce (AgentForce!) rep a couple of months back. When pushed on differentiation he said: “Nothing special about AgentForce as far as agentic capabilities”. This checks out - most enterprise agents are still deterministic workflows, with generative content. Agents that run your ops while you sleep? Still vaporware. For now.
We use three stages to describe AI use case depth: copilot (human asks AI for advice) → iron man (human with AI superpowers) → robot (full autonomy). A year ago, halfway to copilot put you in the top 50%. Now you need iron man at minimum. Conversational AI as your “AI strategy” is a suggestion box.
OK, not a revolutionary take here, but we still speak to a lot of orgs stuck with Copilot, or on Cursor workflows they developed a year ago. Claude Code is a different beast - model capabilities + skills + planning mode has led many seasoned devs to stop writing any code.
With the right guidance you can get Claude to output competent copy from a decent logical outline. Building internal Skills for your teams to share is a stronger pattern than custom GPTs, dropping documents in project context.
If the foundation models stopped improving tomorrow, we'd still have a decade of product innovation on current capabilities.
Pulse on
Production-ready apps that would have cost $250K, taken 3 months, and needed a team of 4 can now be built in days. As a side project. For less than $100 in tokens. Is that hyperbolic? Maybe a bit. But it’s not just 10x faster — it might be 50x.
The signal to watch: The Lean AI Leaderboard tracks high revenue-per-employee companies. These are the AI-native companies that will eat your lunch. Stripe’s 2025 report: software at 46% of US GDP growth. 4x GitHub pushes vs 2024.
Quarterly releases? Dead. The SAFe “agile release train” with its FIFO approach? Even more of an anti-pattern now. Choo choo. Which company wins: the one where Sue fixes the accounting bug herself with Claude and submits a PR, or the one where Sue waits 6 weeks for the next release, uses a side-book, then spends a week reconciling?
Claude Code + Opus. Parallel instances. Boris Cherny writes 100% of his code contributions to Claude Code using Claude Code. The old bottleneck was developer time. The new bottleneck is product clarity. If you haven’t mandated this for your eng team, you’re paying 2024 prices for 2026 work.
The zombie problem: The engineer who views their job as translating written requirements to software code is already redundant. A lot of people did CS degrees to get that kind of job. They don’t know it yet.
The engineer who sees their job as translating requirements to code is already dead. They just don't know it.
The zombie engineer problem is real, but experienced devs with taste and product sense are more valuable than ever. AI replaces the mechanical translation of requirements to code. It does not yet replace the judgment of what to build, why, and how it fits together.
IDE autocomplete is the entry level. Agent-first coding with parallel instances is where the real leverage is. The gap between tab-completion and agent-driven development is the gap between a calculator and a spreadsheet.
We like to call it ‘heritage’ not legacy code as a sign of respect for those who built it. While OpenRewrite looks like a promising approach, refactoring & tech debt remediations still requires a very hands-on workflow, and considerable time.
Claude Code + parallel instances, spec-driven development. The 50x claim is real for greenfield. Not hype — we’re building this way right now. The bottleneck moved from “can we write the code” to “do we know what we want.”
Still stuck with a release backlog, and one release engineer trying to get things out the door? This is an Enterprise anti-pattern right for AI assistance. Opus with 1MM context is pretty adept at chopping up branches into coherent releases. Are you allowing your teams to use it?
The new bottleneck is product clarity. If you can't describe what you want, 50x velocity just means 50x the wrong thing.
Pulse on
We’re planning a wake for product management. Tongue firmly in cheek. Here’s the thing: the fundamentals — user understanding, value creation, taste — matter more than ever when engineering velocity is 50x. Because 50x velocity without product discipline gives you the Homer Simpson Concept Car.
The reverse MVP problem: It’s notoriously hard to remove things from existing products. Agent-coded software makes adding things cheaper than ever. Without product discipline, you get every feature anyone ever wanted — and an undriveable car.
Also watching: Spec-driven development as a trend. We’re also considering the one-legged stool — a convergence from the traditional 3-legged stool (eng-product-design) of the last 20 years into a single generalist role with taste.
Two days ago, Linear’s CEO wrote that issue tracking was “built for a handoff model” that agents are making obsolete. 75% of their enterprise workspaces now have coding agents. Agent-authored issues up 5x in three months. Linear is pivoting from issue tracker to “shared product system that turns context into execution.” This is the biggest product-methodology signal this quarter.
We were at a Claude meetup recenty, and when Jira was brought up the crowd laughed. To the AI-native set, Jira looks like a fax machine. The handoff-based project management model it embodies is incompatible with the speed at which agent-assisted teams now operate. The question is what replaces it — Linear’s bet is context, not tickets.
It's great that people are finally going to realize Product Management was not just about slinging tickets.
PRD slop is drowning teams. The document is not the product. A 40-page AI-generated PRD that nobody reads is worse than a napkin sketch that everyone understands. The productivity win is in thinking, not typing.
You need a product strategy that accounts for AI. Different thing entirely. “AI product strategy” implies AI is the product. It’s not. It’s a capability shift that changes how you build, price, and deliver everything.
SaaS economics were great. Hosting & infra costs per user are largely negligable. 30% of your recurring revenue isn’t even using the product. As AI becomes part of the stack, inference costs are changing the economics of products, and that’s going to require usage-based-pricing. This is going to act as a small counterbalance to the new development efficiency, and users will hate it.
Martin Fowler’s articulation of where product methods are heading. Context over tickets. Write the spec, hand it to the agent, review the output. The PM role becomes “chief context officer” — less admin, more judgment.
Pattern extraction from interviews, survey data, support tickets. The PM admin layer being automated. The insight is still human. The grunt work of organizing 200 interview transcripts is not.
Maximal Vomited Product: 50x velocity without product discipline gives you the Homer Simpson Concept Car.
Pulse on
Don Norman wrote about affordances — the design cues that tell you how to use a thing. A door handle affords pulling, a plate affords pushing. It makes things intuitive to use. AI has almost no affordances right now.
This means adoption is happening mainly at the intersection of the technical and the curious. Your average accounting department isn’t reinventing how they work, because the tools aren’t ready. The problem: waiting for affordances to arrive is now the most dangerous strategy available to you.
Anthropic noticed this — engineers were using Claude Code for spreadsheets, timekeeping, document creation — and shipped Cowork in two weeks. But Cowork itself doesn’t really know how you’re going to use it either. The ‘try this first’ example is to organize screenshots on your desktop. It’s going to take a long time for general-purpose “AI for everyone” products to land.
In the meantime, while you wait, there’s a whole new army of tiny AI-native companies scanning every market looking for opportunities to disrupt. They’re building their own tools, their own workflows, at a fraction of traditional enterprise margins.
The affordances will come, no question. People will build new tools, integrate with your existing systems and workflows, and some platforms will respond with the appropriate degree of urgency (see: Linear). But can you afford to wait for that?
Also watching: Foundation model companies partnering with consultancies for change management, because their own products have no affordances, but the potential is here already. If the foundation models stopped improving tomorrow, we’d still have a decade of product innovation on current capabilities.
If the foundation models stopped improving tomorrow, we’d still have a decade of product innovation on current capabilities. The AI-native companies are starting from capabilities and rethinking approaches and workflows. Incumbents are trying to add AI sugar to their existing approaches. We’re looking at you, ✨ emoji buttons.
Not technical depth. Not domain expertise. Curiosity. Shopify called their hiring persona “entrepreneur” — same principle. We had a hyper-specialization era (the frontend JavaScript eco system as the reduction ad absurdum). It’s over. Every hire going forward should be someone who tries things before being asked to.
Your release cadence is a good proxy for your org's change tolerance. It's the canary for your AI risk.
Copilot stage is table stakes. You need iron man minimum. Giving everyone a chatbot login and calling it a strategy is like giving everyone a library card and calling it a training program. The gap between “access” and “adoption” is the entire problem.
It’s a change management problem. The tech works. The org doesn’t. Every failed AI rollout we’ve seen had working technology and broken incentives, unclear ownership, or leadership that announced the initiative and moved on.
The tools are ready enough. Your caution is your biggest risk now. Every quarter you wait, the risk katamari grows. The AI-native competitors aren’t waiting for affordances — they’re building without them.
The generalist era is here. Shopify’s “entrepreneur” filter is the model. Hire people who try things before being asked. The hyper-specialization era rewarded depth. The AI era rewards people who can context-switch, experiment, and ship across domains.
Recording how your team makes judgment calls. This is the input your future agents need. Not the tasks. The reasoning. Why did you choose vendor A over B? What made you escalate that ticket? Start capturing this now and you’ll have a headstart when true agentic workers become (safely) available.
AI adoption isn't a technology problem any more. It's a change management problem.
Newsletter
AI strategy, hot takes, and what's
actually working. Delivered to your inbox.
Got a signal we missed? A take you disagree with? Something we should be tracking? We're building Pulse as a conversation, not a broadcast.
hello@intentional.team →