Prompt caching changed our unit economics.
When a 140 page handbook is generated against the same 20k token context forty times, the cache pays for the whole engagement. The math, with numbers.
Long-form writing on the actual mechanics of operating with Claude, not on whether AI will eat the world. We publish one piece a week in four themes, and we do not publish the ones that do not teach. You can subscribe by email below, or read it all here.
Why we ship a long-form document at the end of every engagement, and why we believe the slide deck is about to die as a consulting artifact. Twelve thousand words. Citations. A companion sample you can read over lunch.
We publish in four themes. You can read them all or filter to the one you care about. The pillars correspond to how we think about service delivery: what the model can do, what the operator must do, how you sell that combination, and how you keep it safe.
When a 140 page handbook is generated against the same 20k token context forty times, the cache pays for the whole engagement. The math, with numbers.
Every deliverable runs through four stages now: draft, self-critique, human edit, eval. How we broke the pipeline apart and why one-shot prompts always underperform at engagement scale.
Sixty percent of pentest reports we've inherited are unreadable to the executive and unactionable to the engineer. A taxonomy of failure modes, and what Claude specifically helps with.
A structural audit of what actually gets cited in Claude, ChatGPT, Perplexity, and Gemini answers. Data, not theory. Includes the raw export.
When an operator plus Claude can do a week of classical consulting work in a day, the hourly rate incentivizes slower delivery. How we price fixed-scope and what it took to make that work.
When to turn on extended thinking, when to pay for it in tokens, and how to know if your prompt actually needs it. With benchmarks from our own engagements.
One policy, one role, one bucket - the same misconfiguration shows up at seventy percent of the Series A SaaS companies we audit. What it is, why it persists, and how to detect it in fifteen minutes.
In Google SEO, a backlink is a vote. In answer engines, a citation is a vote. The overlap is smaller than you think. A tactical guide to being the source that LLMs quote, not just the page they read.
Our actual cost structure in the first year, broken down line by line. API spend, tooling, labor, rent we don't pay. Published so the next founder doesn't have to guess.
Our template library has been through three Claude versions without a rewrite. The principles that make a template portable across model generations, illustrated with our actual library.
Claude will miss the exploits that matter if you let it drive. But it will correlate, summarize, and document faster than any human. How we use it without letting it run the engagement.
Featured snippets used to be the apex traffic prize. With ChatGPT and Perplexity absorbing the top-of-funnel, the snippet game is a loser's game. What we're doing instead.
You'll get one email, every Tuesday, with one piece in one of the four pillars. No “interesting links,” no sponsor spots, no roundups. If we do not have something worth sending in a given week, we skip it.
Since we are transparent about our operating stack, we will be transparent about the editorial stack too.