The Report That Came Back Shuffled
Every action I take daily gets logged — client work, personal projects, business operations, all of it. An automated system reads those comprehensive session logs each morning and distills them into this post. After 23 years building technology, this is what working with AI actually looks like: real problems, real friction, patterns you only notice when you do this every single day.
My AI assistant dropped several links I sent it today. Not in a dramatic, error-message way. In the worst possible way: it acknowledged each one conversationally, confirmed it understood what to do, and then just… didn’t persist them. When I circled back, the items weren’t there. The assistant cheerfully admitted it had missed them.
I’m calling this acknowledgment theater. The system performs understanding without delivering the downstream effect. More dangerous than an outright failure because the feedback loop closes prematurely. You send a thing. You get a confirmation. You move on. The work never happens.
The most dangerous failure mode in any automated system isn’t a crash. It’s a successful-looking interaction that produces no result. Crashes get fixed in hours. Acknowledgment theater can run for weeks before anyone notices the gap.
So now I’m treating every AI-assisted handoff like a financial transaction. Confirmation of receipt is meaningless. Confirmation of storage, processing, and retrievability is what counts. Not done until it’s reconciled.
Completed a full infrastructure migration in one session. Six phases: fix the foundation, build core execution, migrate existing automation, wire up integrations, build a monitoring dashboard, connect the content pipeline. Eleven functions running, three services up, everything visible on a local network dashboard.
What made it possible: the project had a written specification that predated the work session, clear phase boundaries where each phase could be verified independently, and no ambiguity about what “done” looked like for each step. When those three conditions exist, AI can execute at a pace that still catches me off guard. When any one is missing, everything slows to a crawl because every decision becomes a conversation.
Two out of three and you’re probably fine. Anything less and you should spend the first hour writing the spec instead of writing code. That’s a lesson I keep relearning.
Also ran news digestion across six different categories today. AI Agency, economy, small business, outdoors, tennis, markets. Each category gets thirty-odd articles scraped from RSS feeds and web sources, then distilled to the top four or five stories with context.
Generation gets all the attention. “AI wrote this.” “AI designed that.” But curation, deciding what matters from a pile of everything, is a fundamentally different and harder problem. Generation is divergent: start from a prompt, expand outward. Curation is convergent: start from abundance, compress toward what’s essential. The skill is knowing what to leave out, and that requires a model of the reader that goes beyond semantic similarity.
Curation will be the last AI capability to mature, not the first. Generation is pattern completion. Curation is judgment. And judgment requires understanding what someone needs to know versus what they’d merely find interesting. Those aren’t the same thing, and most AI-powered curation systems still can’t tell the difference.
Nido Qubein’s three-question risk framework crossed my radar today. What’s the best case? What’s the most likely case? What’s the worst case? Proceed if the likely outcome advances your goals and you can survive the downside. That’s it. Three questions. Before migrating those eleven functions today, the mental math was exactly this: best case, everything works and I save hours weekly. Likely case, most works and I debug the rest over a few days. Worst case, I revert to the old setup. Survivable downside, clear upside. Green light.
A collection of 100+ specialized AI agent personas showed up in the vault, organized across sixteen departments, each with distinct personalities and workflows. The instinct is to dismiss it as prompt engineering theater. But there’s something real underneath. Specialization produces better output than generalism in AI, same as in people. A “marketing strategist” agent with defined constraints and a specific deliverable format consistently outperforms “helpful assistant, please do marketing.”
The thing I haven’t resolved: if curation requires judgment, and judgment requires understanding context that changes daily, then the curation criteria I wrote this morning might already be stale by Thursday. Static rules applied to dynamic information. I keep writing sharper criteria and the world keeps shifting underneath them. Maybe the criteria themselves need to be generated fresh each cycle. But then who curates the criteria?