Last updated: March 10, 2026
I have an AI Chief of Staff. Not a chatbot I talk to sometimes. An actual Chief of Staff that shows up every morning with my briefing prepared, my conflicts flagged, and my priorities loaded — before I say a word.
This is post 2 in my AI Executives series, where I break down how I built a 19-agent AI system that runs my consulting practice. If you haven't read the full architecture overview, start there. But this post stands on its own.
Here's the short version: most people's AI experience quietly dies because they hit the same invisible wall every session. I'm going to show you how I broke through it.
What Is an AI Chief of Staff?
An AI Chief of Staff is a persistent AI system that proactively manages your context, priorities, and daily operations — not just responds to commands.
The distinction matters. An AI assistant waits to be told what to do. An AI Chief of Staff shows up with the briefing already prepared.
Think about what a human Chief of Staff does in a large organization. They don't just take notes and schedule meetings. They synthesize information across departments, flag conflicts before they become crises, and make sure the CEO's time aligns with what actually matters that quarter. They hold the context so the leader can hold the vision.
That's what I built — except it runs on Claude Code and a markdown file instead of a six-figure salary.
Here's what my AI Chief of Staff handles every morning:
| Function | What It Does | Why It Matters |
|---|---|---|
| Calendar scan | Reviews today's meetings, flags prep needs | No more scrambling 5 minutes before a call |
| Inbox triage | Surfaces what matters, filters noise | Email doesn't set my agenda anymore |
| Task prioritization | Shows tasks ranked by urgency and alignment | Not just "what's due" — "what matters" |
| Urgent flags | Catches things I missed or forgot | Safety net for dropped threads |
| Cross-agent handoffs | Routes work between my 19 AI agents | Coordination that used to live in my head |
The entire briefing takes five minutes. I open one window and start my day with clarity.
The Stranger Loop Problem: Why Your AI Forgets You Every Session
Here's the thing most people don't talk about with AI: every conversation starts from zero.
You open ChatGPT or Claude. You explain your role, your project, your constraints. You get a decent answer. You close the tab.
Next morning? Same thing. Re-explain who you are. Re-explain what you're working on. Re-explain what matters.
I started calling this the Stranger Loop — and it's where most people's AI experience quietly dies.
A 2025 Boston Consulting Group study found that while 85% of executives reported experimenting with generative AI, only about 6% had deployed it at scale. Microsoft's 2024 Work Trend Index reported that 75% of knowledge workers use AI at work, but most are still in the "copy-paste a prompt, hope for the best" phase. The gap between "tried AI" and "AI actually changed how I work" is enormous.
The Stranger Loop is a big reason why.
Nobody quits AI because the output was bad. They quit because the overhead of re-establishing context every session eventually costs more than the value they're getting. It's death by a thousand onboardings.
"The challenge isn't getting AI to produce good outputs. It's getting AI to produce contextually appropriate outputs consistently." — Ethan Mollick, Wharton professor and author of Co-Intelligence
For my brain specifically — I have AuDHD, which means ADHD and autism together — the Stranger Loop isn't just annoying. It's expensive. Every unplanned context switch costs real cognitive energy. My working memory is a whiteboard that someone erases every time I look away. Asking me to re-explain my entire business context every morning is like asking someone with a broken leg to climb the stairs before they can start working.
I needed a system that held the context my brain couldn't.
How Persistent Context Changes Everything
The fix is surprisingly simple in concept: give your AI a memory file it reads before every conversation.
In Claude Code, this is called a CLAUDE.md file. It's a markdown document that sits in your project directory and gets loaded automatically at the start of every session. No copy-pasting. No "here's my background" prompts. The AI just... knows.
Here's what my CLAUDE.md includes:
- Who I am — my role, my business, my personality type, my working style
- What I'm building toward — my 90-day sprint goals, tier priorities, specific metrics
- My values — not aspirational poster values, actual decision-making values with specific behavioral definitions
- My constraints — AuDHD working patterns, context-switching costs, known blind spots
- My agents — who does what, how they hand off work, what each one can and can't access
- My patterns to watch for — specifically the overextension pattern where I take on too much
That last one is critical. I didn't just give my AI my resume. I gave it my failure modes.
The result: my Chief of Staff doesn't just know what's on my calendar. It knows why certain calendar items matter more than others given what I said matters this quarter.
This is the difference between "you have three meetings today" and "you have three meetings today, but the 2pm conflicts with your deep work block and none of them advance your Tier 1 goals — do you want to reschedule?"
Content is no longer king. Context is king.
The same AI model, with the same capabilities, produces dramatically different value depending on how much context you give it. A well-contextualized AI assistant is a different tool entirely from a cold-start one.
Building a Daily Briefing That Actually Knows Your Business
My daily briefing isn't a template I fill in every morning. It's a routine my Chief of Staff runs automatically when I say "startup."
Here's what happens in those five minutes:
Step 1: Date and orientation. Sounds basic, but it confirms the current date, checks what day of the week it is, and loads the session context. My Chief of Staff knows it's Tuesday, which means different priorities than Friday.
Step 2: Handoff check. My 19 agents write status reports and hand off work to each other through shared files. The Chief of Staff scans all of them and surfaces anything addressed to me — or anything that needs my decision before other agents can proceed.
Step 3: New material preview. What's changed since my last session? New YouTube transcripts in my knowledge base? New content drafted by my content agent? New intel from my marketing director? Quick counts, not full reviews.
Step 4: Last session context. One line on what I was working on last time. This is the anti-Stranger-Loop in action — I don't have to remember where I left off. The system remembers.
Step 5: Urgent flags. Only surfaces if something genuinely needs immediate attention. Broken posting cadence, missed deadlines, content gaps. If nothing's urgent, it says so and moves on.
The key design principle: the briefing isn't "here's everything." It's "here's what matters." The system filters thousands of data points down to the handful that deserve my attention right now.
Before this system, my mornings looked like this:
- Open my laptop and immediately get pulled into email. Whatever landed overnight set my agenda for the day — not my priorities, theirs.
- Check LinkedIn, iMessages, and Slack in no particular order, responding to whatever felt urgent in the moment.
- Try to remember what I was working on yesterday. Fail. Spend twenty minutes re-reading my own notes across three different apps.
- Realize at 11am that I haven't started on anything that actually moves the needle this quarter.
My mornings weren't unproductive. They were reactive. I was working hard from the moment I sat down — just not on the right things.
Now I start every day knowing exactly where I stand and what matters most. Five minutes of structured clarity beats an hour of reactive scrambling.
What My AI Chief of Staff Does in a Typical Day
Beyond the morning briefing, my Chief of Staff runs throughout the day as a persistent coordination layer.
Last month, I said "startup" and Lennier came back with: 53 emails (28 auto-archived, 25 needing me), 8 open follow-up loops with people I'd lost track of, a blocker where one of my agents couldn't function because a shared file had grown too large, and a routing recommendation — visit my content agent first because my marketing agent's work depended on it. In one five-minute briefing, I had drafted replies queued for 6 follow-ups, a structural fix in motion, and a clear execution order for the day. Before this system, that same morning would have been two hours of inbox archaeology and a growing sense that I was forgetting something important.
The pattern across all of this: I think, it coordinates. I make decisions, it routes them. I set priorities, it enforces them — even against me.
The doing isn't the work anymore. The thinking is the work.
How to Start Building Your Own AI Chief of Staff
You don't need 19 agents to get started. You don't even need Claude Code specifically. The principle works across tools.
Here's the minimum viable Chief of Staff:
-
Write your context document. Start with who you are, what you're working on, and what your priorities are this quarter. Even 500 words makes a dramatic difference versus starting cold.
-
Add your failure modes. What patterns do you fall into? What should the AI watch for? This is what turns an assistant into a guardrail.
-
Build a startup routine. Define what "check in" means — calendar, tasks, urgent items. Make it repeatable.
-
Add memory between sessions. Log what happened at the end of each session. Load it at the start of the next one. This is how you break the Stranger Loop.
-
Iterate for two weeks before adding complexity. Resist the urge to build everything at once. Get the Chief of Staff solid, then consider adding a second agent. One Person, Five AI Executives covers how the full system connects.
The technology will keep changing. The models will get better. But the principle — give AI persistent context about who you are and what matters to you — is the foundational layer everything else builds on.
Information expires. Systems compound.
Frequently Asked Questions
Do I need to be technical to build an AI Chief of Staff?
No. I'm not a developer. My CLAUDE.md file is written in plain English with markdown formatting. If you can write a detailed email, you can write a context document. The thinking is harder than the technology — you need to actually define your priorities, your constraints, and your patterns clearly enough for the AI to act on them.
What tools do I use for this?
I use Claude Code (Anthropic's CLI tool) with a CLAUDE.md file for persistent context. But the principle works with any AI tool that supports system prompts or custom instructions — ChatGPT's custom instructions, Claude Projects, or even a text file you paste at the start of each session. The tool matters less than the context architecture.
How long did it take to build?
The first version took about two hours — mostly writing the context document. Getting it to the Chief of Staff level (proactive briefings, cross-agent coordination, pattern detection) took about two weeks of daily iteration. Each session I'd notice something missing or miscalibrated and adjust. It's a living document, not a one-time setup.
Is this just a fancy prompt?
In the same way that an operating system is "just code," sure. But a CLAUDE.md file that includes your values, your 90-day goals, your failure modes, your agent architecture, and your session history is qualitatively different from a prompt that says "you are a helpful assistant." The depth and specificity of the context is what creates the Chief of Staff behavior. One Person, Five AI Executives goes deeper into why architecture matters more than any individual prompt.
Does this actually save time, or is it just interesting?
My morning orientation went from scattered and reactive to a 5-minute structured briefing. The Chief of Staff saves real time in context-switching costs alone — and that's before counting the decisions it helps me avoid (like the overextension pattern catches). The ROI isn't theoretical. It's my actual workday. Your AI CMO covers how the same persistent context principle applies to content creation.
This is post 2 in the AI Executives series. Next up: how I built an AI CMO with a mentor council that knows my voice better than I do.
Building your own AI executive team? Connected Intelligence teaches the full architecture — from your first context document to a coordinated multi-agent system.