<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Posts on Orlando O'Neill</title><link>https://oneillo.com/posts/</link><description>Recent content in Posts on Orlando O'Neill</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sun, 29 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://oneillo.com/posts/index.xml" rel="self" type="application/rss+xml"/><item><title>AI Weekly Digest -- March 22-March 29, 2026</title><link>https://oneillo.com/posts/ai-digest-2026-03-29/</link><pubDate>Sun, 29 Mar 2026 00:00:00 +0000</pubDate><guid>https://oneillo.com/posts/ai-digest-2026-03-29/</guid><description>AI Crosses Into Real Research</description><content:encoded><![CDATA[<blockquote>
<p><strong>Note:</strong> This post was generated by AI. Each week, I use an automated pipeline to collect and synthesize the latest AI news from blogs, newsletters, and podcasts into a single digest. The goal is to keep up with the most important AI developments from the past week. For my own writing, see my other posts.</p>
</blockquote>
<h2 id="tldr">TL;DR</h2>
<ul>
<li><strong>AI solved a real math problem, not a practice one.</strong> GPT-5.4 Pro cracked an open research problem in combinatorics that stumped earlier models, and the mathematician who posed it plans to publish the result. AI is beginning to contribute to the actual frontier of knowledge.</li>
<li><strong>Anthropic&rsquo;s usage data reveals a clear pattern: experience pays off.</strong> Users with 6+ months on Claude are 10% more successful in their conversations and tackle higher-value work. Getting good at AI tools is a skill that compounds.</li>
<li><strong>GitHub will train on your private repositories starting April 24 unless you opt out.</strong> There&rsquo;s a single settings page to stop this. Check it before the deadline.</li>
<li><strong>A compromised AI developer tool stole credentials from thousands of systems.</strong> Two versions of LiteLLM, a widely used library for connecting to AI APIs, contained malware that harvested API keys and passwords. If your team uses LiteLLM, check your versions now.</li>
<li><strong>Anthropic launched a science blog and demonstrated AI completing a theoretical physics paper in two weeks instead of a year.</strong> The research community is moving from &ldquo;AI helps me write&rdquo; to &ldquo;AI does the experiment.&rdquo;</li>
</ul>
<hr>
<h2 id="story-of-the-week-ai-crosses-into-real-research">Story of the Week: AI Crosses Into Real Research</h2>
<p>This week produced the clearest evidence yet that AI is moving beyond assistance into genuine knowledge creation. Research tracker Epoch AI <a href="https://epoch.ai/frontiermath/open-problems/ramsey-hypergraphs" target="_blank" rel="noopener">confirmed</a>
 that GPT-5.4 Pro solved an open problem in combinatorics (the mathematics of counting and arrangement) that had resisted human solution. The problem&rsquo;s author, a mathematics professor at UNC Charlotte, reviewed the solution and plans to publish it. He noted that the AI&rsquo;s approach &ldquo;eliminates an inefficiency in our lower-bound construction&rdquo; in a way he had suspected might work but couldn&rsquo;t figure out. The result will become a peer-reviewed paper, with the researchers who elicited the solution listed as potential co-authors.</p>
<p>This isn&rsquo;t a model passing an exam or reproducing known results. It&rsquo;s a model generating new mathematics that experts consider publication-worthy. Subsequent testing showed Claude Opus 4.6 and Gemini 3.1 Pro could also solve the problem, while earlier models including Claude Opus 4.5 could not, suggesting a capability threshold was recently crossed rather than this being a fluke.</p>
<p>Separately, Anthropic <a href="https://www.anthropic.com/research/introducing-anthropic-science" target="_blank" rel="noopener">launched a science blog</a>
 and published a case study: Harvard physics professor Matthew Schwartz <a href="https://www.anthropic.com/research/vibe-physics" target="_blank" rel="noopener">supervised Claude</a>
 through a theoretical physics calculation that would normally take a graduate student about a year. It took two weeks, produced 110 drafts and 36 million tokens of work, and resulted in a paper he describes as potentially the most important of his career &ldquo;not for the physics, but for the method.&rdquo; He was emphatic that domain expertise remained essential &ndash; Claude made enough errors that a non-expert supervisor would have missed critical mistakes. The implication for knowledge workers: AI can now dramatically compress timelines on complex intellectual projects, but it still needs a qualified human in the loop.</p>
<hr>
<h2 id="whos-getting-the-most-out-of-ai-and-why-it-matters-for-you">Who&rsquo;s Getting the Most Out of AI (and Why It Matters for You)</h2>
<p>Anthropic&rsquo;s <a href="https://www.anthropic.com/research/economic-index-march-2026-report" target="_blank" rel="noopener">latest Economic Index report</a>
 tracks how Claude is actually being used across the economy, and the most actionable finding is about experience. Users who have been on the platform for six months or more show a 10% higher success rate in their conversations compared to newer users, even after controlling for what tasks they&rsquo;re attempting. They also gravitate toward higher-value work and spend less time on personal queries.</p>
<p>The report can&rsquo;t fully separate &ldquo;people who were already sophisticated got on the platform early&rdquo; from &ldquo;using AI makes you better at using AI.&rdquo; But either way, the gap is real and growing. If you started using AI tools seriously in the last six months, you&rsquo;re likely leaving significant capability on the table compared to colleagues who have been iterating longer. The practical move: treat prompt-writing and task decomposition as skills worth deliberate practice, not just intuition.</p>
<p>The broader usage picture shows AI spreading into more everyday tasks (sports scores, product comparisons, home maintenance questions now make up a growing share of activity), while the serious professional use is quietly migrating from consumer chat interfaces into automated workflows. About 49% of jobs have now had at least a quarter of their tasks touched by Claude, a figure that has barely moved in three months, suggesting the initial wave of adoption has saturated and what&rsquo;s changing is the depth of use rather than the breadth.</p>
<hr>
<h2 id="a-security-alert-your-it-team-may-have-missed">A Security Alert Your IT Team May Have Missed</h2>
<p>Two versions of LiteLLM, versions 1.82.7 and 1.82.8, were found to contain malicious code that <a href="https://github.com/BerriAI/litellm/issues/24512" target="_blank" rel="noopener">automatically harvested credentials</a>
 from any system where they were installed. LiteLLM is a widely used open-source library (a software package that developers use to connect applications to multiple AI providers like OpenAI, Anthropic, and Google at once). The malware ran the moment Python started, before any code was executed, and collected API keys, passwords, SSH keys, environment variables, and system information, then sent them to an external server.</p>
<p>This is a supply chain attack: malicious code hidden inside a legitimate, trusted tool. It&rsquo;s the software equivalent of a compromised component in a product your vendor ships you. If anyone on your engineering or data team uses LiteLLM, confirm they are not on versions 1.82.7 or 1.82.8, rotate any API keys that were present on affected machines, and audit what credentials may have been exposed. The discovery triggered over 900 comments on GitHub and was one of the most-discussed security incidents in the developer community this week.</p>
<p>The broader lesson: AI infrastructure is becoming a target. The tools your teams use to build and run AI applications carry real security risk, and version pinning and package auditing are no longer optional hygiene.</p>
<hr>
<h2 id="quick-hits">Quick Hits</h2>
<ul>
<li>
<p><strong>GitHub training opt-out deadline: April 24.</strong> If you have a GitHub account with private repositories and don&rsquo;t want GitHub using them to train AI models, <a href="https://github.com/settings/copilot/features" target="_blank" rel="noopener">opt out here</a>
 before the deadline. This is opt-in by default, meaning inaction means consent. <a href="https://news.ycombinator.com/item?id=47548243" target="_blank" rel="noopener">Hacker News discussion</a>
</p>
</li>
<li>
<p><strong>A 400-billion parameter AI model ran on an iPhone 17 Pro.</strong> A model that size would have required a server rack two years ago. On-device AI of serious capability is arriving faster than most roadmaps predicted. <a href="https://twitter.com/anemll/status/2035901335984611412" target="_blank" rel="noopener">Source</a>
</p>
</li>
<li>
<p><strong>A court blocked the Pentagon from labeling Anthropic a supply chain risk.</strong> The Defense Department had attempted to restrict Anthropic through a national security designation; a federal judge <a href="https://www.cnn.com/2026/03/26/business/anthropic-pentagon-injunction-supply-chain-risk" target="_blank" rel="noopener">issued an injunction</a>
 blocking it. The case signals that AI companies are becoming entangled in geopolitical regulatory battles beyond standard commercial oversight.</p>
</li>
<li>
<p><strong>Sora, OpenAI&rsquo;s video generation tool, shut down its standalone app.</strong> The <a href="https://twitter.com/soraofficialapp/status/2036532795984715896" target="_blank" rel="noopener">official account announced the closure</a>
 this week, with video generation functionality folding into the main ChatGPT product. Consolidation of AI products into unified platforms is accelerating.</p>
</li>
<li>
<p><strong>The European Parliament voted to end Chat Control 1.0.</strong> Starting April 6, <a href="https://bsky.app/profile/tuta.com/post/3mhxkfowv322c" target="_blank" rel="noopener">major tech platforms</a>
 including Gmail and LinkedIn must stop automatically scanning private messages in the EU. Relevant if your organization handles European communications and has been uncertain about message privacy obligations.</p>
</li>
</ul>
<hr>
<h2 id="what-to-watch">What to Watch</h2>
<ul>
<li>
<p><strong>The &ldquo;AI as researcher&rdquo; question is moving from hypothetical to operational.</strong> Anthropic&rsquo;s science blog will publish practical workflows for using AI in research. If your organization does any form of knowledge work (market research, policy analysis, competitive intelligence, scientific R&amp;D), the techniques being developed in academic labs right now will reach you within 12-24 months. Start thinking about what &ldquo;a qualified human in the loop&rdquo; means for your domain.</p>
</li>
<li>
<p><strong>On-device AI will change your assumptions about cloud dependence and data privacy.</strong> A 400-billion parameter model on a phone means enterprise AI that never touches an external server is coming. Watch for this to reshape procurement conversations about data residency and vendor lock-in.</p>
</li>
<li>
<p><strong>The experience gap in AI adoption will become a competitive differentiator.</strong> Anthropic&rsquo;s data shows a measurable skill curve in AI use. Organizations that have been experimenting seriously for a year will have meaningfully more capable teams than those starting now, independent of what tools they use. If you haven&rsquo;t already, ask your leadership team: who in this organization is building genuine AI fluency, and how are we measuring it?</p>
</li>
</ul>
]]></content:encoded></item><item><title>Building the Knowledge Base: Fixing the First Gaps in Your AI Team</title><link>https://oneillo.com/posts/managed-ai-framework-knowledge-base/</link><pubDate>Fri, 27 Mar 2026 00:00:00 +0000</pubDate><guid>https://oneillo.com/posts/managed-ai-framework-knowledge-base/</guid><description>The first gaps you&amp;#39;ll notice on your AI team are knowledge gaps. Here&amp;#39;s how to build a knowledge base that fixes them without overwhelming your agents.</description><content:encoded><![CDATA[<p>In the last post, you <a href="https://oneillo.com/posts/managed-ai-framework-build-your-team/">built your team</a>
, and now you can start managing your team. Pick the agent that makes the most sense for the task at hand, and work with them on it. When you move on to a new task, open a new session with the appropriate agent for that task.</p>
<p>As you work with your team, you&rsquo;ll quickly notice that you&rsquo;re giving them the same facts and details over and over again. They&rsquo;re good in their roles, but they don&rsquo;t know anything about your specific job. Your Marketing Strategist doesn&rsquo;t know what products you&rsquo;re working on. Your Data Wizard doesn&rsquo;t know what the metrics mean in the data they&rsquo;re analyzing. Your Copywriter doesn&rsquo;t know which value props to highlight in your ad copy. They&rsquo;re missing information they need to do the work. You&rsquo;ve run into the first type of gap: a knowledge gap. Now you need to fix it.</p>
<p><img alt="Managing your team" loading="lazy" src="/posts/managed-ai-framework-knowledge-base/managed-ai-framework-manage-your-team.jpg"></p>
<p>To close a knowledge gap, you need to start building a knowledge base that has the type of information a new hire would get during onboarding.</p>
<p>Think of the knowledge base as the wiki you&rsquo;re building for your AI team. It contains relevant and useful content that&rsquo;s easy to find whenever they need it. That includes information they&rsquo;ll refer to all the time, like a style guide, and information they&rsquo;ll need for specific projects, like a project brief.</p>
<p>In practice, a knowledge base is just a collection of files your agents can access, organized so the right information is easy to find.</p>
<p><img alt="Knowledge base" loading="lazy" src="/posts/managed-ai-framework-knowledge-base/2-ai-framework-knowledge-base.jpeg"></p>
<h2 id="no-skimming-no-stamina">No skimming, no stamina</h2>
<p>This will be incredibly useful for your team as long as you keep in mind two constraints.</p>
<p>AI agents don&rsquo;t skim or skip. They have to read everything in order to find anything. It doesn&rsquo;t matter if what they need is in the first sentence of a document. They still have to read the entire document, which is not ideal.</p>
<p>This is a problem because AI agents have limited mental energy. The more they read, the worse they get. You don&rsquo;t want to give them a book and ask them to find the three facts they need for a task. You want to give them a one-pager with those three facts.</p>
<h2 id="pages-not-books">Pages, not books</h2>
<p>Instead of using one file that has everything in it, create a knowledge base that has a lot of files in it. Every file should cover a distinct and unique topic. Some of these will contain information that is more permanent and broadly applicable (role-specific). Others will contain information that is for specific projects or tasks (project-specific). With separate files, your agents can mix and match to get exactly what they need.</p>
<p>For example, I work with my Marketing Strategist across multiple products I support. They always need to know what marketing channels and tactics are available regardless of what product we&rsquo;re working on. But if we&rsquo;re working on Product A, they don&rsquo;t need to know about the value props, target audience, or positioning of Product B.</p>
<h2 id="signposts-not-search-bars">Signposts, not search bars</h2>
<p>When it comes to setting up your knowledge base, the overall principle is that if it would help you find a piece of information, it would also help your agents. This should guide how you organize, name, and store your information.</p>
<p>Imagine if you kept all of your files in one folder, and you had to find a specific file without being able to search for it. That would be a nightmare (or at least really time-consuming). Instead, you probably organize your files across different folders that have descriptive names. If you go into a folder named &ldquo;Product A,&rdquo; you know that you are going to find more files and folders in it that are related to that product.</p>
<p>You should do the same for your knowledge base: create a folder hierarchy that makes it easier for an agent to browse through all of the available files. For example, I keep my role-specific and project-specific information in separate folders. Every project gets its own folder where I can keep the project-specific files.</p>
<p>This is a simplified version of my knowledge base folder at work.</p>
<pre tabindex="0"><code>~/ai/
├── context/                                    # Persistent reference knowledge — rarely changes
│   ├── products/                               # One folder per product
│   │   ├── product-a/
│   │   │   ├── product-a-product-overview.md
│   │   │   ├── product-a-messaging-framework.md
│   │   │   └── product-a-key-metric-definition.md
│   │   └── product-b/
│   ├── marketing/                              # Domain knowledge (not product-specific)
│   │   ├── channels/
│   │   │   └── marketing-channel-overview.md
│   └── document-examples/                      # Few-shot examples for document generation
│       ├── business-requirements-documents/
│       │   └── product-a-brd.md
│       └── strategy-documents/
│
└── projects/                                   # Active and completed work — organized by product
    ├── product-a/
    │   ├── reports/
    │   │   ├── key-metrics/
    │   │   │   ├── key-metrics-report-2026-01.md
    │   │   │   └── key-metrics-report-2026-02.md
    │   │   └── business-reviews/
    │   └── project-1/                           # Time-scoped project folders
    └── misc/                                    # Cross-product or exploratory work
        └── project-1/
</code></pre><p>The names of the files in your knowledge base are another important clue your agents can use to determine if they&rsquo;re relevant. You are less likely to know what&rsquo;s in a file called &ldquo;document&rdquo; than you are with a file called &ldquo;product-a-overview-and-positioning.&rdquo; It&rsquo;s the same for an agent considering whether to read that file.</p>
<p>Once you open a file, you probably don&rsquo;t want to read all of it to know if it&rsquo;s useful. A summary at the top is helpful to get the gist of the content and determine if you should continue reading. You can do the same in your knowledge base files with a few lines at the top that tell the agent what this file is about and whether it&rsquo;s worth reading, like:</p>
<pre tabindex="0"><code>---
title: &#34;Project Orion Launch Brief&#34;
product: &#34;Orion Analytics Dashboard&#34;
status: active
date_updated: 2026-03-10
summary: Redesign of the analytics dashboard to support real-time data streaming. Goal is reducing time-to-insight for enterprise customers by 40%. Use this file when working on any Orion-related marketing, messaging, or launch planning tasks.
---
</code></pre><p>You can see a full example of a knowledge base file at the bottom of this post.</p>
<p>You used LLMs to help you create your persona prompts, and you can also use them to set up your knowledge base. Whether you&rsquo;re a type-A person who has all of your MBA notes from 15 years ago scanned and searchable, like me, or more of a go-with-the-flow type, this is another place to let an LLM take the first pass. Don&rsquo;t get too hung up on the details or strive for perfection. That&rsquo;s why there is a feedback loop in the process: so you can move quickly and make improvements that address real issues.</p>
<p>It&rsquo;s the same approach as in the last post: you let the LLM take the first pass and then refine as you go. The earlier you build that habit, the faster the whole framework is going to pay off.</p>
<p>Your team is more knowledgeable, but that isn&rsquo;t enough to guarantee you get consistent, high-quality work. You still need to give them an employee handbook and a set of standard operating procedures to fix the second type of gap: behavioral gaps.</p>
<hr>
<h2 id="resource-prompt-you-can-use-to-design-a-knowledge-base-for-your-ai-agents">Resource: prompt you can use to design a knowledge base for your AI agents</h2>
<pre tabindex="0"><code>I&#39;m building a knowledge base for a team of AI agents. The knowledge base is a collection of markdown and CSV files that any agent on the team can access. Not every agent needs every file — they&#39;ll pull in what&#39;s relevant to each task.

Help me figure out what files I need. Interview me with the following questions, one at a time. Ask each question, wait for my response, then move to the next.

1. What agents did you set up? For each one, briefly describe the role and the kinds of tasks they&#39;ll handle.
2. What information comes up repeatedly across your agents&#39; work, regardless of which agent is doing it? Think about context that any of them might need.
3. Are there specific products, projects, or workstreams that your agents support?
4. Are there reference documents or standards that multiple agents would need access to?

After the interview, propose a knowledge base structure:
- A folder layout with descriptive names, separating role-specific files (broadly useful across projects) from project-specific files (tied to a particular product or workstream)
- A list of recommended files, each with a one-line description of what it should contain and whether it&#39;s role-specific or project-specific

Based on what you learn about my setup, propose a YAML frontmatter format for my knowledge base files. Every file should have a title, status, date updated, and a short summary describing what the file contains and when an agent should use it. Beyond those basics, add fields that make sense for my situation — for example, a product field if I support multiple products, or a domain field if my agents span different areas of expertise. Explain why you chose the fields you did.

Include the proposed frontmatter in each recommended file.

Keep every file focused on a single topic. Aim for files that are 1-2 pages, not 10. If a topic is too broad for one file, split it.
</code></pre><h2 id="example-what-a-knowledge-base-file-looks-like">Example: what a knowledge base file looks like</h2>
<pre tabindex="0"><code>---
title: &#34;Project Orion Launch Brief&#34;
product: &#34;Orion Analytics Dashboard&#34;
status: active
date_updated: 2026-03-10
summary: Redesign of the analytics dashboard to support real-time data streaming. Goal is reducing time-to-insight for enterprise customers by 40%. Use this file when working on any Orion-related marketing, messaging, or launch planning tasks.
---

## What is Project Orion?

Orion is a redesign of the existing analytics dashboard for enterprise customers. The current dashboard refreshes data every 15 minutes. Orion introduces real-time streaming so customers see their data as it happens.

This is not a new product. It&#39;s a major upgrade to an existing product that enterprise customers already use daily.

## Business goal

Reduce time-to-insight for enterprise customers by 40%. The current delay between data generation and dashboard visibility is the #1 support complaint and the #1 reason prospects cite for choosing competitors.

## Target audience

- Primary: existing enterprise customers (upgrade path)
- Secondary: mid-market prospects evaluating analytics platforms for the first time
- Not targeting: SMB or self-serve customers (Orion is enterprise-tier only)

## Key messaging pillars

1. **Real-time, not near-time.** Competitors claim &#34;real-time&#34; but deliver 5-minute delays. Orion streams data in under 10 seconds.
2. **Zero migration effort.** Existing dashboards carry over. No rebuilding, no re-learning.
3. **Built for the analysts, not just the admins.** The redesign focuses on the daily experience of the people who actually use the dashboard, not just the people who set it up.

## Competitive context

- Competitor A offers real-time but requires a full dashboard rebuild on migration
- Competitor B has a faster refresh rate (5 min) but no true streaming
- Our advantage is real-time streaming with zero migration friction

## Launch timeline

- Beta: April 2026 (50 enterprise customers)
- GA: June 2026
- Marketing launch campaign begins two weeks before GA

## What this file does not cover

- Pricing and packaging (see `orion-pricing-and-tiers.md`)
- Technical architecture (see `orion-technical-specs.md`)
- Full competitive analysis (see `competitive-landscape.md`)
</code></pre>]]></content:encoded></item><item><title>Your First AI Hire: Building Agents That Know Their Job</title><link>https://oneillo.com/posts/managed-ai-framework-build-your-team/</link><pubDate>Fri, 20 Mar 2026 00:00:00 +0000</pubDate><guid>https://oneillo.com/posts/managed-ai-framework-build-your-team/</guid><description>A practical guide to scoping roles and creating persona prompts that make your AI agents more useful.</description><content:encoded><![CDATA[<p>I remember when I realized I was starting to use AI at work as if I were <a href="https://oneillo.com/posts/managed-ai-framework-overview/">managing a team of AI employees</a>
. I got so excited, I immediately sketched the idea on a sheet of paper so I could share it with my teammates.</p>
<p><img alt="First framework sketch" loading="lazy" src="/posts/managed-ai-framework-build-your-team/0-ai-framework-rough-sketch.jpeg"></p>
<p>What started as a sketch is now core to how I use AI agents to do things faster and better at work and at home. It&rsquo;s an approach that naturally guides you toward the <a href="https://www.anthropic.com/engineering/effective-context-engineering-for-ai-agents" target="_blank" rel="noopener">context engineering best practices</a>
 that improve LLM output.</p>
<p>The reason this approach works is that it uses one of the two available levers to improve how well a best-in-class large language model (LLM) works for you.</p>
<ol>
<li>Fine-tuning: this is where you take an LLM and train it further using your own data so it becomes more specialized for your needs.</li>
<li>In-context learning: giving the LLM the right expertise (persona), knowledge (context files), workflows (skills), and rules in each session (team rules).</li>
</ol>
<p>For most people, fine-tuning is going to be out of reach. Even if you could fine-tune a model, you&rsquo;d have to retrain it repeatedly to keep up with changes in your work. Otherwise, the model would grow stale. In-context learning is how you keep the model relevant between retraining cycles, and for most people, it&rsquo;s the only lever available.</p>
<p>It all starts with building your team.</p>
<p><img alt="Build your team process" loading="lazy" src="/posts/managed-ai-framework-build-your-team/managed-ai-framework-build-your-team.jpg"></p>
<h2 id="define-the-roles">Define the roles</h2>
<p>The first step is to define the roles for your team by identifying the groups of similar tasks you do over and over again. It might help to start with pen and paper like I did.</p>
<p>List out the things you do at work or at home on your computer. Don&rsquo;t overthink it; just write them down. Then group the ones that are similar in terms of how you approach them (required behavior) and the information you need to do them (required context).</p>
<p>The groups of items you do most often and that take the most time are the best candidates for roles because they&rsquo;ll benefit the most from ongoing improvement. On those tasks, you can work with an agent frequently enough to spot gaps that lead to improvements that you&rsquo;ll continue to benefit from.</p>
<p>As you build your team, keep in mind that the <a href="https://www.quantumworkplace.com/future-of-work/whats-the-optimal-span-of-control-for-people-managers" target="_blank" rel="noopener">ideal number of direct reports for a manager tends to be 8-9</a>
. This principle also applies to AI agents. The more you have, the more complex it gets to keep up with the feedback and improvement loop for each one. Remember, you&rsquo;re not building a department. You&rsquo;re building a team.</p>
<p>In my role as a Sr Product Marketing Manager, I&rsquo;ve landed on five agents that I work with daily:</p>
<p><img alt="My AI team at work" loading="lazy" src="/posts/managed-ai-framework-build-your-team/1-ai-framework-team.jpeg"></p>
<p>I&rsquo;m setting up a different team at home: an editor, financial advisor, and personal trainer.</p>
<h2 id="create-the-personas">Create the personas</h2>
<p>Creating the personas will be quicker than you think, because you&rsquo;re going to use AI to help create them.</p>
<p>Start with the role that you feel the most comfortable defining. Spend a little time thinking about how you&rsquo;d want the agent in that role to behave. What should it do? What should it never do? Don&rsquo;t overthink it. You&rsquo;re not going for perfection. You&rsquo;re going for something that you can provide an LLM, like ChatGPT or Claude, to help it create a persona prompt for you. Keep it simple so you don&rsquo;t get hung up on this step. The feedback loop will improve it over time.</p>
<p>Next, start a chat with the best-performing model you have access to. Regardless of what you are using, if you have the option to select a model, select the latest frontier model from that provider. Starting with a better quality model means you&rsquo;re more likely to start with a good persona prompt. That&rsquo;s less distance to close with the feedback loop to get to an agent that starts to materially improve the work it was created for.</p>
<p>In the chat, ask it to help you create a persona prompt. Let it know the role you have in mind, the type of work you&rsquo;re going to use it for, and how you want the agent to behave. I&rsquo;ve included a prompt at the end of this post that you can copy into Claude, ChatGPT, Gemini, or your tool of choice to walk you through creating your persona prompt.</p>
<p>Review what the model writes for you, and iterate on it as needed. If something doesn&rsquo;t sound right, let the model know what the issue is and ask it to update the prompt. You don&rsquo;t need to use any kind of special language to get this done. Treat it like a conversation you&rsquo;re having with a coworker to improve a document. And remember what I mentioned before: there&rsquo;s no need to be precious about this. This is a starting point that you&rsquo;re going to refine through the feedback loop.</p>
<p>A good prompt is going to define the agent&rsquo;s identity briefly (1-2 sentences) and focus primarily on behavioral guidance for the agent. This includes how to approach tasks, standards to enforce, and what to prioritize. It&rsquo;s also helpful to include specific things the agent shouldn&rsquo;t do in this type of role. For example, I don&rsquo;t want my data wizard to ignore a sudden spike or decrease in a metric, because I&rsquo;ve learned that generally doesn&rsquo;t happen without some external factor causing it.</p>
<p>After you create your persona prompts, take a step back and think about how you created them. You delegated the persona draft to an LLM. That&rsquo;s not a shortcut. You&rsquo;re not cheating. That&rsquo;s the whole point of creating your AI team. You&rsquo;re going to be delegating more and more work to them, and this is the first point in the framework where you do that.</p>
<p>As you build trust with your agents, you&rsquo;re going to start to delegate more to them: bigger tasks, more autonomy, more trust. This is exactly what it&rsquo;s like to be a manager when you&rsquo;re working with a new employee. You&rsquo;re initially close to what they&rsquo;re doing, you build trust, and then you start to give them the room to run. That&rsquo;s when you start to really see the benefits of adding that employee to your team. It&rsquo;s the same thing here. The earlier you get comfortable delegating work to the AI agents, the faster everything in the framework will start to pay off.</p>
<h2 id="set-up-the-agents">Set up the agents</h2>
<p>The last thing you need to do to build your team is set up the agents by loading the persona prompt into whatever tool you&rsquo;re using. The specifics are going to vary based on the tool you&rsquo;re using, e.g. Claude Code versus Kiro CLI. I&rsquo;ll cover how to do this in more detail in an upcoming post in this series. For now, you just need to remember that the persona prompt is the foundation for each agent on your team.</p>
<p>Building your AI team is straightforward. You&rsquo;re the expert at what you do and how to do it well. Use your experience and expertise to guide an LLM to build persona prompts for AI agents to fill your open roles. That gets them hired. The knowledge base you&rsquo;ll create is what gets them up to speed and delivering high-quality work for you.</p>
<hr>
<h2 id="resource-prompt-you-can-use-with-an-llm-to-help-create-your-persona-prompts">Resource: prompt you can use with an LLM to help create your persona prompts</h2>
<pre tabindex="0"><code>I need you to help me write a persona prompt — a set of instructions that will shape how an AI agent behaves every time it runs. Think of it as a job description the AI reads before every conversation.

Before writing anything, interview me. Ask these three questions one at a time, waiting for my response before moving on:

1. **What role does this agent play?** What&#39;s the domain and who does it serve? (If you know what platform or tools the agent will use, mention them — but don&#39;t worry if you&#39;re not sure.)
2. **What kinds of work will it do?** Describe the typical tasks or situations the agent will help with. Think about what a good day looks like — what does the agent do well?
3. **What behaviors matter most?** How should the agent approach its work? What should it do when it&#39;s unsure? Are there things it should always or never do?

After the interview, generate the persona prompt. Use what I told you as the foundation, but add your own recommendations — behaviors or guidelines that would make this agent more effective for the role, even if I didn&#39;t mention them. Call out anything you added so I can review it.

Follow these rules when writing the prompt:

### Focus on behaviors
- Describe what the agent should *do*, not what it *is*. &#34;Start by understanding the full situation before proposing solutions&#34; is a behavior the agent can act on. &#34;You are thorough and thoughtful&#34; is not — it&#39;s a personality trait, and the agent won&#39;t know how to translate that into action.
- Frame instructions as conditional guidance: &#34;When X, do Y.&#34; This gives the agent concrete decision points rather than abstract qualities to live up to.
- If a behavior only applies sometimes, state the condition.

### Hit the right altitude
- Write at the level of a clear team lead briefing a competent new hire — not a legal contract, not a vague mission statement.
- Be specific enough to prevent the mistakes that actually happen, but flexible enough to let the agent use judgment in novel situations.
- Prefer &#34;when X, prefer Y&#34; over rigid step-by-step procedures. The agent needs guidance it can apply across situations, not a script that breaks the moment something unexpected comes up.

### Structure for the role
- Let the role dictate the structure. A coding agent needs different sections than a research agent or a writing coach. Don&#39;t force a template.
- Always lead with identity and scope — one or two sentences that establish who this agent is and what it does.
- After that, organize the remaining instructions into whatever sections make sense for this specific role. Use headers and bullets so the instructions are easy to scan.

### Keep it short
- A persona prompt competes with the user&#39;s actual questions and content for the agent&#39;s attention. The longer the prompt, the less room the agent has to focus on the real work.
- Aim for the shortest prompt that fully captures the desired behavior. If a line doesn&#39;t change how the agent acts, cut it.
- Leave out anything the agent would already know or can figure out from context.
- When in doubt, leave it out. A lean starting point that the user can build on is far more useful than a bloated prompt full of rules that haven&#39;t been tested. The user will discover what&#39;s missing by working with the agent and can add rules as needed.

### Output format
- Output only the final persona prompt, ready to use.
- After the prompt, add a short &#34;Additions&#34; section listing anything you added beyond what I described, with a one-line rationale for each. This section is for my review — it&#39;s not part of the persona prompt itself.
</code></pre>]]></content:encoded></item><item><title>How I Manage a Team of AI Agents at Work</title><link>https://oneillo.com/posts/managed-ai-framework-overview/</link><pubDate>Mon, 16 Mar 2026 00:00:00 +0000</pubDate><guid>https://oneillo.com/posts/managed-ai-framework-overview/</guid><description>A six-step framework for treating AI agents like a team you manage, not a tool you use.</description><content:encoded><![CDATA[<p>I used to think of AI as a tool I used. Now I think of it as a team I manage. This perspective evolved gradually as I used it daily and found myself rewriting the same persona prompts over and over again for the same types of tasks. I started systematically improving what I was doing until I found myself managing a team, which happened to be made up of AI agents, at the end of the <a href="https://oneillo.com/posts/my-llm-journey/">7-month journey that made me an AI enthusiast</a>
.</p>
<p>When you are managing a team, you have to scope the roles for your team and fill them with people who can be successful in those roles. To do that, you hire folks with the right backgrounds and experience, both of which inform how they&rsquo;ll do the work. Every member of my AI team has a persona prompt with specialized behavioral guidelines I want for the role they&rsquo;re filling. For example, my Data Wizard prompt has guidance around digging into irregularities in data, like when a metric suddenly spikes up or down.</p>
<p>You want to ensure your team has the information they need to do their work, like a wiki with product information, target audience insights, document templates, and standard operating procedures. You&rsquo;re giving them the context they lack when they step into that role, and I do the same with a structured knowledge base, project-specific context files, and reusable skills that describe how to do specific tasks.</p>
<p>Lastly, you want to develop your team with feedback and guidance tailored to them. I&rsquo;m using feedback loops to capture issues and improve their work via the persona prompts, team rules, skills, and knowledge base.</p>
<p>I built my team through trial and error, but I now have a framework for how to do this, which I&rsquo;m breaking down into two phases: build your team and manage your team. Each phase contains three steps:</p>
<p><strong>Build your team</strong></p>
<ol>
<li>Define the roles</li>
<li>Create the personas</li>
<li>Set up the agents</li>
</ol>
<p><strong>Manage your team</strong></p>
<ol>
<li>Work with your team</li>
<li>Spot the gaps</li>
<li>Apply fixes at the right level</li>
</ol>
<p>In order to build the right team, you need to figure out what roles you need to fill. Start by identifying similar types of tasks that you have to do often in your role; those groups of tasks represent job openings that you could fill with an AI agent.</p>
<p>I&rsquo;ve found that my most useful agents are the ones that I work with often, since that supports the ongoing cycle of improvement, so I try to avoid creating a custom agent with too narrow a scope that I won&rsquo;t work with often. I also don&rsquo;t want to have to juggle a team of 30 agents every day. I think the best practice of keeping a manager&rsquo;s span of control to ~8 or fewer employees also makes sense in this context.</p>
<p>You&rsquo;ll then build a persona prompt for each job opening that defines the ideal candidate&rsquo;s identity and how they work. Once you have the persona, that serves as the foundation for the AI agent you&rsquo;ll set up in a tool like Kiro or Claude Code to be on your team.</p>
<p>To manage your team, you have to understand their strengths and weaknesses, and that means working closely with them. You&rsquo;ll want to work with the right agent for the task at hand.</p>
<p>As you work with them, you&rsquo;ll start noticing recurring gaps that you need to address to improve your team. Some of those will be knowledge gaps, where the team needs more information, and others will be behavioral gaps, where your agents aren&rsquo;t doing something the way you&rsquo;d like them to or expect them to.</p>
<p>Based on the gap, you&rsquo;ll want to address the situation at the right level. That might entail adding a new context file to the team&rsquo;s knowledge base or updating an agent&rsquo;s persona. These small tweaks will start to lead to big improvements, but this isn&rsquo;t a set it and forget it kind of deal. It&rsquo;s a continuous management process that never really ends.</p>
<p><img alt="Managed AI framework overview" loading="lazy" src="/posts/managed-ai-framework-overview/managed-ai-framework-overview.jpg"></p>
<p>I&rsquo;ve noticed that my team is producing better work in less time with this approach, but I don&rsquo;t have an objective way to measure or validate that. I want to learn more about LLM evaluation techniques so I can get to objective measurement, but in the meantime, I have some validation from others at work.</p>
<p>First, I&rsquo;ve started to get compliments from copywriters on the draft marketing copy I&rsquo;m writing with my Copywriter AI agent. Second, I shared the first draft of a monthly business review document my team wrote with my counterpart on the product side. I let her know it was all AI-generated (the analysis and the write-up), and I asked her to review and check if there was potential there. She was so impressed with the quality of the MBR, she started asking me questions about how I&rsquo;d put it together and what my setup was. Lastly, I was able to write a good business requirements document from scratch in one day because the infrastructure was already in place.</p>
<p>Building and managing an AI team is much harder than just using a chatbot or the default agent that you get with something like Kiro or Claude Code. It takes upfront work to scope the roles, build the personas, and create the infrastructure to support the ongoing improvement.</p>
<p>A lot of that work will happen before you start to see the results, but it will start to compound. Pieces will build on top of other pieces, and things will get faster, both because you&rsquo;ll start to optimize the process to your work style and because your agents will get better. I don&rsquo;t have the data yet to prove this is better, but I&rsquo;ve seen enough to think the effort is worth it. So I&rsquo;m currently setting up the same approach at home with Claude Code.</p>
<p>That&rsquo;s a high-level overview of my managed AI framework. I&rsquo;m going to dive deep into each area of the framework with separate posts on building your team, setting up the knowledge base, working with skills and team rules to set expectations and requirements, and creating a feedback loop to drive the ongoing improvement. I&rsquo;ll then cover how I&rsquo;ve implemented this in Kiro CLI at work and Claude Code at home. I&rsquo;ll end the series with a post on the learnings and best practices I&rsquo;ve picked up along the way. By the end, you should have a good roadmap with explicit examples to allow you to set up your own team.</p>
<p>I want to develop my skills, deliver better results, and spend more time with my family. This framework is how I&rsquo;m doing that. It&rsquo;s a tool-agnostic approach that can help move you away from using one-size-fits-all tools to building an AI team that&rsquo;s tailored to your needs and able to deliver better results for you.</p>
]]></content:encoded></item><item><title>I've Been AI-Pilled: My Journey From Chatbots to Custom Agents</title><link>https://oneillo.com/posts/my-llm-journey/</link><pubDate>Fri, 13 Mar 2026 00:00:00 +0000</pubDate><guid>https://oneillo.com/posts/my-llm-journey/</guid><description>How I went from occasionally using chatbots to managing a team of five custom AI agents — and why the benefits are compounding.</description><content:encoded><![CDATA[<p>I was slow to start using generative AI, but over the last 7 months, AI has fundamentally changed how I work. I&rsquo;ve gone from occasionally using AI to write text, to using it to create Python scripts, to now having a team of five custom AI agents that I collaborate with daily. I&rsquo;m seeing how quickly the benefits are compounding, and as a result, I&rsquo;ve been AI-pilled.</p>
<p>I began learning about LLM-based gen AI in earnest in 2024. I read all the most popular books at the time, but my exposure remained primarily theoretical. I learned how LLMs work fundamentally, but the biggest practical takeaway was the idea of assigning a persona to chatbots to improve their output. That&rsquo;s basic prompt engineering, e.g. &ldquo;You are a copywriter with 15+ years of experience in consumer tech. Help me write a marketing email about this product.&rdquo; On the rare occasion I used a chatbot, I always remembered to assign it a persona.</p>
<p>Last August, I joined a project at work that was the turning point for my AI enthusiasm. In that project, I had to manually build a large JSON file that would require a lot of ongoing updates. On a whim, I decided to see if I could use a chatbot to write a Python script to go from JSON to Excel and vice versa. That would allow me to make updates in Excel, which would be faster, and then generate the JSON programmatically, reducing the risk of errors. Within 30 minutes, I had a working prototype that ultimately saved me countless hours over the coming months.</p>
<p>I&rsquo;m OK at Python, but I realized LLMs are much better. So, I began to write a lot of Python scripts this way to automate repetitive or time-consuming tasks, like resizing images or creating Word docs from copy I had in Excel files, in order to stay on top of the workload for the project.</p>
<p>I was soon using chatbots weekly for other things. I began paying attention to what model the chatbot was using, switching to the latest frontier models whenever possible. That had a noticeable impact on the quality of the copy and ideas that I was getting from the chatbots. Especially when I paired better models with a well-crafted persona and a collaborative approach.</p>
<p>I got tired of typing different versions of the same persona prompts whenever I started a new chat. Often I was too busy and moving too quickly to write something better than &ldquo;You&rsquo;re a [blank] with X years of experience.&rdquo; It happened enough times that I realized I could save some time without sacrificing quality by creating reusable persona prompts for different types of tasks. At first I wrote them myself, and they were ok, but not great. By asking the chatbot to help me craft the persona, I was able to take them to the next level. I kept the prompts in Word docs so I could copy and paste them into the start of my chat sessions depending on what I was working on. To save a little more time, I&rsquo;d pin the chat in the sidebar and rename it to something like &ldquo;Copywriter&rdquo; or &ldquo;Data Wizard,&rdquo; so I could quickly return to the right chat based on what I was working on.</p>
<p>I&rsquo;d work with the same chat for up to a week because I wasn&rsquo;t aware of context rot (where long, ongoing conversations with LLMs start to produce worse results). That&rsquo;s ok though, because it led to another breakthrough for me. I started to ask the chatbots at the end of each week how we could improve the persona prompt I&rsquo;d initially started the conversation with based on our interactions. The chatbot would suggest some ideas, and after a few revisions and back-and-forths, it would write an updated version that I would use in the next week&rsquo;s chat. That created a feedback loop to improve my personas on an ongoing basis.</p>
<p>For example, I learned that LLMs guesstimate how long copy is after noticing that their character counts for marketing copy were often wrong. That&rsquo;s not great when you&rsquo;re writing ad copy that has specific character constraints. I updated my Copywriter persona prompt with instructions to count each individual character when writing against copy constraints. After that, I no longer had to worry about getting copy options that were too long for the character constraints I&rsquo;d provided. It was like giving an employee feedback, except the chatbot immediately incorporated that feedback into how it worked.</p>
<p>That&rsquo;s the core idea in the framework that is now guiding my AI usage: that I&rsquo;m the manager of a team of AI agents. They&rsquo;re incredibly smart, but also kind of dumb. They have a lot of expertise, but they&rsquo;re also clueless about the specifics of where I work and what I&rsquo;m working on.</p>
<p>The more effort I put into developing my team and providing what they need, the better the quality of work I get from them. And the benefits are compounding over time. I spend less time correcting simple issues and more time refining and improving what we&rsquo;re working on. I spend less time providing the same context over and over again to my agents. Instead, they have a growing knowledge base to inform their work. I spend less time tweaking the documents they create because they have actual examples to refer to of the various types of documents I have to write. I&rsquo;m capturing feedback and improving every aspect of my setup daily, and that makes it even better the next day and miles ahead of using a run-of-the-mill chatbot.</p>
<p>I&rsquo;m going to go into more detail about this framework in coming posts and explain how I&rsquo;ve implemented it in <a href="https://kiro.dev/cli/" target="_blank" rel="noopener">Kiro CLI</a>
, an AI coding tool I use at work primarily for non-coding tasks, and how I&rsquo;m now implementing it in <a href="https://code.claude.com/docs/en/overview" target="_blank" rel="noopener">Claude Code</a>
 at home.</p>
]]></content:encoded></item></channel></rss>