Loading video player...
I sit down with Ras Mic to break down how AI agents actually work and why most people are using them wrong. Ras Mic explains the mechanics of context windows, makes the case that agent md files are largely unnecessary, and shares his step-by-step methodology for building custom skills that make agents dramatically more productive. Whether you're coding with Claude Code or automating workflows with OpenClaw, this episode gives you the foundational knowledge to stop wasting tokens and start getting real results from your AI tools. Timestamps 00:00 – Intro 00:42 – The Models Are Good Now 01:20 – How Context Windows Actually Work 04:55 – The Power of Skills 09:17 – How to create Skills 16:35 – Skill Maxxing 19:05 – What you need too build a project 20:40 – Recursively Building and Improving Skills 29:23 – Context Window Management and Token Efficiency 33:02 – Closing Thoughts Key Points * The models (Opus 4.6, GPT 5.4) are exceptionally good now — the differentiator is the context and harness you build around them. * Agent md and claude md files get loaded into context on every single turn, burning tokens and degrading performance as the context window fills up. 95% of users can skip them entirely. * Skills use progressive disclosure: only the name and description sit in context until the agent determines it needs the full file, saving thousands of tokens per conversation. * The best way to create a skill is to walk through the workflow with the agent step by step, achieve a successful run, and then have the agent write the skill based on that real context. * Recursively refine skills by feeding failures back into the agent and having it update the skill file so the same mistake is avoided going forward. * Scale for productivity by starting with one agent and building up workflows before adding sub-agents — start simple, then expand. Numbered Section Summaries 1. The Models Are Good — Context Is What Matters Ras Mic opens by declaring that the current generation of models, Opus 4.6 and GPT 5.4, are exceptionally capable. The conversation is no longer about which model is "better" in a general sense. What matters now is the quality of context you feed them — that is what separates quality output from slop. 2. How Context Windows Work Ras Mic walks through the anatomy of a context window: system prompt, agent.md files, skills, tools, the codebase, and the user conversation. All of these stack up as tokens, and the window has a hard limit (around 250,000 tokens). When you hit that limit, agents compact — and performance drops. Understanding this structure is the foundation for everything else in the episode. 3. Skills and Progressive Disclosure Skills solve the token-bloat problem. A skill file contains a name, description, and the detailed instructions — but only the name and description are loaded into context. The agent reads the full file only when it determines the skill is relevant. This means a skill costs roughly 53 tokens per turn versus 944+ for an equivalent agent.md file. 4. Building Skills the Right Way Ras Mic shares his methodology: identify a workflow, walk through it with the agent step by step, correct mistakes in real time, and only create the skill after you have completed a successful run. He illustrates this with his sponsor email screening agent — the first attempt returned all-positive results because the agent had no criteria for rejection. 5. Recursively Improving Skills Even after a skill is created, the agent will still hit edge cases and fail. Ras Mic treats each failure as an opportunity: identify the error, have the agent fix it, then tell the agent to update the skill so the failure is documented. After five iterations of this loop on his YouTube analytics report generator, the agent now executes flawlessly across eight data sources in about ten minutes. 6. Scaling for Productivity Over Flash Ras Mic started with a single agent handling everything — email, spreadsheets, research. Only after building reliable skills did he add sub-agents for marketing, business, and personal tasks. He argues that jumping straight to multi-agent architectures (or adopting tools like Paperclip without building foundational workflows first) optimizes for what looks cool rather than what is productive. The #1 tool to find startup ideas/trends - https://www.ideabrowser.com/ LCA helps Fortune 500s and fast-growing startups build their future - from Warner Music to Fortnite to Dropbox. We turn 'what if' into reality with AI, apps, and next-gen products https://latecheckout.agency/ The Vibe Marketer - Resources for people into vibe marketing/marketing with AI: https://www.thevibemarketer.com/ FIND ME ON SOCIAL X/Twitter: https://twitter.com/gregisenberg Instagram: https://instagram.com/gregisenberg/ LinkedIn: https://www.linkedin.com/in/gisenberg/ FIND MIC ON SOCIAL X/Twitter: https://x.com/Rasmic Youtube: https://www.youtube.com/@rasmic