2026-04-08 · 6 channels · 20 videos
Claude Code dominates this week — Nate and Cole are both deep in optimizing how developers work with it, but from opposite ends: Nate is focused on cost reduction and token efficiency while Cole is building persistent memory and multi-agent orchestration layers on top of it. Karpathy's viral post about LLM knowledge bases landed in both their ecosystems simultaneously, making it the connective thread. For an app builder, the practical signal is that Claude Code's ecosystem is maturing fast — planning, memory, and cost tooling are all leveling up in the same week.
Nate covered Karpathy's work as a Claude Code force multiplier — framing it as something that '10x'd everyone's Claude Code' by giving it better context and knowledge management. His angle is practical productivity gain.
Cole took Karpathy's concept and built on it — creating self-evolving memory for Claude Code where your own coding sessions become the raw data that gets compiled into a personal knowledge base. He sees it as the foundation for an AI second brain, not just a one-time boost.
If you're building with Claude Code daily, the divergence matters: Nate's approach gets you quick wins now, while Cole's approach compounds over time — consider which matches your workflow horizon.
Nate dedicated multiple videos to cost reduction — Ollama integration for 99% cheaper local inference, 18 specific token hacks, and tips for hitting rate limits less. He's treating cost as the primary adoption barrier.
Cole is less focused on per-token cost and more on efficiency through orchestration — his Archon harness and multi-agent workflows aim to get more done per session by coordinating multiple agents, effectively reducing cost through better task decomposition.
Two valid strategies for the same problem: Nate optimizes the unit economics of each call while Cole optimizes the architecture of the overall workflow — serious builders should probably combine both.
Cole built a full AI second brain with Claude Code — it checks email, calendar, and tasks every 30 minutes, drafts replies in his voice, and gets better over time. He's treating persistent memory as the killer feature that turns a coding assistant into a personal operating system.
Chuck covered Milla Jovovich's AI memory tool as a consumer product — framing it as surprisingly good and accessible. His angle is that memory-augmented AI is hitting mainstream awareness, not just developer tooling.
AI memory is showing up at both the developer-tool layer and the consumer-product layer simultaneously — this is a strong signal that persistent context is becoming a baseline expectation, not a differentiator.
Chris uses Replit Agent to ship apps quickly (built a government data app, hit rate limits from real traffic). His focus is speed-to-market and revenue — the app doesn't need to be perfect, it needs to exist and charge money.
Cole's Archon is about building the infrastructure that makes AI coding agents ship better code — harness engineering, not just rapid prototyping. He's optimizing for quality and repeatability over raw speed.
The market is splitting into 'ship fast with AI' (Replit/Chris) vs 'ship well with AI' (Archon/Cole) — your choice depends on whether you're validating an idea or building something you'll maintain.
Nate is running a full-court press on Claude Code productivity — cost hacks, token optimization, and workflow upgrades. His thesis right now is that Claude Code is the primary dev tool and the bottleneck is cost and token management, not capability.
Ollama integration can cut Claude Code costs by ~99% for certain workflows, and there are 18+ specific token-saving techniques that compound when stacked together.
Andrej Karpathy Just 10x'd Everyone's Claude Code
Deep dive into Karpathy's viral approach of using LLMs to build personal knowledge bases and how to apply it to supercharge Claude Code with better context via Obsidian integration.
Planning In Claude Code Just Got a Huge Upgrade
Covers Claude Code's new planning capabilities (referenced as 'ultraplan' in the URL) that improve how the agent breaks down and executes complex tasks.
How to Use Claude Code for 99% CHEAPER
Condensed or follow-up version of the Ollama + Claude Code cost reduction video.
Ollama + Claude Code = 99% CHEAPER
Tutorial on integrating Ollama (local LLM runner) with Claude Code to offload cheaper tasks to local models and dramatically reduce API costs.
Hitting Claude Code Limits? Here Are My Best Tips.
Shorter companion piece to the 18 token hacks video, focused on strategies for dealing with Claude Code rate limits and usage caps.
18 Claude Code Token Hacks in 18 Minutes
Rapid-fire walkthrough of 18 specific techniques for reducing token consumption in Claude Code, covering prompting strategies, context management, and workflow optimizations.
Claude Code Just Gave Everyone Virtual Pets (April Fools?)
Coverage of Claude Code's April Fools feature — virtual pets inside the terminal — with speculation on whether it hints at persistent agent personality/state.
Chuck is covering the consumer-facing AI wave — local models on phones, celebrity-backed AI tools, and Anthropic platform drama. He's playing the accessible tech explainer role, bridging normie audiences to real AI capabilities.
Gemma 4 running locally on an iPhone without internet (fast enough for real-time Japanese translation) signals that on-device AI is crossing the usability threshold for real apps.
Gemma 4 on the iPhone (local AI, no internet required)
Hands-on demo of Google's Gemma 4 model running locally on an iPhone with no internet connection, including real-time Japanese translation from a pill bottle.
Anthropic says NO MORE OpenClaw!!
Coverage of Anthropic shutting down or restricting OpenClaw (likely a community project or API usage pattern), sparking debate about platform control and developer access.
Cole is building Archon into an open-source 'harness builder' for AI coding agents and layering persistent memory systems on top of Claude Code. His current thesis: the real leverage isn't in the model — it's in the orchestration layer and the memory that accumulates between sessions.
Archon has evolved from an agent framework into a harness engineering platform — Cole is betting that building the scaffolding around coding agents (memory, multi-agent coordination, knowledge bases) is where the defensible value lives.
Full Archon Guide - Build AI Coding Harnesses That Actually Ship (LIVE)
Full live walkthrough of Archon, Cole's open-source platform for building AI coding harnesses — covering what harness engineering is, why it matters, and how to build one from scratch.
I Built Self-Evolving Claude Code Memory w/ Karpathy's LLM Knowledge Bases
Cole implements Karpathy's viral knowledge base concept but feeds it with his own Claude Code session data instead of external articles, creating memory that improves itself over time.
I Taught My Second Brain to Run Multi-Agent Coding Workflows (Live Session)
First public demo of Archon's multi-agent orchestration capabilities — showing how Cole's AI second brain coordinates multiple coding agents on complex tasks.
Full Guide - Build Your Own AI Second Brain with Claude Code
Step-by-step guide to building a personal AI assistant with Claude Code that monitors your email, calendar, and tasks every 30 minutes and learns your preferences over time.
Chris is in full 'build and sell' mode — using AI tools like Replit Agent to ship apps fast, and pitching low-barrier AI services (missed-call text-back, product sourcing) to small businesses. His angle is always revenue-first: what can you build and charge for this week?
The easiest AI-to-revenue path right now may be selling simple automation (like missed-call text-back via GoHighLevel) to local businesses — low technical bar, high perceived value, recurring revenue.
I Built an App the Government Doesn't Want You to See
Chris built an app using government data (likely public records or spending data) with Replit Agent and hit rate limits from real user traffic — demonstrating the build-fast-ship-fast approach.
The Most Profitable Business Everyone Overlooks
Deep dive into agricultural spray drone businesses — how to start one, the economics, and why it's an overlooked high-margin opportunity in 2026.
This Might Be the Easiest Way to Sell AI to Businesses
Tutorial on selling missed-call text-back automation to local businesses using GoHighLevel — positioned as the lowest-barrier entry to selling AI services.
This AI Finds Cheap Products You Can Sell for 5x More
Sponsored walkthrough of Accio 2.0, an AI-powered product sourcing tool that finds arbitrage opportunities across Alibaba, AliExpress, and other platforms.
Central topic across multiple channels — tutorials on cost optimization, token hacks, planning upgrades, memory systems, and harness engineering. Treated as the primary AI coding tool by both Nate and Cole.
From: Nate Herk | AI Automation, Cole Medin
Tutorial on integrating Ollama with Claude Code to offload cheaper tasks to local models, claiming 99% cost reduction for certain workflows.
From: Nate Herk | AI Automation
Cole's open-source harness builder for AI coding agents — evolved from an agent framework into a platform for building coding harnesses with multi-agent orchestration.
From: Cole Medin
Hands-on demo running Google's Gemma 4 locally on iPhone with no internet — real-time translation demo. Framed as a breakthrough in on-device AI usability.
From: NetworkChuck
Referenced as the knowledge base frontend for implementing Karpathy's LLM knowledge base approach with Claude Code.
From: Nate Herk | AI Automation
Used by Chris Koerner to rapidly build and ship a government data app — positioned as a fast prototyping tool for non-technical or speed-focused builders.
From: Chris Koerner on The Koerner Office Podcast
Platform for building and selling AI automation services to local businesses, specifically missed-call text-back workflows.
From: Chris Koerner on The Koerner Office Podcast
AI product sourcing tool for ecommerce arbitrage — sponsored content, prompt-based interface for finding products to resell.
From: Chris Koerner on The Koerner Office Podcast
Viral concept from Andrej Karpathy — using LLMs to compile raw articles into interconnected wikis with health checks. Both Nate and Cole built on this idea, applying it to Claude Code workflows.
From: Nate Herk | AI Automation, Cole Medin