About this episode
This episode focused on the shift toward local, always-on AI systems and the tools making that possible. The conversation started with Pokémon Go as an example of users generating valuable spatial AI data, then moved into NVIDIA GTC, inference hardware, and the broader push toward on-device agents. The second half centered on building workflows with Claude Code, Open Jarvis, mobile coding limitations, Google’s new embeddings model, and how agent permissions change the way people work with coding tools.Key Points Discussed00:01:38 Pokémon Go as unpaid spatial AI field work00:07:13 NVIDIA GTC and the shift from training to inference00:10:08 How chipmakers plan for agentic AI and local inference00:24:16 Stanford Open Jarvis and fully on-device personal AI agents00:30:05 Beth’s Podcast Buddy build and weekend app experiments with Claude Code00:31:45 Claude’s one million token context window discussion00:34:06 Claude usage limits doubling outside peak hours00:35:45 What Claude Code on a phone can and cannot do00:38:34 Google’s new embeddings model for locating objects and multimodal search00:41:05 Brian’s cruise ship hot-and-cold app idea using geolocation and embeddings00:43:29 How Claude remote works from a phone00:50:41 Bypass permissions mode and the risks of letting coding agents run freely00:55:37 Codex full access mode and why Carl prefers its UI00:58:57 Brian’s story about building for fun versus building on deadlineThe Daily AI Show Co Hosts: Brian Maucere, Andy Halliday, Beth Lyons, and Karl Yeh