This is the first of a special, three-part Reid Riffs miniseries. Instead of a news-and-headline driven conversation, Reid sits down one-on-one with Parth Patil, an AI engineer and strategist, for a deeper exploration of what it actually means to become AI-native. In this first episode of the series, Parth and Reid discuss how individuals can better leverage LLMs, agents, and creative tools daily. They trace the shift from seeing AI as a productivity boost to understanding it as a meta-tool, as well as unpack techniques like role-based prompting, meta-prompting, and voice as a high-bandwidth thinking interface. Along the way, they discuss the humility required to collaborate with these systems, the move from a single copilot to orchestrating fleets of specialized agents, and how these tools are already reshaping workflows.
Subscribe below to catch the second episode on how large companies can integrate AI, as well as the third episode for startup founders and their early teams building AI-native companies.
For more info on the podcast and transcripts of all the episodes, visit https://www.possible.fm/podcast/
01:07 – When ChatGPT became an “everything tool”
03:11 – Role-based prompting and meta-prompting
07:04 – Ego, humility, and the GPT-4 inflection point
10:41 – Why voice is the highest-bandwidth interface
14:15 – Choosing models and building an AI stack
18:09 – From one copilot to fleets of agents
21:11 – When agents go wrong
25:40 – Using AI as an agent, not a chatbot
28:34 – Building real systems with AI agents
32:49 – Context engineering and advanced prompting
36:03 – Becoming AI-native
40:34 – Closing