The New Stack Podcast

From Cloud Native to AI Native: Where Are We Going?

Nov 28, 2025
In this engaging discussion, guests include Kate Goldenring, a software engineer at Fermyon Technologies who champions WebAssembly for secure edge model deployment, and James Harmison from Red Hat, focusing on AI infrastructure. Shaun O'Meara, CTO at Mirantis, emphasizes the need for deep infrastructure awareness for AI workloads. Sean O'Dell from Dynatrace highlights how AI amplifies observability challenges. Jonathan Bryce of CNCF stresses the importance of robust inference layers, ensuring that discussions extend beyond just chatbots to the critical aspects of AI-native systems.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Inference Is The Core AI Layer

  • Jonathan Bryce argues inference is the overlooked core of AI deployment, not just chatbots and agents.
  • He says inference layers must serve models to answer questions and make predictions at scale.
ADVICE

Use WebAssembly For Low-Latency Inference

  • Kate Goldenring recommends using serverless WebAssembly to run inference quickly at the CDN and device edge.
  • She advises bundling models with Wasm components for low-latency, sandboxed execution on GPU hardware.
INSIGHT

Wasm Enables Portable AI Tooling

  • WebAssembly provides polyglot portability and sandboxed security for AI components across regions.
  • Kate highlights MCP server implementations in Wasm as a growing, interoperable pattern for AI tooling.
Get the Snipd Podcast app to discover more snips from this episode
Get the app