ConTejas Code

Shivay Lamba: How to run secure AI anywhere with WebAssembly

5 snips
Jun 23, 2025
In this engaging discussion, Shivay Lamba, a Conference speaker and WebAssembly expert, shares his insights on running AI securely in browsers. He covers the powerful synergy between WebAssembly and AI, emphasizing privacy and performance gains for edge computing. Shivay explains how developers can optimize machine learning models for web use, including techniques like quantization and leveraging TensorFlow.js. He also dives into emerging hardware solutions and the future potential of AI inference APIs, making complex topics accessible and relevant for software developers.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

WebAssembly Enhances AI Privacy

  • Running AI models in the browser using WebAssembly enables privacy by keeping data on the device.
  • It also reduces latency and operating costs compared to cloud-based APIs.
INSIGHT

WebAssembly Runs Beyond Browsers

  • WebAssembly runs anywhere, not just in browsers, including on edge devices with runtimes like WasmEdge.
  • This allows AI workloads on diverse architectures like x86 and RISC-V without needing JavaScript.
ADVICE

Optimize and Convert Models for Browsers

  • Use model quantization to reduce model size and enable running LLMs efficiently on edge devices.
  • Convert models to JavaScript-compatible formats using tools like TensorFlow.js or ONNX.js for browser inference.
Get the Snipd Podcast app to discover more snips from this episode
Get the app