Talks AI

Distributing AI with Python in the Browser: Edge Inference and Flexibility Without Infrastructure

Friday, May 15th, 2026 2:45 p.m.–3:15 p.m. in Grand Ballroom A

Presented by

Fabio Pliger

Description

AI models are often complex, require a lot of resources to run and many considerations in terms of security, privacy, and performance. An emerging ecosystem of tools is making it possible to use the Browser as a great platform to run Python and consume (either by connecting or directly running) AI models. This opens the door to distributing AI to the edge without provisioning servers, managing GPUs, or handling complex infrastructure.

This talk shows how to use Python in the browser as a secure, sandboxed runtime for both edge inference or using it in agentic workflows where models are provided via an API. We will walk through an architecture where a Python runtime and models are delivered as assets, with all inference happening in different ways in (or through) the user’s browser. Along the way, we will look at how this fits into the broader AI‑in‑browser ecosystem: interoperating with existing JavaScript model runners, leveraging WebGPU/WebNN for acceleration, mixing Python‑based preprocessing with client‑side inference engines or simply a gateway to AI services like OpenAI or Anthropic.

Security and privacy are first‑class concerns: the browser sandbox limits what code can do, and keping inference and related compute on the client means user data does not need to leave the device. We will also be honest about the trade‑offs: bundle sizes, model constraints, performance, and when a hybrid edge/cloud approach makes more sense than going fully client‑side. Attendees will leave with concrete patterns, example architectures, and a clear path to experimet with Python‑powered AI at the edge, no special infrastructure required.

Search