AI today is powerful, but increasingly centralized around cloud platforms, proprietary APIs, and infrastructure-heavy deployments. This makes experimentation expensive and raises concerns around privacy, accessibility, and control — values that the open-source community deeply cares about.
This session introduces WebLLM, an open-source runtime that enables large language models to run directly inside modern web browsers using WebGPU and WebAssembly. With WebLLM, developers can build AI-powered web applications without backend servers, without API keys, and without sending user data outside the device.
The talk focuses on why browser-based, local-first AI matters, what trade-offs it brings, and where it realistically fits in today’s development workflows. Rather than pitching WebLLM as a replacement for cloud AI, the session presents it as a powerful complementary tool that aligns strongly with FOSS principles.
Attendees will walk away with a clear mental model of how WebLLM works, why it’s worth exploring, and how open-source developers can start experimenting with it responsibly.
A clear understanding that AI does not always need cloud servers or backend infrastructure
Awareness of local-first, privacy-preserving AI and why it matters for open-source software
Insight into when browser-based LLMs make sense and when they don’t
Motivation to experiment with WebLLM as an accessible, cost-free entry point into AI
A renewed perspective on how the FOSS community can shape decentralized, user-owned AI