On-device AI is becoming increasingly important as it provides increased privacy, security, performance and personalization while lowering costs and energy consumption compared to cloud-based AI. Some key milestones in the history of on-device AI include:
The exponential growth in client-side compute capabilities, particularly GPUs, is making on-device AI feasible for more large language models. Currently, on-device AI is being integrated at the system layer in products like:
Apple Intelligence uses techniques like LoRA (Low-Rank Adaptation)that enables real-time training and personalization of AI models on-device with a similar feature set planned to be supported from Chrome. WebAssembly and WebGPU have risen as a frontrunner for deploying cross platform ML solutions leveraging browsertech with popular libraries like transformers js leveraging ONNX ecosystem that targets these backends.
To showcase the potential of on-device AI today, here are some prototype demos:
On-device AI requires rethinking operating systems and computational architectures. Some key shifts include:
We at Tiles Research are on a mission to advance the communication of human intent with machines to design a more natural way of working. We're developing an intelligence-age operating system built with Rust, WebAssembly, and WebGPU. - To build such a novel system we are actively researching at the intersection of:
The future of AI is mediated through an on-device, personalized and private intent router that acts on behalf of the user. An exciting road lies ahead as we architect new intelligence systems that augment the way we do work.