Join Michael Yuan as he explores lightweight large language model (LLM) inference with WebAssembly (WASM). In this tech video demo, Michael demonstrates how to run full-scale LLMs like LLaMA on various platforms, from personal laptops to cloud servers, with the efficiency of WASM. He addresses the challenges of running LLMs in cloud environments, offers practical demos, and discusses future applications.
Stay up to date
Sign up to the Navigate mailing list and stay in the loop with all the latest updates and news about the event.