Michael Peng
Panza Personal AI
2024–2025
Electrical Engineering and Computer Science
- AI and Machine Learning
Nir N. Shavit
The advent of large language models (LLMs) has significantly transformed human-computer interaction by providing more intuitive and effective communication. However, the high computational demands of LLMs typically require powerful GPUs and cloud infrastructures, limiting their accessibility and making them difficult to deploy. The “Panza Personal AI” project proposes a novel approach to bring the capabilities of LLMs directly to personal devices, such as laptops and desktops that only have CPU resources. This initiative focuses on employing advanced model compression techniques, sparsification, and sparse execution strategies, enabling efficient local deployment of sophisticated assistant models without sacrificing performance.
I’m participating in the SuperUROP program to learn more about SOTA research techniques in a field I’m passionate about. As a previous researcher in an adjacent field, I’m hoping to refine my understanding of ML research with Prof. Shavit and learn about how low-latency LLM systems are implemented and deployed at scale.