Olivia Weng (UC San Diego)- Efficient and Resilient Neural Networks for On-chip Inference
Abstract: Scientific applications are increasingly using neural networks (NNs) at the edge as a fundamental tool for advancing fields such as particle physics and materials science. As the scientific instruments used in these experiments become more advanced, they produce a lot more data (e.g., 40 TB/s) than before. As a result, scientists are relying on edge NNs, which have more capabilities than traditional algorithms, to process the data. To process data quickly enough, these scientific edge NNS have unique requirements. They must (1) be heavily quantized and (2) execute fully on chip. Even more so, these scientific NNs often operate in high radiation environments (1000X that of space). My talk focuses on using hardware-software co-design to create efficient, fault-tolerant computer architectures for NNs that execute fully on chip so that they meet the strict latency and throughput requirements laid out by these scientific experiments to advance research in their fields. I will talk about two projects: (1) scaling up lookup-table-based NNs and (2) codesigning fault-tolerant edge NNs. This work provides insights and tradeoffs that will help scientists better run their NNs on specialized hardware such as FPGAs and ASICs to successfully perform their experiments.
Speakers
Olivia Weng
Olivia Weng is a PhD candidate in Computer Science and Engineering working with Professor Ryan Kastner at UC San Diego. She received her BS in Computer Science at the University of Chicago. Her research focuses on using hardware-software co-design to create efficient, fault-tolerant computer architectures for machine learning. Her work has led to collaborations with researchers at AMD as well as high energy physicists at Fermilab and CERN who seek to deploy machine learning at the edge to discover new physics.