- When: Wednesday, March 16, 2022 from 11:00 AM to 12:00 PM
- Speakers: Tong (Tony) Geng
- Location: ZOOM only
- Export to iCal
Abstract:
In the past decade, Artificial Intelligence through Neural Networks (NNs) has penetrated virtually every aspect of human life. A basic problem for many NN deployments is that their target applications pose stringent requirements on latency, throughput, and accuracy. Much research has therefore gone into various aspects of improving NN performance. My research focuses on the hardware acceleration of NNs.
The problem in creating hardware to accelerate NNs is that optimized NN models, with redundant and superfluous computations largely pruned, typically have significant irregularities, making them hardware-unfriendly. As these algorithmic optimization methods continue to be developed, NN models become ever more irregular, a trend likely to continue for some time and expand beyond Machine Learning. One common way to manage irregularity is by eliminating it through offline approaches, e.g. model regularization through training and preprocessing, but all irregularities cannot be eliminated this way, and many require mechanisms that handle irregularity on-the-fly.
In this talk, I will discuss my approach: instead of regularizing models to make them hardware-friendly, I leverage hardware flexibility from reconfigurability to create novel architectures and systems that are friendly to irregularity in applications. These architectures can handle irregularity for a wide range of NN domains, from real-time single-node inference to high-performance large-scale training. I will begin my talk by briefly introducing my overall framework of hardware flexibility as the key to accelerating optimized NNs, followed by an in-depth discussion of my architectural solutions to handle extreme irregularities in Graph Neural Networks (GNNs). GNNs are drastically expanding the applications of machine learning methods and appear to pose the most significant computational challenges yet. Finally, I will present my vision for the future of graph intelligence and the importance of heterogeneity in future architecture and system research, especially for ML in the post-Moore-law era.
BIO:
Tong (Tony) Geng is a postdoc in the Physical & Computational Sciences Directorate (PCSD) at the Pacific Northwest National Laboratory (PNNL). He received his Ph.D. in Computer Engineering at Boston University in 2021. His research interests are at the intersection of computer architecture & systems, machine learning, graph intelligence, and high-performance computing. He is the recipient of the Outstanding Postdoc Award at PNNL in 2021 and the best paper award in ICCD 2021. He has served on the TPC of HPCA, IPDPS, FPL, and etc. His papers have appeared in MICRO, HPCA, SC, TPDS, TC, ICS, and ICCAD.
Posted 2 years, 8 months ago