Machine Learning (ML) techniques, especially Deep Neural Networks (DNNs), have been driving innovations in many application domains. These breakthroughs are powered by the computational improvements in processor technology driven by Moore's Law. However, the need for computational resources is insatiable when applying ML to large-scale real-world problems. Energy efficiency is another major concern of large-scale ML. The enormous energy consumption of ML models not only increases costs in data-centers and decreases battery life of mobile devices but also has a severe environmental impact. Entering the post-Moore’s Law era, how to keep up performance and energy-efficiency with the scaling of ML remains challenging.
This talk addresses the performance and energy-efficiency challenges of ML. The core hypothesis can be encapsulated in a few questions. Do we need all the computations and data movements involved in conventional ML processing? Does redundancy exist at the hardware level? How can we better approach large-scale ML problems with new computing paradigms? This talk presents how to explore the elasticity in ML processing and hardware architectures: from the algorithm perspective, redundancy-aware processing methods are proposed for DNN training and inference, as well as large-scale classification problems and long-range Transformers; from the architecture perspective, balanced, specialized, and flexible designs are presented to improve efficiency.
Bio: Liu Liu is an Assistant Professor in the department of Electrical, Computer, and Systems Engineering at RPI. He has a Ph.D. in Computer Science at the University of California, Santa Barbara. His research interests reside in the intersection between computer architecture and machine learning, towards high-performance, energy-efficient, and robust machine intelligence.
Date
Location
Sage 3510
Speaker:
Liu Liu
from Rensselaer Polytechnic Institute