Song Han

Assistant Professor, MIT EECS

Efficient AI

with Tiny Resource

Accelerate Deep Learning Computing


Song Han is an assistant professor at MIT’s EECS. He received his PhD degree from Stanford University. His research focuses on efficient deep learning computing. He proposed “deep compression” technique that can reduce neural network size by an order of magnitude without losing accuracy, and the hardware implementation “efficient inference engine” that first exploited pruning and weight sparsity in deep learning accelerators. His team’s work on hardware-aware neural architecture search (ProxylessNAS, Once-for-All Network (OFA), MCUNet) was integrated in FacebookAmazon, Microsoft, received six low-power computer vision contest awards in flagship AI conferences. Song received Best Paper awards at ICLR’16 and FPGA’17, multiple faculty awards from Amazon, SONY, Facebook, NVIDIA and Samsung. Song was named “35 Innovators Under 35” by MIT Technology Review for his contribution on “deep compression” technique that “lets powerful artificial intelligence (AI) programs run more efficiently on low-power mobile devices.” Song received the NSF CAREER Award for “efficient algorithms and hardware for accelerated machine learning” and the IEEE “AIs 10 to Watch: The Future of AI” award.

Group Website, Google Scholar, YouTube, Twitter, Github, LinkedIn

Research Interests

TinyML, enable deep learning on mobile devices, efficient algorithms and hardware for computation-intensive AI applications. 

We actively collaborate with industry partners (AI+IoT, automotive, semiconductor, consumer electronics, cloud). Many research projects have successfully influenced many products. Welcome to email me for collaboration.

Model Compression / AutoML / NAS: [MLSys’21][NeurIPS’20, spotlight][NeurIPS’20][ICLR’20][CVPR’20][CVPR’20][ICLR’19][CVPR’19, oral][ECCV’18][ICLR’16, BP][NIPS’15]
Efficient AI on edge devices: Video / Point Cloud / NLP / GAN: [ICRA’21][CVPR’21][NeurIPS’20][ACL’20][CVPR’20][ECCV’20][ICLR’20][NeurIPS’19, spotlight][ICCV’19]
HW for ML: [MICRO’21][HPCA’21][HPCA’20][FPGA’17, BP][ISCA’16]
ML for HW: [DAC’21][DAC’20][NeurIPS’19 W]


  • Samsung Global Research Outreach (GRO) Award, 2021
  • IEEE “AIs 10 to Watch: The Future of AI” Award, 2020
  • NSF CAREER Award, 2020
  • NVIDIA Academic Partnership Award, 2020, 2021
  • MIT Technology Review list of 35 Innovators Under 35, 2019
  • SONY Faculty Award, 2017/2018/2020
  • Amazon Machine Learning Research Award, 2018/2019
  • Facebook Research Award, 2019
  • Best paper award, FPGA’2017
  • Best paper award, ICLR’2016

Competition Awards


MCUNet [NeurIPS’20 spotlight]:
– WiredAI Algorithms Are Slimming Down to Fit in Your Fridge 
– MIT NewsSystem brings deep learning to “internet of things” devices
– Stacey on IoTResearchers take a 3-pronged approach to Edge AI 
– Morning Brew, Researchers Figured Out How to Fit More AI Than Ever onto Internet of Things Microchips
– IBMNew IBM-MIT system brings AI to microcontrollers – paving the way to ‘smarter’ IoT
– Analytics Insight, Amalgamating Ml And Iot In Smart Home Devices
– TechableMITがIoTデバイス向けのコンパクトなAIシステムを開発!
– Tendencias, El aprendizaje profundo impulsa el Internet de las cosas

SpAtten [HPCA’21]
– MIT News, A language learning system that pays attention — more efficiently than ever before

DiffAugment [NeurIPS’20]:
– Venture BeatMIT researchers claim augmentation technique can train GANs with less data

Once-For-All Network [ICLR’20]:
– Venture BeatMIT aims for energy efficiency in AI model training
– MIT NewsReducing the carbon footprint of artificial intelligence
– AI DailyNew MIT Architecture May Lead To Smaller Carbon Footprints For Neural Networks
– Qualcomm, Research from MIT shows promising results for on-device AI
– TechHQ, 
How MIT is making ground towards ‘greener’ AI
– Singularity Hub, This ‘Once-For-All’ Neural Network Could Slash AI’s Carbon Footprint
– InhabitatMIT moves toward greener, more sustainable artificial intelligence

Temporal Shift Module [ICCV’19]:
– NVIDIANew MIT Video Recognition Model Dramatically Improves Latency on Edge Devices
– MIT Technology ReviewPowerful computer vision algorithms are now small enough to run on your phone
– EngadgetMIT-IBM developed a faster way to train video recognition AI
– MIT NewsFaster video recognition for the smartphone era

ProxylessNAS [ICLR’19]:
– IEEE SpectrumUsing AI to Make Better AI
– MIT NewsKicking neural network design automation into high gear



Ph.D. Stanford University, advised by Prof. Bill Dally

B.S. Tsinghua University


Email: FirstnameLastname [at] mit [dot] edu

Students who are interested in joining us for remote collaborations, please email:

We accept on-site visiting undergrad/MS students starting from Jan 2022.