TinyML¶
Rather than adding more compute power, focus on improving compute efficiency
Will mainly focus on the following applications: Speech, Computer Vision, NLP
Topics¶
- Hardware
- Architecture & Dataflow
- Metrics and Analysis
- Efficiency
- Micro-architecture/Circuits
- Model Optimization
- Quantization
- Pruning
- Knowledge distillation
- AutoML
- Software: Optimize DNN operations through software compilation/kernel implementations
- Domain-specific compilers; eg: TVM
- Kernel implementations
- Mapping onto hardware
- Systems
- Pre/Post Processing
- Distributed training
- Federated learning
- Environmental issues
Pre-Requisites¶
- Computer archictecture
- Machine Learning
- Python programming
- PyTorch Basics
Reading¶
References¶
- Machine Learning Hardware and Systems (Cornell Tech, Spring 2022)
- Videos
- Material
- TinyML and Efficient Deep Learning Computing | EfficientML.ai - MIT HAN Lab
- Tiny Machine Learning | UPenn
- AutoDL | Applied Deep Learning