Lecture 10 - Knowledge Distillation | MIT 6.S965

  Рет қаралды 16,401

MIT HAN Lab

MIT HAN Lab

Күн бұрын

Lecture 10 introduces knowledge distillation, including self and online distillation, distillation for different tasks. This lecture also introduces network augmentation, a training technique for tiny machine-learning models.
Keywords: Knowledge Distillation, Online Distillation, Self Distillation, Network Augmentation
Slides: efficientml.ai...
--------------------------------------------------------------------------------------
TinyML and Efficient Deep Learning Computing
Instructors:
Song Han: songhan.mit.edu
Have you found it difficult to deploy neural networks on mobile devices and IoT devices? Have you ever found it too slow to train neural networks? This course is a deep dive into efficient machine learning techniques that enable powerful deep learning applications on resource-constrained devices. Topics cover efficient inference techniques, including model compression, pruning, quantization, neural architecture search, and distillation; and efficient training techniques, including gradient compression and on-device transfer learning; followed by application-specific model optimization techniques for videos, point cloud, and NLP; and efficient quantum machine learning. Students will get hands-on experience implementing deep learning applications on microcontrollers, mobile phones, and quantum machines with an open-ended design project related to mobile AI.
Website:
efficientml.ai/

Пікірлер
Better not Bigger: Distilling LLMs into Specialized Models
16:49
Snorkel AI
Рет қаралды 4,9 М.
БОЙКАЛАР| bayGUYS | 27 шығарылым
28:49
bayGUYS
Рет қаралды 1,1 МЛН
UFC 287 : Перейра VS Адесанья 2
6:02
Setanta Sports UFC
Рет қаралды 486 М.
Quantization vs Pruning vs Distillation: Optimizing NNs for Inference
19:46
How ChatGPT Cheaps Out Over Time
9:28
bycloud
Рет қаралды 43 М.
Lecture 10 - Knowledge Distillation | MIT 6.S965
1:07:26
MIT HAN Lab
Рет қаралды 4,5 М.
Understanding AI from Scratch - Neural Networks Course
3:44:18
freeCodeCamp.org
Рет қаралды 498 М.
Knowledge Distillation: A Good Teacher is Patient and Consistent
12:35
20. Savings
1:14:29
MIT OpenCourseWare
Рет қаралды 1,1 МЛН
Lecture 05 - Quantization (Part I) | MIT 6.S965
1:11:43
MIT HAN Lab
Рет қаралды 16 М.
Distilling the Knowledge in a Neural Network
19:05
Kapil Sachdeva
Рет қаралды 20 М.
Visualizing transformers and attention | Talk for TNG Big Tech Day '24
57:45