Compressing Large Language Models (LLMs) | w/ Python Code

Опубликовано: 31 Август 2024
на канале: Shaw Talebi
7,490
320

🗞️ Get exclusive access to AI resources and project ideas: https://the-data-entrepreneurs.kit.co...
🧑‍🎓 Learn AI in 6 weeks by building it: https://maven.com/shaw-talebi/ai-buil...
--
Here, I discuss 3 ways to do model compression on LLMs (i.e. Quantization, Pruning, and Knowledge Distillation/Model Distillation) with example Python code.

Resources:
📰 Blog: https://medium.com/towards-data-scien...
🎥 Training the Teacher Model:    • Fine-Tuning BERT for Text Classificat...  

💻 GitHub Repo: https://github.com/ShawhinT/YouTube-B...
👩‍🏫 Teacher Model: https://huggingface.co/shawhin/bert-p...
🧑‍🎓 Student Model: https://huggingface.co/shawhin/bert-p...
👾 4-bit Student Model: https://huggingface.co/shawhin/bert-p...
💿 Dataset: https://huggingface.co/datasets/shawh...

References:
[1] https://arxiv.org/abs/2001.08361
[2] https://arxiv.org/abs/1710.09282
[3] https://machinelearning.apple.com/res...
[4] https://arxiv.org/abs/1710.09282
[5] https://arxiv.org/abs/2308.07633
[6] https://arxiv.org/abs/2402.17764
[7] https://arxiv.org/abs/1710.01878
[8] https://arxiv.org/abs/1503.02531
[9] https://crfm.stanford.edu/2023/03/13/...
[10] https://arxiv.org/abs/2305.14314
[11] https://www.researchgate.net/publicat...

--
Homepage: https://www.shawhintalebi.com

Intro - 0:00
"Bigger is Better" - 0:40
The Problem - 1:35
Model Compression - 2:14
1) Quantization - 3:11
2) Pruning - 5:44
3) Knowledge Distillation - 8:04
Example: Compressing a model with KD + Quantization - 11:10


Смотрите видео Compressing Large Language Models (LLMs) | w/ Python Code онлайн без регистрации, длительностью часов минут секунд в хорошем качестве. Это видео добавил пользователь Shaw Talebi 31 Август 2024, не забудьте поделиться им ссылкой с друзьями и знакомыми, на нашем сайте его посмотрели 7,490 раз и оно понравилось 320 людям.