OPT-175B: Open Pretrained Transformer | ML Coding Series

  Рет қаралды 6,737

Aleksa Gordić - The AI Epiphany

Aleksa Gordić - The AI Epiphany

Күн бұрын

Пікірлер: 16
@TheAIEpiphany
@TheAIEpiphany 2 жыл бұрын
Continuing on with Large Language Models! In this one I walk you through the codebase behind OPT-175B model from MetaAI. Feedback is welcome it'll help shape this channel
@vishalsingh-yj8bk
@vishalsingh-yj8bk 2 жыл бұрын
Hi Aleksa I just had one doubt. OPT's architecture is the same as GPT-2 right?
@dimitrismit6714
@dimitrismit6714 2 жыл бұрын
Very good video, goes in a lot of depth. Thanks for the explanation!
@TheAIEpiphany
@TheAIEpiphany 2 жыл бұрын
Thanks Dimitris!
@dimitrismit6714
@dimitrismit6714 2 жыл бұрын
@@TheAIEpiphany Although, I have to note that the format I liked best for your videos is when you first explain it theoretically then show the implementation in code, like the video about VQ-VAEs. I think that way it ties it very well
@TheAIEpiphany
@TheAIEpiphany 2 жыл бұрын
@@dimitrismit6714 thanks for that!
@j.hanleysmith8333
@j.hanleysmith8333 2 жыл бұрын
Next level!
@hooshisar2435
@hooshisar2435 2 жыл бұрын
Awesome thank you for this! Just at a curiosity, would it be easier for Windows users to use the Windows Subsystem for Linux(WSL2) for repos meant to be using Linux?
@TheAIEpiphany
@TheAIEpiphany 2 жыл бұрын
Thanks! Yes in some cases but I think there are still rough edges and I felt more comfortable going this route this time. Have a dedicated Linux machine for ML - that's the best option. I am currently working on assembling my own deep learning rig.
@JupiterNj
@JupiterNj 2 жыл бұрын
Great video, thanks!
@TheAIEpiphany
@TheAIEpiphany 2 жыл бұрын
🙏
@oc1655
@oc1655 Жыл бұрын
hey Aleksa, thank you for these. i love your patience and approach of step-by-step analysis. at 1:05:29 why is the code multiplying the loss by four? for gradient accumulation, i know we divide by steps. but i'm not understanding the reason behind multiplying with the number of (nodes? devices?)
@lhomme_flaneur
@lhomme_flaneur 2 жыл бұрын
hey man i love your videos, just keep it lit. can you also make a tutorial on how to make a really basic ml framework from zero? that would be awesome. everybody is using libs and apis but that'd be cool just to have a simple package, maybe only for linear regression.
@TheAIEpiphany
@TheAIEpiphany 2 жыл бұрын
Thanks!
@TheAIEpiphany
@TheAIEpiphany 2 жыл бұрын
github.com/gordicaleksa/metaseq
GPT-NeoX-20B | BigScience BLOOM | OPT-175B | Training Large Language Models | Papers Explained
47:17
BigScience BLOOM | 3D Parallelism Explained | Large Language Models | ML Coding Series
1:12:00
Aleksa Gordić - The AI Epiphany
Рет қаралды 6 М.
бабл ти гель для душа // Eva mash
01:00
EVA mash
Рет қаралды 6 МЛН
龟兔赛跑:好可爱的小乌龟#short #angel #clown
01:00
Super Beauty team
Рет қаралды 113 МЛН
How I animate 3Blue1Brown | A Manim demo with Ben Sparks
53:41
3Blue1Brown
Рет қаралды 915 М.
Run ALL Your AI Locally in Minutes (LLMs, RAG, and more)
20:19
Cole Medin
Рет қаралды 206 М.
DRM explained - How Netflix prevents you from downloading videos?
18:17
Mehul - Codedamn
Рет қаралды 125 М.
Lucas Beyer (Google DeepMind) - Convergence of Vision & Language
55:08
Aleksa Gordić - The AI Epiphany
Рет қаралды 4,9 М.
What are AI Agents?
12:29
IBM Technology
Рет қаралды 582 М.
When Vision Transformers Outperform ResNets without Pretraining | Paper Explained
23:14
Aleksa Gordić - The AI Epiphany
Рет қаралды 4,6 М.
Open Pretrained Transformers - Susan Zhang  | Stanford MLSys #77
1:00:05
Stanford MLSys Seminars
Рет қаралды 18 М.
OpenAI CFO Says AI Isn't Experimental Anymore
14:02
Bloomberg Technology
Рет қаралды 26 М.
бабл ти гель для душа // Eva mash
01:00
EVA mash
Рет қаралды 6 МЛН