At 46:50 that is not bigram count table at down rather those are counts resembling smoothed probabilities. Bigram counts won’t come down due to smoothing. I I count cannot be 3.8, probability adjusted count can become float.
@muralidhar40Күн бұрын
DONALD TEUMPS DEATH is correct
@hermanhjaramillo3 күн бұрын
Thank you for your wonderful lectures. It is not clear to me what is the definition of the proximity function phi(x,y)?
@anmoljain11314 күн бұрын
GOD , all my doubts are cleared. Struggled for months and here it comes to end. Thank a lot.
@noadsensehere91954 күн бұрын
Amazing! ALL the stuff of deep learning and graph, fine-tuning, alignment, transformer architecture, etc IS CLUBBED here, Love it!
@hermanhjaramillo6 күн бұрын
At time 59:38 why you have only one positive sample int eh log likelihood and k negative samples. I thought there were the sane number of positive and negative samples "k".
@mauliktailor77537 күн бұрын
This is an awesome course, started as a KZbin suggestion now binge-watching whole series. Thanks for all the handwork to all instructors
@hermanhjaramillo7 күн бұрын
You show the equation C^* = (c+1) N_c+1/N_c. Where did this formula come from? what is the explanation for it? any mathematical proof or at least any intuition for it?
@hermanhjaramillo7 күн бұрын
Excellent courses! It would be best if you considered repeating the questions asked. They are inaudible when coming from students.
@chetansonigara8 күн бұрын
Any Diffrence Between this course and previously completed course this channel on large language model ?
@lablcs27 күн бұрын
Different Instructors, slightly modified course structure, see details here: onlinecourses.nptel.ac.in/noc25_cs45/preview!
@monk-thecoder53888 күн бұрын
are these the same videos as the LLM playlist in this channel.
@lablcs27 күн бұрын
No, you can check the course details here: onlinecourses.nptel.ac.in/noc25_cs45/preview!
@BhuvanWebOsmo8 күн бұрын
Thanks for bringing this course!!
@rounaksaha24948 күн бұрын
That's the course I need also following LLM lecture playlist
@harishravi993610 күн бұрын
@30:42, the definition of monotonicity seems to be incorrect.
@Justuy12 күн бұрын
really enjoyed this lecture. Had an idea of perplexity before but never got its intuition this clearly
@Rahul-by2nn14 күн бұрын
that's great ❤️
@amritkumar-ge4tn14 күн бұрын
As usual awesome content and tshirt ! thanks Yatin :D.
@harshjha677414 күн бұрын
amazing
@amritkumar-ge4tn16 күн бұрын
love Yatin's t-shirts :D
@ShivangiTomar-p7j16 күн бұрын
VERY VERY GOOD. THANKS!!!
@digambarkilledar00329 күн бұрын
Thank you so much for this wonderful series !!
@sanskarkhandelwalАй бұрын
Sir Can you list some good NLP Courses in description or pinned comments
@lablcs2Ай бұрын
I request you to visit our website's materials section: lcs2.in/llm2401, Thank you!
@sanskarkhandelwalАй бұрын
@@lablcs2 ok sir 👍
@IamPotato_007Ай бұрын
I'm so thankful for these lectures 🙏
@tsgaming1173Ай бұрын
👍🏻
@sanskarkhandelwalАй бұрын
Nice lec sir , There is a lack of such quality content on llms ❤
@ashwinkumar5223Ай бұрын
Thankyou Sir
@MohdAthar-t4pАй бұрын
Great! There should be a few lectures that include practical implementations.
@none-hr6zhАй бұрын
In palm ,how does palm model knows when to use python interperter for calculation and when to use only llms?
@Lancer_Soldier8349Ай бұрын
very helpfull lecture
@GauravJain-zo8gtАй бұрын
thank you sir for making this course in public
@amritkumar-ge4tnАй бұрын
beautiful explantation of the concept of Perplexity. Thank u so much prof for enlightening me
@WithRakshithАй бұрын
Sir being from a tier 3 college this series is a real gem for a student like me. I am still in my 5th sem Btech in cse but this pg level course is so well taught i really liked the playlist hatsoff for ur efforts please i beg u to continue more concepts and not to stop
@lablcs2Ай бұрын
I'm glad you are finding the course beneficial. We will continue to explore new concepts.
@MariaM-pu4fxАй бұрын
Is there an error on this slide? 17:19 • Bigram: P(begun | has) - „begun” after „has”. • Trigram: P(begun | season has) - „begun” after „season has”.
@lablcs2Ай бұрын
No. This seems correct. Could you please explain where the confusion is?
@jbm5195Сағат бұрын
@@lablcs2The slide in the video seems wrong.
@funshorts55452 ай бұрын
I never commented on any video, this is my first comment, I am working as AI Engineer currently, thought of refreshing some basics, and started watching Retrieval-based Language Models-I lec 16, I felt it's worth watching complete playlist. whenever I watch some recent technology or subject content on youtube from stanford, used to feel like when will our IIT's teach such latest industry ready techology right in the college. But now I can see that happening, I am heartfully thanking you guys for making it public and accessible on youtube. India needs more such kind of courses on latest trends and techonlogy. I never did it, but I bow to your team of teachers 🙏....I beg you to continue such work, hope other branches follow your path of teaching industry ready/required skills&topics to UG/PG students(btw I am mechanical engineer enjoying working AI ML domain, that's why this last point 😉)
@gopalrkateАй бұрын
I am also on the same path
@chetansonigaraАй бұрын
as data scientists , i am also previous same approch read book and reserch paper & watch stanford lecture
@utkarshtripathi91182 ай бұрын
please continue making these types of videos please
@utkarshtripathi91182 ай бұрын
very ossm exllent lacture sir
@dineshpandey21222 ай бұрын
Thank you Sir for sharing these videos.Really helpful.
@RADHESHYAM-mv5vd2 ай бұрын
What are the pre_requisite to study the course??
@lablcs22 ай бұрын
Watch at kzbin.info/www/bejne/sH7RZGqPrbV3bKs.
@RADHESHYAM-mv5vd2 ай бұрын
Thank you 👍🏻
@kiit83372 ай бұрын
Professor what are the prerequisite to follow the whole playlist
@sanskarkhandelwal2 ай бұрын
Such amazing quality content, On latest technologies, That's why they are IIT profs
@Santosh-t9d6b2 ай бұрын
how the router picks an expert in case all experts have same score ?
@lablcs22 ай бұрын
It is highly improbable due to the random initialization of weights. Should such a scenario occur, the selection process would typically rely on the implementation specifics; most commonly, PyTorch's `topk` function returns the first occurrence.
@Santosh-t9d6b2 ай бұрын
In regular transformer, each token had positionwise FFN. In the presentation, every token is seeing same set of FFNs, why is this? Shouldn't it be FFN1_0, FFN2_0,.. for 1st token and FFN1_n,FFN2_n,... for nth token?
@lablcs22 ай бұрын
Each token is associated with a unique positional encoding, rather than a distinct feedforward neural network (FFN). Within a layer, there is a single FFN that processes all tokens simultaneously. This design is crucial for handling sequences of variable lengths. If separate FFNs were used for each token, it would raise practical concerns about the number of FFNs required to accommodate different sequence lengths.
@Santosh-t9d6b2 ай бұрын
is flash attention of any use in case of edge devices?
@Santosh-t9d6b2 ай бұрын
how is the window size determined?
@lablcs2Ай бұрын
It is a design choice (I.e., a hyperparameter).
@Santosh-t9d6b2 ай бұрын
pls elaborate the autoregression part, specifically how the i/p tokens are fed to decoder . Let say decoder takes 512 input tokens, at first only <SOS> token is sent, but what about rest 511 tokens?
@lablcs2Ай бұрын
During training, the whole input sentence, I.e., all input tokens, are fed together as part of a batch. As it is an auto regressive model, we use masked self-attention so that the transformer can't attend to the future tokens while handling the current token. During inference, the tokens are fed one-by-one as they are generated successively.
@mohdathar55912 ай бұрын
Great
@chiragjain62012 ай бұрын
How to get access to the PPT??
@lablcs22 ай бұрын
Visit lcs2.in/llm2401
@chiragjain62012 ай бұрын
How can I download the slides for this video?
@lablcs22 ай бұрын
Visit lcs2.in/llm2401.
@muhammadmuneebtariq8612 ай бұрын
Sir still same issues persits in google trasnlators, I checked He is docter and She is docter