Stanford CS25: V4 I Overview of Transformers

  Рет қаралды 40,722

Stanford Online

Stanford Online

Ай бұрын

April 4, 2024
Steven Feng, Stanford University [styfeng.github.io/]
Div Garg, Stanford University [divyanshgarg.com/]
Emily Bunnapradist, Stanford University [ / ebunnapradist ]
Seonghee Lee, Stanford University [shljessie.github.io/]
Brief intro and overview of the history of NLP, Transformers and how they work, and their impact. Discussion about recent trends, breakthroughs, applications, and remaining challenges/weaknesses. Also discussion about AI agents. Slides here: docs.google.com/presentation/...
More about the course can be found here: web.stanford.edu/class/cs25/
View the entire CS25 Transformers United playlist: • Stanford CS25 - Transf...

Пікірлер: 36
@fatemehmousavi402
@fatemehmousavi402 29 күн бұрын
Awesome, thank you Stanford online for sharing these amazing video series
@3ilm_yanfa3
@3ilm_yanfa3 29 күн бұрын
Can't believe, ... Just today, we started the part about LSTM and transformers in my ML course, and here it comes Thank you guys !
@Drazcmd
@Drazcmd 28 күн бұрын
Very cool! Thanks for posting this publicly, it's really awesome to be able to audit the course :)
@mjavadrajabi7401
@mjavadrajabi7401 29 күн бұрын
Great!! Finally It's time for CS25 V4🔥
@marcinkrupinski
@marcinkrupinski 25 күн бұрын
AMazing stuff! Thank you for publishing this valuable material!
@benjaminy.
@benjaminy. 24 күн бұрын
Hello Everyone! Thank you very much for uploading these materials. Cheers
@lebesguegilmar1
@lebesguegilmar1 28 күн бұрын
Thanks for sharing this course and palestry Staford. Congratulations . Here the Brazil
@JJGhostHunters
@JJGhostHunters 23 күн бұрын
I recently started to explore using transformers for timeseries classification as opposed to NLP. Very excited about this content!
@styfeng
@styfeng 28 күн бұрын
it's finally released! hope y'all enjoy(ed) the lecture 😁
@laalbujhakkar
@laalbujhakkar 28 күн бұрын
Don't hold the mic so close bro. The lecture was really good though :)
@gemini22581
@gemini22581 28 күн бұрын
What is a good course to learn NLP?
@siiilversurfffeeer
@siiilversurfffeeer 26 күн бұрын
hi feng! will there be more cs25 v4 lectures upload in this channel?
@styfeng
@styfeng 25 күн бұрын
@@siiilversurfffeeer yes! should be a new video out every week, approx. 2-3 weeks after each lecture :)
@liangqunlu1553
@liangqunlu1553 22 күн бұрын
Very interesting summarization
@RishiKaura
@RishiKaura 23 күн бұрын
Sincere students and smart
@GerardSans
@GerardSans 28 күн бұрын
Be careful using anthropomorphic language when talking about LLMs. Eg: thoughts, ideas, reasoning. Transformers don’t “reason” or have “thoughts” or even “knowledge”. They extract existing patterns in the training data and use stochastic distributions to generate outputs.
@ehza
@ehza 26 күн бұрын
That's a pretty important observation imo
@junyuzheng5282
@junyuzheng5282 26 күн бұрын
Then what is “reason” “thoughts” “knowledge”?
@DrakenRS78
@DrakenRS78 17 күн бұрын
Do individual Neurons have thoughts , reason or knowledge - or is it once again the collective which we should be assessing
@TheNewton
@TheNewton 13 күн бұрын
This mis-anthropomorphism problem will only grow because each end of the field/industry is being sloppy with it , so calls for sanity will just get derided as time goes on. On the starting side we have academics title baiting like they did with "attention" so papers get attention instead of just making a new word|phrase like 'correlation network' , 'word window' , 'hyper hyper-networks' etc ; or just overloading existing terms 'backtracking', backpropagation etc. And the on other end of the collective full court press is corporations continuing to pass assistant(tools) off as human like with names such cortana, siri etc for the sake of branding and marketing.
@TheNewton
@TheNewton 13 күн бұрын
@@junyuzheng5282 `Then what is “reason” “thoughts” “knowledge”?` reason,thoughts,knowledge, etc are more than are hallucinated in your linear algebra formulas
@GeorgeMonsour
@GeorgeMonsour 22 күн бұрын
I want to know more about 'filters.' Are they human or computer processes or mathematical models? The filters are a reflection, I'd like to understand more about. I hope they are not an inflection, that would be an unconscious pathway. This is a really sweet dip into the currency of knowledge and these students are to be commended however, in the common world there is a tendency developing towards a 'tower of babel'. Greed may have an influence that we must be wary of. I heard some warnings in the presentation that consider this tendency. I'm impressed by these students. I hope they aren't influenced by the silo system of capitalism and that they remain at the front of the generalization and commonality needed to keep bad actors off the playing field.
@IamPotato_007
@IamPotato_007 27 күн бұрын
Where are the professors?
@TV19933
@TV19933 28 күн бұрын
future artificial intelligence i was into talk this probability challenge Gemini ai talking ability rapid talk i suppose so it's splendid
@hussienalsafi1149
@hussienalsafi1149 29 күн бұрын
☺️☺️☺️🥰🥰🥰
@Anbu_Sampath
@Anbu_Sampath 27 күн бұрын
it would be great if CS25: V4 created another playlist in youtube.
@riju1956
@riju1956 29 күн бұрын
so they stand for 1 hour
@rockokechukwu3343
@rockokechukwu3343 28 күн бұрын
Is it okay to cheat in an exam if you have the opportunity to do so?
@ramsever5087
@ramsever5087 21 күн бұрын
what is said in 13:47 is incorrect. Large language models like ChatGPT or other state-of-the-art language models do not only have a decoder in their architecture. They employ the standard transformer encoder-decoder architecture. The transformer architecture used in these large language models consists of two main components: The Encoder: This encodes the input sequence (prompt, instructions, etc.) into vector representations. It uses self-attention mechanisms to capture contextual information within the input sequence. The Decoder: This takes in the encoded representations from the encoder. It generates the output sequence (text) in an autoregressive manner, one token at a time. It uses self-attention over the already generated output, as well as cross-attention over the encoder's output, to predict the next token. So both the encoder and decoder are critical components. The encoder allows understanding and representing the input, while the decoder enables powerful sequence generation capabilities by predictively modeling one token at a time while attending to the encoder representations and past output. Having only a decoder without an encoder would mean the model can generate text but not condition on or understand any input instructions/prompts. This would severely limit its capabilities. The transformer's encoder-decoder design, with each component's self-attention and cross-attention, is what allows large language models to understand inputs flexibly and then generate relevant, coherent, and contextual outputs. Both components are indispensable for their impressive language abilities.
@laalbujhakkar
@laalbujhakkar 28 күн бұрын
Stanford's struggles with microphones continue.
@jeesantony5308
@jeesantony5308 28 күн бұрын
it is cool to see some negative comments in between lots of pos... ✌🏼✌🏼
@laalbujhakkar
@laalbujhakkar 28 күн бұрын
@@jeesantony5308 I love the content, which makes me h8 the lack of thought and preparation that went into the delivery of all that knowledge even more. Just trying to reduce the loss as it were.
Stanford CS25: V4 I Jason Wei & Hyung Won Chung of OpenAI
1:17:07
Stanford Online
Рет қаралды 76 М.
Stanford CS25: V4 I Aligning Open Language Models
1:16:21
Stanford Online
Рет қаралды 13 М.
ПЕЙ МОЛОКО КАК ФОКУСНИК
00:37
Masomka
Рет қаралды 9 МЛН
Follow @karina-kola please 🙏🥺
00:21
Andrey Grechka
Рет қаралды 22 МЛН
Joven bailarín noquea a ladrón de un golpe #nmas #shorts
00:17
Did you find it?! 🤔✨✍️ #funnyart
00:11
Artistomg
Рет қаралды 114 МЛН
MIT Introduction to Deep Learning | 6.S191
1:09:58
Alexander Amini
Рет қаралды 160 М.
The Turing Lectures: The future of generative AI
1:37:37
The Alan Turing Institute
Рет қаралды 545 М.
Geoffrey Hinton in conversation with Fei-Fei Li - Responsible AI development
1:48:12
Arts & Science - University of Toronto
Рет қаралды 96 М.
Intelligent Thinking About Artificial Intelligence
1:04:48
World Science Festival
Рет қаралды 127 М.
Quantum Computing for Computer Scientists
1:28:23
Microsoft Research
Рет қаралды 2,1 МЛН
The Possibilities of AI [Entire Talk] - Sam Altman (OpenAI)
45:49
Stanford eCorner
Рет қаралды 443 М.
Andrew Ng: Opportunities in AI - 2023
36:55
Stanford Online
Рет қаралды 1,8 МЛН
Jeff Dean (Google): Exciting Trends in Machine Learning
1:12:30
Rice Ken Kennedy Institute
Рет қаралды 166 М.
Stanford CS25: V3 I Retrieval Augmented Language Models
1:19:27
Stanford Online
Рет қаралды 131 М.
ПЕЙ МОЛОКО КАК ФОКУСНИК
00:37
Masomka
Рет қаралды 9 МЛН