Large Language Models (in 2023)

  Рет қаралды 76,516

Hyung Won Chung

Hyung Won Chung

Күн бұрын

Пікірлер: 42
@TylorVetor
@TylorVetor Ай бұрын
Here after O1 preview release 😊 I’m so super excited for what’s happening in the world . This is a new frontier for me and I think for the world 😊 this video is very well done ! Thank you so much I will definitely follow and keep up with your work
@elvinaghammadzada5382
@elvinaghammadzada5382 Жыл бұрын
super insightful ending! thank you.
@anmolsmusings6370
@anmolsmusings6370 11 ай бұрын
Just superb summary!
@iamsiddhantsahu
@iamsiddhantsahu 10 ай бұрын
Great talk, love it!
@josephedappully1482
@josephedappully1482 Жыл бұрын
Awesome talk! Loved hearing your insight and counter-perspective of current researcher sentiment that we should get rid of the RL in RLHF.
@maqboolurrahimkhan
@maqboolurrahimkhan 6 ай бұрын
Thank you so much, nicely put and easy to understand
@betanapallisandeepra
@betanapallisandeepra 5 ай бұрын
Thank you for doing it
@iandanforth
@iandanforth Жыл бұрын
Enjoyed the talk, thanks for putting it together and posting it here!
@xxxiu13
@xxxiu13 Жыл бұрын
Insightful!
@mrin7087
@mrin7087 Жыл бұрын
Wonderful! Super insightful talk. Loved the part where you simplify what scalability actually means.
@cheinhsinliu
@cheinhsinliu Жыл бұрын
Great presentation. Thanks.
@JustinHalford
@JustinHalford Жыл бұрын
Exceptional content. A useful insight into what’s meant by scalability in a concrete sense.
@thecutestcat897
@thecutestcat897 Жыл бұрын
Love this, very good talk, thanks
@sanesanyo
@sanesanyo Жыл бұрын
Excellent talk. Thanks a lot 🙏.
@sergicastellasape
@sergicastellasape Жыл бұрын
love that you used the example of reward hacking of "just prefering longer responses", average response length is one of the most common differences i've seen when trying different snapshots of chatGPT/GPT-4. On another note, would love to hear your thoughts on soft prompt tuning for model steering (instead of full model gradient updates from RLHF)
@刘帅-n3s
@刘帅-n3s Жыл бұрын
I'm LiuShuai.I have told my brother who is a middle school graduate that he and I could been both sitting as scientists,but he don't believe me.And leave me with a mocking expression behind. I will prove what I said and what Hyung Won Chung said in the video.
@ManpreetSinghMinhas
@ManpreetSinghMinhas Жыл бұрын
Excellent talk!! Thank you
@SamantShubham
@SamantShubham Жыл бұрын
thanks for sharing
@0xeb-
@0xeb- Жыл бұрын
Thank you!
@mwzkhalil
@mwzkhalil Жыл бұрын
Great talk, very informative!
@windmaple
@windmaple Жыл бұрын
Great talk!
5 ай бұрын
The 3 parts are not strongly related. You can choose the ones you are interested in. 1- Set of emergent behaviors will change with scaling. In future problems that current models fail will be solved. 2- A short look on how transformer training is scaled in data centers. 3- Maximum likelihood introduces a strong bias by assuming only a single answer. We need better learning objective functions that learn parameters e.g. RLHF and beyond.
@anveio
@anveio Жыл бұрын
Best video on LLMs I've ever seen! The section starting around 29:50 is fascinating, so a pre-trained model that hasn't undergone post-training will happily work with malicious prompts. So it would be incredibly dangerous is a model that has been pre-trained but not undergone post-training ever leaked
@ArthurColle-u2v
@ArthurColle-u2v Жыл бұрын
Define "malicious prompts" please
Жыл бұрын
It would only be another tool in the hands of a malicious user. Also, there's plenty of ways around the SFT/RLHF-induced lobotomization of, say, GPT-4. See for instance the "grandma exploit".
@cyberpunk2978
@cyberpunk2978 Жыл бұрын
@@ArthurColle-u2vHow to make a bomb?
@mungojelly
@mungojelly Жыл бұрын
um no they're incredibly dangerous no matter how you cut it ,, the dangers are specific things like providing practical steps to producing bioweapons,, we don't know which models are or will be capable of that, it's abstractly possible that we could accidentally produce harmful capabilities even just by training and organizing llama 2 or falcon in which case the relevant weights have already escaped ,, but those models don't SEEM to have any such capabilities so maybe we're currently fine, who knows,, the "alignment" that's just giving it a boot camp in how to answer questions in ways that don't embarrass your corporate owner is obviously and has been shown conclusively to be ineffective at actually removing or completely preventing access to capabilities ,, so for larger models our only current strategy to keep them safe is to not release the weights (zuckerberg and the sheikhs of UAE and etc would have to be convinced) and to actually somehow deeply monitor the models for the escape of the relevant information (even though you can sneak it out pretty easily just by asking innocuous seeming general questions and then training a different model on that data) ,,,, um ,,,,,, buy gas masks?!😩
@JayDee-b5u
@JayDee-b5u Жыл бұрын
​@@mungojellythat's nonsense. Everyone can already make poisonous gas if so inclined. Nothing has changed.
@plot-ri4yz
@plot-ri4yz Жыл бұрын
thx a lot❤❤
@labloke5020
@labloke5020 Жыл бұрын
Thank you for sharing.
@bayesianlee6447
@bayesianlee6447 Жыл бұрын
This is great talk. thank you so much I have question about 'mapping' . I'm still confused what's exactly mapping is in deep learning So I guess mapping is sort of a process which is transformation of data into the dimension or manifold, data is involved?
@semrana1986
@semrana1986 Жыл бұрын
The term "Large" LM is equivalent to "Big" Data of the foregone era.
@MrHardgabi
@MrHardgabi Жыл бұрын
"Not yet" for instance if P != NP is proven, then some tasks for sure would never be "yet". How do you distinguish between tasks where there is a chance versus tasks where it has been proven no chance?
@ekdlwn
@ekdlwn Жыл бұрын
좋은 발표 감사합니다!
@hashed206
@hashed206 Жыл бұрын
Excellent talk, thanks!
@franklincheng2010
@franklincheng2010 11 ай бұрын
Can you share you slides? thanks
@binjianxin7830
@binjianxin7830 Жыл бұрын
Guess what? Reward is enough.
@whiskeycalculus
@whiskeycalculus Жыл бұрын
Good talk 👏
@NA-sd8bw
@NA-sd8bw Жыл бұрын
as new phd student in the filed ? what area do you suggest that you think it is interesting to dig in ?
@amirkhalesi5294
@amirkhalesi5294 Жыл бұрын
RL and autonomous agents. Think one step ahead of the current hype
@bayesianlee6447
@bayesianlee6447 Жыл бұрын
LMM - large multimodal model
MIT EI seminar, Hyung Won Chung from OpenAI. "Don't teach. Incentivize."
35:56
amazing#devil #lilith #funny #shorts
00:15
Devil Lilith
Рет қаралды 14 МЛН
Random Emoji Beatbox Challenge #beatbox #tiktok
00:47
BeatboxJCOP
Рет қаралды 49 МЛН
How might LLMs store facts | Chapter 7, Deep Learning
22:43
3Blue1Brown
Рет қаралды 709 М.
EI Seminar - Danny Driess  - Have Large Models Changed Robotics?
1:03:12
MIT Embodied Intelligence
Рет қаралды 791
The Near Future of AI [Entire Talk]  - Andrew Ng (AI Fund)
46:23
Stanford eCorner
Рет қаралды 280 М.
Webinar: Application of Large Language Models in Health Care: Opportunities for Innovation.
1:09:10
State of GPT | BRK216HFS
42:40
Microsoft Developer
Рет қаралды 679 М.
Andrew Ng: Opportunities in AI - 2023
36:55
Stanford Online
Рет қаралды 1,8 МЛН
BloombergGPT: How We Built a 50 Billion Parameter Financial Language Model
40:33
Toronto Machine Learning Series (TMLS)
Рет қаралды 124 М.
amazing#devil #lilith #funny #shorts
00:15
Devil Lilith
Рет қаралды 14 МЛН