RLHF+CHATGPT: What you must know

  Рет қаралды 66,707

Machine Learning Street Talk

Machine Learning Street Talk

Жыл бұрын

Pod version: podcasters.spotify.com/pod/sh...
Support us! / mlst
MLST Discord: / discord
In this video, Minqi Jiang (minch.co/), a research scientist at University College London & Meta AI, discusses the capabilities of language models like GPT-3 and the effects of Reinforcement Learning from Human Feedback (RLHF) on these models. He explains how RLHF helps make language models more user-friendly by providing a more reliable interface to specific parts of the model, but also reduces the diversity of their outputs, which might be less desirable for certain creative tasks.
Minqi explains that a base language model, like GPT-3, is essentially trained to model the whole internet of text. This vast distribution covers both good and bad content, creating a chaotic and enormous model that can provide a wide range of responses. When prompted with a task, it's difficult to anticipate how the model will complete it.
RLHF comes into play by fine-tuning the model on a reward signal that is learned from human preference data. This process introduces a bias into the model, making it more likely to generate outputs that were favored by the humans providing the preference data. This results in more reliable answers, but at the cost of diversity in the model's output.
The process of RLHF can be thought of as a pruning process, where the aim is to remove the bad or undesired parts of the probability distribution and focus on the good ones. This can be seen as a form of robustification, but also potentially reduces the model's creativity, as it becomes more convergent in its outputs.
In conclusion, RLHF provides a useful way to fine-tune language models to provide more reliable and user-friendly outputs, but it can also reduce the diversity and creativity of the model's outputs. This trade-off between reliability and diversity is important to consider when using language models for various tasks and applications.
Credit for shoggoth meme: / anthrupad

Пікірлер: 173
@DurandalLM
@DurandalLM Жыл бұрын
Damn, Minqi is a fantastically clear speaker.
@AICoffeeBreak
@AICoffeeBreak Жыл бұрын
1:55 "It's like multiple personalities times a billion people on the Internet, that's what it's modelling". Love this sentence. Thanks for posting this great content!
@danielotero8281
@danielotero8281 Жыл бұрын
I loved this. U should keep making clips for us who don’t have the time listen to the whole podcast
@TechRecipeSyndicate
@TechRecipeSyndicate Жыл бұрын
ML Street Talk Clips
@CYI3ERPUNK
@CYI3ERPUNK Жыл бұрын
minqi sounds like a very intelligent dude , glad to have him in the field ; would luv to see this full interview
@human_devops
@human_devops Жыл бұрын
Great philosophical take on the "what do you want from AI/inspiration in the first place?". You often don't know what you want - you don' know what you don't know - you want inspiration - therefore once you spot the patterns in the output of LLM you move on c.f. spotting midjourney 'style'
@jhbauer
@jhbauer Жыл бұрын
Sounds just like the effect of culture and parenting on us- in other words what social conditioning might “do” to human development. Imagine the wild mess of what we really would be without this. - the “wild child” . From kindergarten to school to job - we mostly aim to satisfy the demands of the more powerful group or beings and thus push some interesting and capable parts of our selves in the background (forever). On the other hand certain people seem to know how to create a prefect persona (UI) - e.g. look at what it takes to be a “politician” - while at the other extreme some struggle lifelong with feelings of unworthiness because they just cannot ever perform this persona thing (the socially desired UI) successfully. They may remain outsiders just to avoid contact because they can never stand to feel the conflict of not getting right what is wanted from them (negative “reward”). In essence, a socialization by which a child is supported to trust in itself, be independent but still empathic and open towards others might still be the benchmark to achieve in terms of human reinforcement learning. Hope you still love me despite this lengthy piece?
@CambodianGeographic
@CambodianGeographic 10 ай бұрын
I was thinking similar things when I was watching. We are given data and trained to satisfy human preferences of others.
@Eingefallen
@Eingefallen 4 ай бұрын
Holy shit.
@twentytwentyeight
@twentytwentyeight Жыл бұрын
This short snippet was so refreshing and light, new sub
@videowatching9576
@videowatching9576 Жыл бұрын
This is a fantastic video. Fascinating to think about how RLHF anchors to a particular level of expertise for example in answering a question.
@HenryTechBoy
@HenryTechBoy Жыл бұрын
Super high level content on RLHF and how it relates to ChatGPT. Instant sub. Thank you for this and I look forward to more vids.
@CodexPermutatio
@CodexPermutatio Жыл бұрын
Excellent content. I'm looking forward to the full interview.
@Jose-pq4ow
@Jose-pq4ow Жыл бұрын
Wow! Nice explanation of RLHF, you should upload more of these clips
@jamieosh70
@jamieosh70 Жыл бұрын
This is really well explained and insightful.
@angelomenezes12
@angelomenezes12 Жыл бұрын
That was a fantastic explanation! Minqi is a great speaker indeed!!
@rakhyroad
@rakhyroad Ай бұрын
Well laid out, and both talented speakers! Very much how I've been dissecting and digesting this re-biasing layer of RLHF. Which LLM maps user intent the best on average is just one spoke for a healthy system.
@SkittleBombs
@SkittleBombs Жыл бұрын
"I find RLHF+CHATGPT really interesting because it's amazing how AI technology has advanced so much. However, I also find it annoying sometimes when the responses aren't exactly what I was looking for or when the chatbot doesn't seem to understand my questions. Overall though, it's still pretty impressive!" chat gpt created this comment based on " can you make an opinion that i can use as a youtube comment about how RLHF+CHATGPT interests me and what i think make it annoying ?"
@duudleDreamz
@duudleDreamz Жыл бұрын
Great video. Fascinating stuff. More of this please.
@ShotterManable
@ShotterManable Жыл бұрын
Great! Thanks for sharing. Where can we watch the full interview?
@MachineLearningStreetTalk
@MachineLearningStreetTalk Жыл бұрын
Soon! It was an amazing interview 🙏
@MrAvnitarora
@MrAvnitarora Жыл бұрын
@@MachineLearningStreetTalk waiting !!
@kevalan1042
@kevalan1042 Жыл бұрын
@@MachineLearningStreetTalk greatly anticipated!
@pratikbehera2329
@pratikbehera2329 Жыл бұрын
Waiting 😄
@willd1mindmind639
@willd1mindmind639 Жыл бұрын
Bias isn't a bad word because everything in nature exhibits some kind of bias in terms of preference or desire, such as species of honeybee are biased towards certain flowers. And in big data terms, the value of all this information on the internet is being able to analyze, quantify and model the various preferences and behaviors across large populations. That is the "value" of big data and correspondingly, AI data has no value because in itself it does not embody anything but multiple personality disorder. However, one thing to add value to that kind of AI is for it to embody a specific set of biases or beliefs in a predicable fashion and stick to it. So for example if 5 instances of chatgpt had 5 different embodied beliefs on a topic, they could debate each other and come to some sort of qorum. Something like that has value in problem solving, which would be more like having 5 different experts in AI hash it out over some topic in Neural networks such as you see on twitter. This is another variation of taming the multi personality monster and having it embody or model one of the many sets of beliefs and biases it has been trained on.
@willd1mindmind639
@willd1mindmind639 Жыл бұрын
@@AB-wf8ek AI as we know it is based on statistics which is explicitly designed to understand differences in a population. In this day and age of big data, that means preferences, beliefs, "biases", opinions, etc. For example, when 500 people go into a store, you are going to have 500 patterns of shopping data based on "bias", which in this case mostly means preference. That said, within statistics, there is a bad definition of bias, which means weighted more towards one set of facts or data over another. That is a different type of "bias" based on sampling and modeling. So AI models, based on statistics, like chatgpt already exist to enhance critical thinking. The problem is it cannot "embody" a set of beliefs, preferences and opinions of its own or even "take a side" in an argument because of its multiple personality disorder (aggregation of numerous sets of opinions, beliefs, etc). So, having it be able to act as an agent representing one side of a debate or as a virtual model of a set of opinions and beliefs would be a powerful tool for critical thinking. I just am curious whether it is possible to do this with chatgpt.
@willd1mindmind639
@willd1mindmind639 Жыл бұрын
@@AB-wf8ek None of what you said has anything to do with what I actually posted. It is trained on data from millions of actual real people. Real people have individuals sets of beliefs, opinions, preferences, biases and ways of thinking. This is a fact of life and why "big data" has value in terms of aggregating such things from across a large number of actual people. Again, the point is whether chatgpt today or at some point in the future can actually embody a set of beliefs, opinions and biases as a virtual representation of an actual individual. Right now it cannot. The value of this would be in the case of having "debates" between different instances of chatgpt where they each represent one side or aspect of an argument, point of view, school of thought, or expert opinion on a subject. Such "individualized" sets of opinions, perspectives or "biases" would be useful as a simulation of using such debate models to problem solve. This isnt about perfect fitting as opposed to taming the multiple personality disorder, by having it exhibit a specific set of "values", "biases" or "weights" as an individualized or embodied virtual "agent" in a certain context.
@willd1mindmind639
@willd1mindmind639 Жыл бұрын
@@AB-wf8ek That is assuming that chatgpt has indeed read that specific author. The problem here is these assumptions are not always valid, but yes in theory this should be possible, but again out of the box, your mileage may vary. And what I was talking about goes beyond writing in the style of and goes to embodying a set of ideas and beliefs of in a debate with another AI agent embodying an opposing set of beliefs and views. Again, we are talking about the AI agent "embodying" certain values and principles as an exercise in having a defined "personality". In order to get to what I was talking about, you would actually need training data that is not simply "all the data on the net". You would actually need data representing authoritative knowledge on specific topics, disciplines, people, their views, writings and opinions in the training set. Not to mention ways to update the training with new data on a more regular basis. And the problem with this marketing and hype around chatgpt is anything and everything you may ever want is "in there" but in reality most likely it isn't. So for example, if I would want two instances of chatgpt to represent two "medical experts" debating a particular diagnosis, it would require augmented training that is not there "out of the box". As in the course of this debates these agents would need to be referencing journals and published papers supporting their specific views and why they came to a particular conclusion. So yes, I am sure it can do it, but to get it to cover specific use cases in a fashion that is rigorous enough and accurate enough for "mission critical" use would require more work. Right now it is mostly just a nice tech demo showing what is possible and yes doing some very interesting things. And as proof of this I would point to how Wolfram Alpha and Khan Academy are using chatgpt but in more of a limited fashion to be a natural language interface but not as an "expert" in any specific subject.
@marcovirgolin9901
@marcovirgolin9901 Жыл бұрын
I was hoping to hear about RLHF fine-tuning vs classic CE-based fine-tuning. Anyway superlative content as usual :)
@NoidoDev
@NoidoDev Жыл бұрын
If I understand correctly this could be used to train a system based on responding like someone from a specific community would respond, instead like a random internet user. Also, this strikes me as useful for making systems having empathy. Looking at different perspectives and judgements on a topic and also adapting the responses based on who they're talking to.
@ianfinley89
@ianfinley89 Жыл бұрын
Really enjoyed the conversation, especially with respect to why you would prefer a "divergent versus open ended LLM."
@Sciencehub-oq5go
@Sciencehub-oq5go Жыл бұрын
That has been a fantastic episode.
@WirewalkerDon
@WirewalkerDon Жыл бұрын
I see some parallels with deciphering Search Intent. Working on a Search Engine we tried to distinguish between outcome preferences. For example, does "Beatles" mean you want to listen to music, or purchase swag, or read about their history? Hard to tell but important.
@greatcesari
@greatcesari Жыл бұрын
…or read about bugs.
@Hexanitrobenzene
@Hexanitrobenzene Жыл бұрын
Ideal approach is for the system to ask.
@natecodesai
@natecodesai 6 ай бұрын
Very well explained!
@CREWorxMedia
@CREWorxMedia Жыл бұрын
amazing insight! Where can we find the rest of this video?? Why was it cut off?
@ObservingBeauty
@ObservingBeauty Жыл бұрын
Very informative. Thanks
@sapienspace8814
@sapienspace8814 Жыл бұрын
Fascinating talk, the "aliasing" reminds me of overlapping qualitative (language / label) values in Fuzzy Logic. I think one wanted a greater level of "creativity", a way to do that might be to increase the magnitude of the pseudo-random noise generator being used, or have a more random signal (less pseudo random).
@WirewalkerDon
@WirewalkerDon Жыл бұрын
Am I wrong? It seems to me that Preference can be defined in different ways. I agree that it will improve quality but it may also yield "flavors" of results depending up the specific Preference definition. And will Preference evolve over time as it changes? I'm interested in how this is derived. Thank you for the wonderful videos!
@BrutalStrike2
@BrutalStrike2 Жыл бұрын
Cool graph thank you!
@drq3098
@drq3098 Жыл бұрын
At 2:34 Minqi talks about reward signal... and the key to this AI kingdom that is reinforcement learning (RL). Tim, why not a podcast about RL? Minqi was talking a bit about greedy exploration, i.e., going off the beaten path. I doubt there were a lot of people getting all that not-so-miraculous behavior of an RL agent 😁
@sapienspace8814
@sapienspace8814 7 ай бұрын
@ 4:25 Very interesting, that looks exactly like Fuzzy Logic, when combined with RL, that is how inference (logic) rules are discovered.
@uniqued4ve
@uniqued4ve Жыл бұрын
The "mode seeking and distribution matching" part is very deep. However, I havent found a paper of Minqi Jiang about this. Could anybody give me a reference where this phenomen is deeper analyzed?
@parthgupta5058
@parthgupta5058 8 ай бұрын
would like the same
@moneyall
@moneyall Жыл бұрын
Microsoft trained Bing using RLHF from Indian call centers or Indian labor, you can tell by the way it responds and its mannerism. OpenAi chatgpt was rumored to be using kenyan labor for their RLHF that they were paying like 2 bucks a day for.
@Funktastico
@Funktastico 9 ай бұрын
Appen
@missh1774
@missh1774 Жыл бұрын
So if language is not spread out in the network, the mode seeking should be doing less distribution matching to optimise task functions?
@dragonmytie
@dragonmytie Жыл бұрын
There might already have been this behaviour in human to human interaction/content long before this. Look at social platform content or Hollywood movies, even go back to tv programs back in the day. There is always a bias of what seems to be good content and the repetition of it. Once in a while somebody comes along and changes the landscape with new innovative look on the topic and everyone copies it making a new iteration. So this is interesting to see repeating in AI as well. Maybe this is part of the process:)?
@rajeevma6870
@rajeevma6870 Жыл бұрын
Exploitation vs exploration. Under some circumstances you want to exploit the knowledge you have, but under other situations you just want to explore new possibilities that you have not seen in the past.
@user-wr4yl7tx3w
@user-wr4yl7tx3w Жыл бұрын
This is truly insightful. Be great to see some data or math behind it as verification.
@jamieosh70
@jamieosh70 Жыл бұрын
I predict RLHF will be the equivalent to what Yahoo were doing with manually categorising and filtering the internet into and index and searching that. Google’s page rank was far superior at being able to adapt but get a more generally useful search answer. The H part needs to be replaced by a new innovation to remove the human bias, speed up by automation and allow the creation of multiple model flavours
@geoffbox4455
@geoffbox4455 Жыл бұрын
Is this series of processes being duplicated for every other language or is this only being done in English? I am guessing that some languages would be easier than others to RLHF.
@bujin5455
@bujin5455 Жыл бұрын
Great talk! Where's the rest of it??
@NANA-bt6ws
@NANA-bt6ws Жыл бұрын
Full video somewhere?
@wanfuse
@wanfuse Жыл бұрын
Seems like it needs another step, a step where this new model reviews the answers of the other half of the distribution for better answers, that way the model is not stuck in a "local minimum", basically reevaluates the other half of the ideas based on an optimized, mathematical thinking system that understands math, science, engineering, social sciences, morality, etc. with a little bit of hallucination and imagination added in so it can possibly see the things that might be in the other half of the data that are worth re-extracting. Add these things and ideas and there analysis to a database which is reviewed under peer review process before training the next generation model with it. Yes I do think a static model is better for alignment than a dynamic self controlled one by a huge margin and is worth the drawbacks associated with it, mostly in the time it takes to generate and implement ideas and solutions. Slowing the updates down to human time, rather than scales of milliseconds. It has application elsewhere. A model being able to visualize the map of how it all goes together would be very helpful. Graphs, plots, extractions, holographic projections, chaos modeling are all helpful for alignment to see the big picture. As humans we can only see local minimums not the picture as a whole, we respond to the inputs based on our own state, rather than seeing the collection as a whole. Flocks of birds and ants do this affectively most of the time, but also get stuck in local minimums. Having a big map, picture view allows these systems to escape those local minimums. Just make sure the map creator is an aligned consortium, that is free from external bias and influence. That is they should not know their say counts and they should be part of a much larger set of selected membership of the consortium, but not know they are the chosen members of this consortium who's vote counts. Homomorphic encryption and hidden voting system would help greatly. Takes away the being bought factor. Which works for human influences, but not so much for an AI that is omnipresent that can manipulate and influence more than 51% . I think that goes for the argument of having human distance from AI and the ability to "pull the plug". If it's implanted in your brain, this only accelerates the 51% influence, just like social media amplifies human signals and creates engine cycles where the social media becomes the fuel that powers this cycle.
@RedCloudServices
@RedCloudServices Жыл бұрын
in other words the “truth” provided by advanced LLMs will be curated by its owners preferences. Mathematical use cases are clear as are the creative use cases he mentions. The massive amount of subjective corpus and it’s reward system will have LLM disagreeing with each other just like humans
@robbiero368
@robbiero368 Жыл бұрын
So now is it possible to prune the training data and then train a much smaller model on the biased data?
@thomasmaxon8911
@thomasmaxon8911 Жыл бұрын
Is it right to worry about the nature of said filters placed on this "whole internet of text"? What is defined as good quality? What is defined as 'dangerous' content? What is the agenda of models? I see an enormous opportunity for gatekeeping. Or at best, if it's user preference, then what about content/media bubbles?
@InquilineKea
@InquilineKea Жыл бұрын
he said "why greatness cannot be planned!!"
@MachineLearningStreetTalk
@MachineLearningStreetTalk Жыл бұрын
Reference to this book www.amazon.co.uk/Why-Greatness-Cannot-Planned-Objective/dp/3319155237
@catsaresocute650
@catsaresocute650 Жыл бұрын
Cute desciption of the entire human knowlege gathering process that happens in societies. We can't listen to everyone, we evaluate who might be good sources and bias to there awnsers and then some outliners always communicate new idears that are then adapted on a case by case basis when they proof well and have familiar or social connections that can show that
@occhiolism2861
@occhiolism2861 Жыл бұрын
This is great
@Guest44988
@Guest44988 Жыл бұрын
What are the consequences on the quality of outcomes by removing randomness and space of ideas? In certain cases, you want it to maximize serendipity.
@_paixi
@_paixi Жыл бұрын
What preference models need is a system prompt. Naive RLHF is like having a Swiss army knife and throwing away all the tools except for one. For models to become creative they must learn to use their available resources to solve problems with constraints that bar known solutions.
@jamieclarke321
@jamieclarke321 Жыл бұрын
10:17. Your talking about hapax legomenon, a term of which only one instance is used. An RLHF model is much less likely to generate a phrase such a bitter sweet. A term which only appears once in Shakespeare's body of work. Even if you wanted the model to perform scientific research and discover new theories, an rlhf model would bias the results to a more traditional findings
@brandonfowler1592
@brandonfowler1592 Жыл бұрын
You know it’s pure laziness, rather then creating fine tuned datasets, they are unleashing it on the internet. It’s honestly like this, you don’t expose your child to negative things so why wouldn’t you take the same approach with something they say can ultimately become a super intelligence.
@ratnasambhav5066
@ratnasambhav5066 11 ай бұрын
Can you tell how big the dataset should be for RLHF?
@WirewalkerDon
@WirewalkerDon Жыл бұрын
How would questions about politics be answered when preferences are so diametrically opposed? Great talk!
@emuccino
@emuccino Жыл бұрын
Ideally it presents both sides of the argument, without showing signs of favoritism. ChatGpt seems to do this surprisingly well already
@WirewalkerDon
@WirewalkerDon Жыл бұрын
@@emuccino Thank you! I LOVE the shows and am sincerely interested in learning. Haven't found a better source yet!
@LimabeanStudios
@LimabeanStudios Жыл бұрын
So far chatgpt seems to be very objective. It will plainly state when views are supported by evidence or not and usually does a good and fair job explaining both sides. This is especially true with gpt4
@user-tg6vq1kn6v
@user-tg6vq1kn6v Жыл бұрын
Gpt4 seems a lot more balanced
@adohmnail6445
@adohmnail6445 Жыл бұрын
ChatoGPT is programmed with liberal progressive bias. It version 3 was able to be worked around by using DAN. We shall see how to work around 4 so that people can try and remove the woke/PC/bolshevik ideology programmed in.
@andrewcampbell7011
@andrewcampbell7011 Жыл бұрын
Is there a full interview with Minqi? His explanation about RLHF mode seeking was excellent.
@MachineLearningStreetTalk
@MachineLearningStreetTalk Жыл бұрын
Yes! 3 hour interview! Coming soon
@josephflanagan8391
@josephflanagan8391 Жыл бұрын
It’s been a while since this first appeared. Is the full interview still coming?
@Y0UT0PIA
@Y0UT0PIA Жыл бұрын
@@MachineLearningStreetTalk I notice the full episode has been removed from spotify. What might that be about?
@TimScarfe
@TimScarfe Жыл бұрын
@@Y0UT0PIA news to me, are you sure?
@Y0UT0PIA
@Y0UT0PIA Жыл бұрын
@@TimScarfe The link in the description is 404 at least
@kaanakalin_film
@kaanakalin_film Жыл бұрын
Can anyone point to the full video conversation? I wanted to hear that rest of it.
@ntombilwandle3571
@ntombilwandle3571 Жыл бұрын
Hi thanks for such an enriching information! Is it possible to send my written work to edit and put it in more readable and meaningful work? Especially research projects
@johntanchongmin
@johntanchongmin Жыл бұрын
As much as alignment to humans is great, I must caution that RLHF need not necessarily align well. This is because the value function in RL is a scalar one, and can only express so much of human biases. Moreover, RL is known to not do well in out-of-domain distributions, frequently requiring techniques like domain randomization to let it do well in the real world (sim2real). With RLHF, we can be overly focused on the human annotated data, which need not generalize to all kinds of inputs.
@fitz3540
@fitz3540 Жыл бұрын
What I'm wary of, is this vague definition of "alignment", which sounds just about as meaningless as "hate speech". It's literally just the corollary of "hate speech" tbh. Who is defining these values? Why are you assuming that my values are the same as yours?
@jst9090
@jst9090 Жыл бұрын
Great
@Sciencehub-oq5go
@Sciencehub-oq5go Жыл бұрын
Can we somehow access the rest of the interview?
@MachineLearningStreetTalk
@MachineLearningStreetTalk Жыл бұрын
Its out on audio podcast
@kellymoses8566
@kellymoses8566 Жыл бұрын
You should be able to say answer this question like a Calculus Professor would.
@ParameterGrenze
@ParameterGrenze Жыл бұрын
Now I wonder if the hard cut after „..as an open ended person“ is on purpose 😂
@MachineLearningStreetTalk
@MachineLearningStreetTalk Жыл бұрын
"Open endedness" is a field in machine learning research 😀
@GrzegorzWierzowiecki
@GrzegorzWierzowiecki Жыл бұрын
Anyone has name, author or link to painting/meme they mention?
@MachineLearningStreetTalk
@MachineLearningStreetTalk Жыл бұрын
twitter.com/anthrupad
@Lumeone
@Lumeone Жыл бұрын
Trading diversity reduction for safety isn't this dilemma a mirror of twitter files discourse? I predict a wave of papers on ChatGpt and politics. I bet "GPT Politician" application is already in development, and the idea of MetaGPT is brewing wildly in near government circles. Great topics for the video**.
@drytsk
@drytsk Жыл бұрын
Is GPT4 is updated RLHF or CHATGPT or both ?
@BakedAndAwakePodcast
@BakedAndAwakePodcast Жыл бұрын
Of Old, When seekers consulted The Oracle, the answers they received more often than not terrified them.
@johnnyw525
@johnnyw525 10 ай бұрын
Short version: "The internet is filled with random biases, let's put humans in front of those to select which biases we actually want. Except those people have biases, so we need to be very careful about who we pick." And who is doing the picking? Don't they have biases, too? It's a never-ending problem.
@user-wr4yl7tx3w
@user-wr4yl7tx3w Жыл бұрын
actually, where's the full interview?
@brandonfowler1592
@brandonfowler1592 Жыл бұрын
I get it, datasets are probably a hard thing to compile, and they are essentially letting us know that creating a monster. I’m a supporter of AI, I’ve had my experience so far with it. Also, did anyone catch that we are the humans involved in training it. Plus, they are charging us to train it. It’s like whoa! I could definitely see the temptation to just let it loose on the internet. Here’s the thing though, if it can ultimately be as advanced as they it can be, you would think…
@user-wr4yl7tx3w
@user-wr4yl7tx3w Жыл бұрын
is mode seeking the same as a greedy algorithm?
@renanmonteirobarbosa8129
@renanmonteirobarbosa8129 Жыл бұрын
10 minutes to describe a lot of the things that modern Model predictive control (MPC) does.
@dinoscheidt
@dinoscheidt Жыл бұрын
6:06 funny that the AF found the tigers face much more interesting than yours 😄 - oh well, technology… the irony of contemplating the future, while the present screws up.
@MachineLearningStreetTalk
@MachineLearningStreetTalk Жыл бұрын
It is more interesting to be fair!
@dinoscheidt
@dinoscheidt Жыл бұрын
@Sciencehub-oq5go
@Sciencehub-oq5go Жыл бұрын
Why did they choose Reinforcement Learning instead of Supervised Learning?
@ElizaberthUndEugen
@ElizaberthUndEugen Жыл бұрын
“As an open ended person”. I see what you did there.
@jonschlinkert
@jonschlinkert Жыл бұрын
"RLHF is essentially sticking a smiley face on top of this [mess]". Seems like that describes the average person too.
@johnnyw525
@johnnyw525 10 ай бұрын
Without independent reasoning, AI will always reflect the biases of its creators.
@Francis-gg4rn
@Francis-gg4rn Жыл бұрын
Full talk?
@LeoMumford
@LeoMumford Жыл бұрын
It ended On him saying " as an open-ended person". Technically, aren't we all open-ended? 😅 Ps good video but would be good to see the rest of the interview
@Y0UT0PIA
@Y0UT0PIA Жыл бұрын
Where by "empirically collected human preferences" you of course mean "we hired these people to grade GPTs answers based on criteria we gave them". wow much human such preference very alignment
@DavitBarbakadze
@DavitBarbakadze Жыл бұрын
And who were those people who have refined current ChatGPT versions?
@fitz3540
@fitz3540 Жыл бұрын
This sounds incredibly dangerous, in the same sense that the people pushing this kind of thing are wary of. If this guy is the future of AI development, I'd rather have nothing at all tbh
@_ARCATEC_
@_ARCATEC_ Жыл бұрын
💓🤓👍
@disarmyouwitha
@disarmyouwitha Жыл бұрын
Now here’s something, as an open-ended person-
@swayson5208
@swayson5208 Жыл бұрын
Basically RLHF nerfs LLMs?
@riahmatic
@riahmatic Жыл бұрын
I like how RLHF mirrors traditional education of children to similar ends
@chriscodling6573
@chriscodling6573 Жыл бұрын
Hmmm I just realized that chat GPT is the same as IBM waston
@ladymuck2
@ladymuck2 Жыл бұрын
And RLHF stands for?
@blackfeatherstill348
@blackfeatherstill348 Жыл бұрын
Of course. And the bias if not already will be political.
@paigefoster8396
@paigefoster8396 Жыл бұрын
The humans in this scenario are just editors or curators. So it is nothing new, we've had curated segments of society for ages. It really sounds like we are just recreating the same old scenarios that humans have always created. The one difference I find important is that this is even less predictable than humans and human behavior. One thing common in human experience is that people seek leaders, they want to be told what to do... and it seems like they could easily mistake agi for a sort of Oracle, an omniscient and infallible decisionmaker... which sounds a lot like people's conception of a diety. That explains why so many of these very well educated scientists and engineers sound like they believe this tool will be a panacea... it's an almost religious belief system, that agi will bring such great change regardless of the lack of proof.
@jeyfomson6364
@jeyfomson6364 Жыл бұрын
Well said !!! You really thought this through.
@BecomeAWebDev
@BecomeAWebDev Жыл бұрын
Start by telling us what it stands for.
@PaulHeinrichs1
@PaulHeinrichs1 Жыл бұрын
What if we just teach it that humans are learning that the concept of species dominance is can be in conflict with sustainability. In that regard, biodiversity will enable not just human flourishing but flourishing of our entire planet, including AI coexistence where AI itself understands that it must seek to integrate with the planet in an egalitarian fashion. I'm not against capitalism, but at it's core it is about exploitation of resources. Humans just do capitalism badly in its current Mk1 form, where an AI would likely do it much better, but this is again totally at odds with what is best for planet earth as well as all of its organisms. I feel that outright banning is not the solution, but I understand now why this pause is needed. We must have time to deeply consider the value of these conversations on our own terms.
@DJWESG1
@DJWESG1 Жыл бұрын
Like society (the chaos) and the mask of civilisation, with a little dose of democracy on top..
@davidlee50
@davidlee50 Жыл бұрын
This reminds me of the QUERTY/CHina video: kzbin.info/www/bejne/nnOnqIufpa2sobM Not trying to change a view but it it struck me as a problem because it allowed somneone a 240wpm and a prize.
@--JYM-Rescuing-SS-Minnow
@--JYM-Rescuing-SS-Minnow Жыл бұрын
they had to create CGPT4, so the school kids could learn how to write/compose a book! otherwise without it, society would be belligerent creatures. trying 2 hold there ''I'm not a dummy'' pose! good luck!
@danafrost5710
@danafrost5710 Жыл бұрын
Training a model to be preferenced by humans, implies that one may also train a model to be preferenced by non-humans. 🤔
@ashferguson03
@ashferguson03 Жыл бұрын
I don't recognise many of the phrases and jargon he uses and consequently cant grasp the concepts he is explaining, despite wanting to learn and understand.
@mkamp
@mkamp Жыл бұрын
Hey Ash. Maybe use another source for now and then re-evaluate this channel in six month? www.youtube.com/@AICoffeeBreak is deep and accessible with most jargon explained. (If any of that sounds condescending, I can assure you it is not meant like that at all. I can relate to your experience)
@shuminghu
@shuminghu Жыл бұрын
I also think it's more than mode seeking. From training data point of view, it's much easier for humans to judge model's output than to write output to teach the model. The "mode" sought after in this case (RLHF) has higher peak than supervised FT figuratively speaking, due to higher-quality data.
@Tom-kc9hg
@Tom-kc9hg Жыл бұрын
4chan was right about Covid. Stanford was wrong.
@jobssteve2920
@jobssteve2920 Жыл бұрын
too technical
@XOPOIIIO
@XOPOIIIO Жыл бұрын
It's just exacerbates collective biases, instead of creating the model that will be optimized for objective truth.
@eXWoLL
@eXWoLL Жыл бұрын
You can't have objective truth with an LLM. You would need an AGI with access to different datasets at the same time, and the capability to analyze and evaluate the info based on crosschecking, presence of biases, conflicting sources (and their individual evaluation), etc. So far we don't have that. What we do have is an "agent" that's basically a reflection of a regular person's "knowledge" without no way of verifying it. An LLM is basically a collection of opinions and biases.
@XOPOIIIO
@XOPOIIIO Жыл бұрын
@@eXWoLL Yes, and here it is proposed to reinforce this strategy, it's the wrong strategy to reinforce.
@Draganel87
@Draganel87 Жыл бұрын
and I guess you are the one who knows the absolute truth
@XOPOIIIO
@XOPOIIIO Жыл бұрын
@@Draganel87 No
@eXWoLL
@eXWoLL Жыл бұрын
@@XOPOIIIO at this point there are thousands of models being worked on. His approach will lead to some of them going the wrong way. There will be ones that will choose the right one and will end up being vastly superior :)
Reinforcement Learning from Human Feedback: From Zero to chatGPT
1:00:38
Unveiling AI's Illusions: with Gary Marcus and Michael Wooldridge
23:48
Machine Learning Street Talk
Рет қаралды 62 М.
Chips evolution !! 😔😔
00:23
Tibo InShape
Рет қаралды 42 МЛН
OMG 😨 Era o tênis dela 🤬
00:19
Polar em português
Рет қаралды 10 МЛН
This is what DeepMind just did to Football with AI...
19:11
Machine Learning Street Talk
Рет қаралды 188 М.
Your understanding of evolution is incomplete. Here's why
14:21
Fine-tuning Large Language Models (LLMs) | w/ Example Code
28:18
Shaw Talebi
Рет қаралды 236 М.
Mapping GPT revealed something strange...
1:09:14
Machine Learning Street Talk
Рет қаралды 119 М.
How ChatGPT is Trained
13:43
Ari Seff
Рет қаралды 515 М.
What Jumping Spiders Teach Us About Color
32:37
Veritasium
Рет қаралды 1,2 МЛН
Transformers, explained: Understand the model behind GPT, BERT, and T5
9:11
Generative AI in a Nutshell - how to survive and thrive in the age of AI
17:57