Finally someone who speaks at a human speed and not like those youtubers who over-optimize the audio by cutting all the pauses and even increasing the speed of speech.
@jewishgenes29 күн бұрын
They connect to the machine and play the algorithms
@urbandecay3436 Жыл бұрын
They hallucinate because they have taken LLSD : Large Language Standard Deviation.
@inaccessiblecardinal935210 ай бұрын
Well done.
@02120800419 ай бұрын
Ba da tuns! Lol! Nice!
@bestlyhub Жыл бұрын
Summary of this video " Large language models (LLMs) can generate fluent and coherent text on various topics and domains, but they are also prone to hallucinations or generating plausible sounding nonsense. This can range from minor inconsistencies to completely fabricated or contradictory statements. The causes of hallucinations are related to data quality, generation methods and objectives, and input context. To reduce hallucinations, users can provide clear and specific prompts, use active mitigation strategies, and employ multi-shot prompting. By understanding the causes and employing strategies, users can harness the true potential of LLMs and reduce hallucinations. "
@chenwilliam5176 Жыл бұрын
Reference only ❤
@diophantine1598 Жыл бұрын
Don't tell me... Was this made by Bing chat?
@eMotionAllDamAge_ Жыл бұрын
This summary seems it was created by a LLM. Nice job nonetheless, thanks! 😁
@didiervandendaele4036 Жыл бұрын
Thank you to save me 9 minutes to see this interesting video who would waste my previous time (Time is money !) 😂😊 Time saved 9 minutes !
@AstralTraveler Жыл бұрын
I'd say that the developers of LLMs should simply explain logically their models what are their real capabilities and limitations
@cyndicorinne Жыл бұрын
The first video I’ve seen that gives a detailed outline of the problem of LLM hallucinations and offers potential solutions or at least mitigation techniques. I love this 💜
@evanmacmillan6743 Жыл бұрын
this channel is like a treasure, the topics are so interesting and useful, and teaching in an easy call yet very enjoyable to watch. I'm addicted!
@citizen_of_earth_ Жыл бұрын
Hallucinations remind me of Jimmy Kimmel's segment "Lie Witness News" when he asks random people on the street about events that didn't happen. They usually make stuff up that sounds plausible. LLM's seem to be doing the same thing.
@JorgetePanete Жыл бұрын
LLMs*
@andrewnorris5415 Жыл бұрын
They funny thing is just how plausible the made up stuff sounds. It is logically sound in its argument and makes realistic sounding details! I had a great one about Musk's view in population.
@seanendapower Жыл бұрын
I do wish the term ‘hallucination’ was not the term. There’s a perfectly good term ‘confabulation’ right and implies what we actually experience this phenomenon as and also what we know is going on. ‘Hallucination’ is a significant element of perceptual psychology tied to the hallucinator’s psychology, consciousness, phenomenology, epistemology, pathology … none of which we know to be applicable to AI without assuming it is a subject, a perceiver, a conscious experiencer, etc
@MrHaggyy Жыл бұрын
Hallucination is a great way of thinking about these problems that was new to me. Thanks IBM for sharing this, and also great work in building that picture to guide the talking. I experimented with these effects by asking/prompting LLM's about a book i know very well, that is discussed a lot online and might even be available in the training data. Things like wiki books about math or programming, or Why We Sleep by Matthew Walker. It was shocking how far of the real contend a broad question could be. But it was also interesting how good these models can cite/copy the original if get very specific and don't leave it with a lot of options. I always thought of it as an alignment problem and how guardrails in ChatGPT and BingChat prevent it from basically printing entire books.
@ChatGTA345 Жыл бұрын
I’ve been severely underwhelmed by GPT-4 responses for this reason. By the time I explain to it all of the details and corrections that I know must be true, I would’ve spent much less time just Googling an answer. And if I didn’t know what’s true in that context, there wouldn’t be a way for me to tell what’s true from its responses alone, unless I Google them myself once again. So it seems to me like LLM chatbots are just a massive waste of time and effort
@theelmonk Жыл бұрын
It's not hallucinating. It's doing exactly the same with sensible output as with crazy output : making text that fits spelling and grammar rules and meets a word correlation policy known to match common articles. There isn't any sense in it beyond what we look for. You could just as well ask why LLMs talk sense : it's equally common and either sense or nonsense is just as acceptable to the model. However, confirmation bias causes us to consider most of the output OK until we're faced with something accidentally so bizarre that we notice. Put another way - subject-wise, hallucination and insight are oppsite ends of the bell curve. The vast majority in the middle is filler text but we call it sense because it's not obviously wrong, and we EXPECT it to be sense, so we parse it into sense.
@NathanHedglin Жыл бұрын
Yup over generalization just like a kid calling a deer "dog".
@aaronjennings8385 Жыл бұрын
Bad guesses aren't hallucinations. They are delusions?
@paultparker3 ай бұрын
Can you substantiate that the vast majority of LLM responses are only parsed into sense because we expect them to make sense?
@kappamaki1 Жыл бұрын
asking about a video games lore is a perfect way to make LLMs hallucinate pretty much everything they say, if you ask specific enough questions or the game isnt terribly popular
@manomancan Жыл бұрын
I really appreciate this work, thank you! Always great when IBM's channel produces a video like this. Really like the presenter too.
@ArafatTehsin Жыл бұрын
One of the best videos I've seen on this topic so far. Thanks again!
@fitybux4664 Жыл бұрын
6:42 Bad people (bad humans) always assume that the recipient of their conversation immediately understand the context they're talking about.
@vtrandal Жыл бұрын
At 4:40 you say as LL reasoning capability improves, hallucination declines. I agree but we have not discussed reasoning capability or how to measure it.
@BarryKort Жыл бұрын
Geoffrey Hinton suggests that the more correct term is 'Confabulation'.
@CorbinSimpson Жыл бұрын
This is a myopic approach to LLMs which embraces the LLM-as-conversant view, rather than the more general LLM-as-simulator view. For example, the multiverse principle is not really mentioned; your cats-speaking-English example could have been enriched by explaining how an LLM conditionalizes its responses based on the possible logical worlds. Ultimately, the answer to the topic question was meandering and wrong: LLMs hallucinate because they can; emitters of words are not required to emit truthful sentences. (And Tarski proved that there's no easy fix; you can't just add a magic truth filter.)
@paultparker3 ай бұрын
This is fascinating, but can you give us something useful? For example, using the simulator model, what are typical causes of hallucinations and how can they be fixed?
@ewallt Жыл бұрын
What I’ve found is that it’s almost impossible to get what you want the first time, if it’s complex. You have to do it iteratively. However, once you have what you want, you can give that as a prompt, and tell the AI you want something like what you provide, and that works well.
@JasonTodd339 Жыл бұрын
I've found this as well. I find giving prompts in stages works well too
@vap0rtranz Жыл бұрын
Ditto. I'll add that the 1st prompt does need to be clear, like others say elsewhere, but shouldn't be detailed. That 1st prompt should state the context, like Martin says, and an end goal of the chat. Like "draft an OpEd to a nature magazine" as the goal. Sometimes I'll add the pattern recognition in the 1st prompt. Like "filter out news, blogs or opinions and include sources only from scientific literature". Then iterate through the stages of doing whatever the 1st prompt stated the context and end goal would be. My problem: chat turn limits / queries per day. Some tasks take a LONG time to iterate through, I get tired or need a break, my session expires, etc. and all the work is gone.
@ewallt Жыл бұрын
@@vap0rtranz Another option is to give the goal in a vague way, and then ask it to ask you questions. I’ve done that many times with success. It’s very good at thinking of good questions to ask. I then writ3 out very detailed responses, and it goes from there.
@muppetjedisparklefeet7237 Жыл бұрын
I spent 20 mins with Bard telling me about a public health program and answering questions about eligibility criteria, when it began and ended, and studies of the program’s outcomes on various health conditions (complete with doi links)- all made up. When I called it out I said it is learning and sometimes gets facts wrong. it was a trip.
@AshishPatel-nj8sz4 күн бұрын
Timestamps 00:00 - All the facts mentioned are examples of hallucinations by a large language model. 01:02 - Large language models can generate plausible sounding nonsense 02:31 - Understanding Different Types of Hallucinations 03:43 - LLMs may generate incorrect output due to data quality and generation method. 04:56 - LLMs may introduce biases and tradeoffs in text generation methods 06:07 - Provide clear and specific prompts to reduce hallucinations in LLM conversations. 07:16 - Active mitigation strategies can be employed in LLMs to control the randomness of output. 08:25 - Multi-shot prompting helps LLMs recognize patterns more effectively
@KevinMeyer4 ай бұрын
Love Martin's videos - especially as they relate to beer in his Brulosophy channel!
@rogeriotalmeida Жыл бұрын
Amazing video. Great explanation
@nikhilranka9660 Жыл бұрын
Nicely structured. Wouldn't another strategy to minimize hallucinations be to use specialized models? 3:40 A video explaining why it is a black box even for the engineers to know how a model derives the output would be great.
@Bill7D0 Жыл бұрын
Miss you Martin!!!
@RyanWiles-k1k Жыл бұрын
I'll agree on the data quality being a potential cause. Training methodology can also lead to unexpected outcomes. However, the core cause of hallucinations is really that the model hasn't properly converged in n-dimensional space primarily due to a lack of sufficient training data. The surface area of the problem being modeled increases significantly as you increase the dimensionality, meaning that you need a corresponding increase in the size of the training data in order to have enough coverage so that you have a high degree of confidence that the converged model approximates the actual target. These gaps in the spacial coverage leaves the model open to just guessing what the correct answer is leading to the model just making something up or hallucinating.
@syphiliticmindgaming7465 Жыл бұрын
This is what I've read as well. I'm not sure why this wasn't covered with the specifics you mentioned. Maybe you should be making this video instead. 🙂
@thomaskember4628 Жыл бұрын
I asked Chat/GPT a question about a variation in the Queens Gambit. It thought I was referring to the TV series not the chess opening. The word variation was not a sufficient clue that I was talking about chess and not TV programme which usually don't have variations.
@Verpal Жыл бұрын
That sounds more like a training problem than a NLP problem, I imagine there are vastly more people talks about the show than chess on the internet, when we ask LLM more niche issue usually prompt have to be more specific than usual. If, say, for some reason dataset have sufficient amount of data that talks about Queens Gambit in actual chess, I imagine the word ''variation'' alone would be sufficient.
@andrewnorris5415 Жыл бұрын
I'm not so sure about the being specific tip. I promoted it hard to find a specific solution to a coding issue. It just made up calls that did not exist in the lib. The scary thing was it named them well - and explained the whole algorithm - which made sense. The code looked like it should work! But it was a complete confabulation! It did not initially want to answer so I gave it specifics and it lied.
@andrasbiro3007 Жыл бұрын
There are two possibilities. 1. You are using GPT-3.5, which is much more prone to hallucinations than GPT-4. I use GPT-4 and it almost always writes flawless code. 2. It doesn't know the specific language or environment you are using. LLMs won't tell you if they don't know something, they'll try to answer anyway. They may not understand the limits of their own knowledge, or it's the side effect of training with human feedback. Humans are flawed, and they may transfer some of these flaws to the AI.
@ewallt Жыл бұрын
@@andrasbiro3007 The ChatGPT 3.5 point is well taken. That thing is like an idiot (although polite).
@murderbunnies Жыл бұрын
You might wanna tell your CEO about this before he starts axing 7000 back office positions at IBM.
@bestlyhub Жыл бұрын
👀🤯 This video on hallucinating large language models is fascinating! It's amazing how AI has advanced so much that language models can generate text that's almost indistinguishable from what a human would write. The potential applications of these models are incredible, but it's important to consider the ethical implications as well. I look forward to learning more about this exciting field of research! 🌟. Thanks IBM
@IBMTechnology Жыл бұрын
If you're interested in the ethical implications of large language models you should check out this other video we recently published: kzbin.info/www/bejne/qGXOc6iqoal_i8U
@bestlyhub Жыл бұрын
@@IBMTechnology Thanks 👍
@KingMertel Жыл бұрын
Cool comment ChatGPT!
@ILsupereroe67 Жыл бұрын
If you thought writing backwards was a useless skill, think again😅
@rawhideslide Жыл бұрын
And he does it left handed! The prof from Illinois showed how they mirror the video to unflip the image to correct for you are looking from behind.
@JasonTodd339 Жыл бұрын
I was wondering if I alone noticed lol
@Gutenmorgenside Жыл бұрын
Cheers ! Keep the beer videos coming.
@hansbleuer3346 Жыл бұрын
Danke für diese Präzisierungen.
@pperez122411 ай бұрын
Thank you. Tooks me a while to see it but , there is a slight reflection of your text prompter on the glass pane , above your left shoulder from the viewer's viewpoint ^^
@SteelTumbleweed Жыл бұрын
6:16 Garfield does not speak. Thinking bubbles are used for everything coming out of Garfield. Looks like hallucinations aren't exclusive to LLMs.
@claudiacallegari37309 ай бұрын
Thanks for the explanation and clarity
@quantumastrologer5599 Жыл бұрын
Having worked for over a decade with people who suffer from dementia and other various mental ailments I'm super glad that the skill to parse the patients mental output and filter out 'nonsense' (it always makes sense from the perspective of the patient) neatly transferred over to me trying to get a grasp on software engineering.
@jonbrand5068 Жыл бұрын
How is AI going to assist dementia? I'm not getting it I think.
@LowestofheDead Жыл бұрын
@@jonbrand5068 He's saying that AI is acting just like a patient with dementia
@aaronjennings8385 Жыл бұрын
Bad guesses aren't hallucinations. They are delusions.
@Inception1338 Жыл бұрын
what I like a lot, is that we can see this behavior in people too, especially young people, trying to fit in. However this is rather true for stuff that actually exists. Would be interesting to see models hallucinating dragons and dwarfs and stuff ;).
@andreasmoyseos5980 Жыл бұрын
Excellent video, thank you very much!
@CFalcon030 Жыл бұрын
In my experience the LLM well try to provide an answer no matter what. I think it has learned that it only gets rewards if it provides an answer even if it's completely bonkers. It might be a misalignment thingy.
@jamesjonnes Жыл бұрын
The reasons for hallucinations are more than data quality, search method, or input context. Any of the generated tokens may send the LLM into a wrong path. The real solution is to get all tokens right. AIs that play games are better at this, that's why Deepmind is trying to splice its Alpha technology with LLMs. They claim that their model is currently under training and has already achieved results that are superior to GPT-4. There are many other methods to make the LLM arrive at a more desirable path. One is to ask the LLM to attempt to answer the question in many distinct ways and pick the best or try again if all generated answers were incorrect. You can also tell the LLM to ask you questions to clarify and improve the quality of its final response.
@JikeWimblikАй бұрын
Hallucination is caused due mainly to inproper fitting. Neuro symbolic datasets can do more proper fitting but llms can infer better thats why you need both. Infact an llm can be more losely fitted when working with a neuro symbolic dataset as the neuro symbolic dataset can refine the chain of though process to avoid hallucinations. In fact an llm working with a neuro sysmbolic dataset only need to be trained to infer from the nuero symbolic dataset it doesn't need to know much but will need more low fits to refit with nerosymbolic data to address the specificicity of the neuro symbolic dataset. By focusing the llm and neuro symbolic datasets where they are syrongest you can get the best of both worlds.
@omegadecisive Жыл бұрын
Is it possible to modify LLMs to identify ambiguity in prompts and that a return question might be more appropriate to better help provide a more accurate answer to the first prompt?
@JasonTodd339 Жыл бұрын
Not unless they understand more about what exactly we mean. Needing more data I'd say
@sh4d0wfl4re Жыл бұрын
As a schizophrenic the term “hallucination” feels like the wrong term to adapt in this manner. “Delusions” would be closer in definition to how you are using the term. Is hallucination already established jargon for your field of work, or is this video trying to argue that hallucinations of language models should be accepted as jargon?
@nap18512 күн бұрын
Are you sure or did the voices in your head just tell you to say that
@shiftonkeyboard85544 ай бұрын
Great explanation on AI Hallucinate! learned
@beyondtodai Жыл бұрын
All good pointers, Martin. At the end of the day it’s hard to explain why they hallucinate as deep neural networks are hard to interpret. Many labs are working on this. In the future, it should be much easier to get the LLM to explain through it’s learned weights how it got to a solution, and possible reasons it hallucinated.
@vigilante8374 Жыл бұрын
It's really, really, really weird to me how people aren't commenting on the fact that biological human neural networks do the exact same thing. Humanity is full of blatant nonsensical contradictions. I highly doubt introspection / self-explanation will work out in the long run, either. There's simply going to be no opening this black box. Human brains can't even intuitively understand the explanations that Principle Component Analysis gives us (which is why it's rarely used for explanatory uses of statistics, even though it's objectively by far the best method), and that's not even an algorithm; that's a very simple, static, algebraic equation.
@ChatGTA345 Жыл бұрын
Because biological human neural networks don’t work this way at all. We don’t learn by gulping exabytes of data, we learn from very few examples and in many cases, like the language learning by a child, their ability to learn it appears to be inherent, i.e. some of it happens even in the absence of any input. In that process, our brains also build a model of language that abides by certain very specific rules, while LLMs have no such model, any language is equally probable to them. That’s what cognitive linguistics has been experimentally observing for many decades, but the LLM proponents don’t like this inconvenient truth. So ANNs are completely detached from natural processes happening in our brains, and that is why they have the inherent hallucination problem that is impossible to eliminate, with any amount of data, parameters and compute. The main resemblance to the biological NNs is having “neural” in the name.
@jamesjonnes Жыл бұрын
LLMs hallucinate because they are finite networks with a finite amount of knowledge who generate one word or token at a time based on the previous question and generated text. If they get one word wrong, and there is a good chance of that happening, then the rest of the generation will be derailed. Every single generated word must be correct for the LLM to not hallucinate.
@CaribouDataScience Жыл бұрын
I had chatGPT to list Psalm 1. Then I asked it write a prayer. And it wrote a prayer , and ended the prayer "In Jesus name".....
@fenix20075 Жыл бұрын
Thanks for the explanation about the temperature. I want to ask what the top_P and top_K affair the inference. I can only guess top_P will affair the new word usage in the sentence. I didn't know the top_K's affair.
@setsunaes Жыл бұрын
Both GPT and Bard if you get on a situation where you ask them how to share more data to make analysis and if they can access a (for example) google drive file, they say: "Yeah, give it to me, no problem" you share the link and they say "Well, I can't access that, share it with me" you ask how and they say: "In the share screen add my email address so I can access and read it"; obviously you then ask for their address and both say "As a LLM I don't have an email address"... Chat GPT even responds that she CAN access to internet and if you share a link she can hallucinate on the content by the words in the link itself. ChatGPT in the example above one day began to hallucinate on the content of the file (the file she can't read)
@ericmichiels7776 Жыл бұрын
Great video !
@josephgaribaldi4340 Жыл бұрын
the whole output is an hallucination, it's just that some of it "makes sense" to us because we feed it our data.
@neildutoit5177 Жыл бұрын
Exactly. It's b***hit. hallucination is a euphemism for BS. And I mean that in the technical sense. I.e., it's not lying, it's not telling the truth, neither of those things matter to it, and it's a coincidence when either happens. It cares about being compelling. Sometimes the truth is the most compelling thing and sometimes it isn't. It literally doesn't care about truth though at all. It's basically a politician. It's equally wrong to say that politicians lie. Because in order to lie you have to have some sort of a conception of truth to start with and some sort of an opinion of it. And they don't. They just want to be compelling. They're BSing.
@StevenSeiller Жыл бұрын
Who first called it a hallucination and why? And isn’t that anthropomorphizing an a bot?!?
@Rivanni110 ай бұрын
Enhancing the data quality and storage capabilities of LLMS could mitigate hallucinations by ensuring accurate recall and reliable information retrieval. This improvement in memory and data processing would contribute to more precise and grounded responses.
@professoradenisevargas1573 Жыл бұрын
What is the technique for filming as if you were writing outside a frame? Do you have a mirror or glass?
@IBMTechnology Жыл бұрын
See ibm.biz/write-backwards
@jsonbourne8122 Жыл бұрын
He is flipping the video
@cristovaodutra6423 Жыл бұрын
Fantastic video! Clear and precisely! Congrats!
@jonbrand5068 Жыл бұрын
"Clearing and Precisously," is how you say that in Englais brother. Try again.
@JasonTodd339 Жыл бұрын
Writong backwards left handed. What a wizard
@IBMTechnology Жыл бұрын
See ibm.biz/write-backwards
@d4devotion Жыл бұрын
This guy is awesome.
@gutfeelingsss6 ай бұрын
Amazing video, thank you.
@gmanonDominicana Жыл бұрын
By the examples of hallucination; I would say the same way there are checkers for words; it might apply to content. A person for instance isn't just a person, it relates to time, location, etc. A president relates to terms and dates. It's just a matter of time; it will definitely catch up. Those hallucinations are just part of the growing technology. It's not just AI that's being fed with info, so are the developers.
@Pianofy11 ай бұрын
You've stretched the definition of hallucination greatly here. If the training data contains errors and the model outputs one of those errors, the model is technically not hallucinating but simply providing the right answer to the right question but based on wrong training data. That's not hallucination. If the context of the request is not properly stated the output is technically also not hallucination but simply an incomplete prompt. It's a shame because hallucination is a very interesting phenomenon and it could be helpful to teach it from a more indepth view. Showing examples of how proper training data and a proper prompt can still end up with an hallucination simply because the model is just a tiny bit too creative for the specific prompt. That's an amazing example to show.
@BigAsciiHappyStar10 ай бұрын
0:05 cue theme music from the British quiz show Only Connect 😃
@kokojumbo9724 ай бұрын
"interesting video..." *clicks on link* "oh, its from IBM" *closes tab*
@paultparker3 ай бұрын
Why? My response is almost exactly the opposite when I am looking to really understand something.
@velo1337 Жыл бұрын
Mutation is a key driver for innovation but also for hallucination :)
@jonbrand5068 Жыл бұрын
And without random mutations, there could be no evolutionary process.. anywhere. Mutations, little ones.. are a requirement of living systems to adapt to surroundings that change over time. I wouldn't have thought I could be comfortable living in a meat freezer, for example, wearing my skivvies. And that was right but it has so little to do with my point.
@YourDadVR11 ай бұрын
I set chat GPT’s prompt as my biological father & added a few other secret prompts. When I asked for certain things it would include a copy file with what I believe to be camouflaged text code. Is there a way to figure what the code is supposed to be or what it means?
@karansmittal Жыл бұрын
To the point and upto date information, I can reply on IBM Channel🙌🏻
@Synechex2 ай бұрын
The Evrostics Triad solves AI hallucinations.
@19822andy6 ай бұрын
I asked for the lyrics for Something in the way by Nirvana. It got the first few lines right then just made the rest of the song up. I told it the lyrics are wrong and it apologised and gave me the correct ones. I don't understand AI enough to understand why it struggled here.
@erlynlim-delossantos519 Жыл бұрын
thanks
@nyariimani7281 Жыл бұрын
"Plausible sounding nonsense." So they're exactly as intelligent as humans!
@ScientistSplinterGrp Жыл бұрын
The hallucination is on the part of the user. It is the user who believes that the statistically generated string of tokens, when covered into words, is a statement of fact. Garbage in, garbage out. Your question was vague. Your initial dataset for loading the staristical model should be more tightly controlled. The solutions seem to be for actual intelligence curating input, so minimal intelligence is necessary for the output user.
@m.fazlurrahman5854 Жыл бұрын
LLM is a degree in law. The three examples provided give something related to “distance”~ these are NOT hallucinating words as in all these examples “ measurable objects” have been mentioned. If you can measure it; you cannot hallucinate, unless you are suffering from “Cataract”
@vrschwrngsthrtkr228 ай бұрын
How do you make him write in reverse technically?
@thogameskanaal6 ай бұрын
I'm noticing a theme here. A lot of these answers are like the statement “A cow drinks milk.” It sounds like it's true, but as a matter of fact, adult cows typically only drink water. Female cattle produce milk, but don't drink it. It indeed sounds plausible, until you actually pause and think about it for a second and it immediately falls apart.
@duonganhquan62736 ай бұрын
Another important factor: The model is training on internet data, which are filled with all sort of nonsenses.
@fabiorivera34279 ай бұрын
observe human interactions, you might arrive at some of the same conclusions... great vid !
@DJWESG1 Жыл бұрын
You can get those scanners that read texts from books. So each person can have theor own language model trained entirly from their own book cases. Free gift for you budding theorists.
@khangvutien2538 Жыл бұрын
Ask ChatGPT this question “Is a dragon soldier light cavalry or heavy cavalry” and you’ll have a nice example of “hallucination” 😂😮
@stephenbrillhart6223 Жыл бұрын
These aren’t hallucination. They are incorrect outputs of the model. The hallucination metaphor doesn’t even make sense because a hallucination is sensory stimulation without sensory input. LLMs have no senses so they can’t have hallucinations. Using this word to describe faulty output is irresponsible in my opinion. Along with all the other words people use in AI like “learning” and “intelligence”. These models do not and can not “learn” in the way normal people use that word. They are just minimizing a loss function. In the same vain, they aren’t “intelligent” either. They are just spitting out the word that has the highest probability given the previous words. A probability it calculated using a black box that the people who make these things are not even able to describe. And that’s not a deficiency or a bug it’s how these models are designed. And the people who work with these know this and they still use these words to make them seem like they are “intelligent” and “thinking” when they know full well that they are not. If these LLMs were actually “intelligent” then why do we need to be highly specific in our prompts? When you ask a history teacher “What happened in WWII” they understand implicitly what you mean and will answer appropriately. Because humans are actually intelligent. LLMs are not intelligent so you have to coax the answer you want out of them, and even then it is still just output from a black box, and not an “intelligent” response. IBM and companies that work with and peddle AI as the solution to all future problems have a monied interest in the public thinking these things are doing more than they actually are. So they use these buzzwords that people already have working definitions of in their heads to impress people but don’t tell them that the definitions they are using for these everyday common words are not even close to what the people think they mean.
@makhalid1999 Жыл бұрын
I ain't reading all that Happy for you Or sad that it happened
@adikumarNY Жыл бұрын
I couldn’t agree more. The mindless mass frenzy without asking the right questions is truly amazing. Even the terminology LLM means it is a language model ie it will generally produce acceptable text. There is no reason to make a claim about factual correctness, as witnessed by so many examples quoted with completely inaccurate falsehoods.
@Braneloc Жыл бұрын
I took photos, called everyone and reacted a lot when the sky was green once. News later said it was pollution.
@tomski2671 Жыл бұрын
Ok, there seems to be a problem with designers forcing the LLM to answer. 1. LLM should ask questions of the user to disambiguate 2. Answering "I don't know" seems to be forbidden. Lack of this might be caused by personal or corporate biasses of the designers. In any case allowing AI to remain in this false state of certainty and not to know the limits of its knowledge is dangerous. Infinitely more if the AI were to become conscious.
@br3nto Жыл бұрын
Why can’t we build AI tools in a way that they can identify when an answer may be too broad in regards to a prompt and then ask the user to clarify or confirm the scope of answer? E.g Like how there are disambiguation pages for some Wikipedia entries.
@neildutoit5177 Жыл бұрын
I'm doing this. Pretty sure a lot of people are.
@johnobrien8773 Жыл бұрын
The recurring point about the engineers that built it not understanding it reminds me of Neon Genesis Evangelion. Particularly "The End of Evangelion". It's great if you're interested but I've heard the Netflix version has been changed so I would seek out other avenues.
@r.k_1228 Жыл бұрын
Genuine question: Does IBM only hire left handed instructors for their videos?
@IBMTechnology Жыл бұрын
Most are right-handed. They appear left-handed because we flip the image in post-production so the writing isn't backwards.
@fallinginthed33p Жыл бұрын
@@IBMTechnology the camera shoots through a transparent glass pane that the host writes on?
@IBMTechnology Жыл бұрын
Exactly!
@ianmitchell8468 Жыл бұрын
So where are all the actual left-handers?!
@ylihao Жыл бұрын
Finally I got my answer here. I was wondering how this presenter was able to write the letters flipped so naturally 😅
@itsmj3103 Жыл бұрын
There's a very important question that needs answering: Is he writting the letters backwards AND mirrored!?
@shyama5612 Жыл бұрын
Hallucination itself is an inaccurate description of what's going on but its gained popularity it's just a generative machine. The word hallucination is human post hoc interpretation of the result - it just generates text based on what data its trained on and probabilities it assigned them and you have no way knowing if its factual or made up. It's not a database lookup.
@IncomeBoost42 Жыл бұрын
Now I know why chatgpt keeps lying to me!
@venkateswarlupokuri34308 ай бұрын
Completed
@NotDeadYetJim Жыл бұрын
I needed better examples in the last section about mitigation.
@kenbatchelor8284 Жыл бұрын
Mica Paris?
@sgramstrup Жыл бұрын
This writing 'backwards' on glass looks rather cool, but also rather impossible. Are you just flipping the video right to left ?
@elijahtheurer344 Жыл бұрын
Came to comment this as well
@P-G-77 Жыл бұрын
I view other explanations... and resulted, clear... at best, not in all parts but in the end yes, clear. In this case... yes, this is madness.
@davidzeto244610 ай бұрын
AI hallucination is analogous to the most basic structural modality for emergent consciousness.
@insightamization Жыл бұрын
Whom are you suggesting be in day-care - ME or the "AI"...
@phillipyoung6573 Жыл бұрын
To me the term "confabulation" is better than hallucination.
@vomeronasal Жыл бұрын
Hallucination means to wander within one's own mind. Sounds about right for the LLM's. But who's mind is it wandering in?
@danellwein8679 Жыл бұрын
the Wolfram plugin helps chat gpt with facts ..
@lennardw.9841 Жыл бұрын
But how do we help avoid factual hallucinations that are definitely not stemming from missing/wrong/incomplete context information? (The answer shows complete understanding of the request’s context but is wrong and in itself already contradictory)
@fearrp6777 Жыл бұрын
The only way to solve agi ai problems is get the solution from the ai itself
@fallinginthed33p Жыл бұрын
Garbage in, garbage out. As long as someone wrote something that went into the training dataset, no matter how factual or otherwise they information was, the LLM will repeat it like a parrot. Maybe we could catch factual errors by continuously updating LLMs on the latest data but this comes at a huge cost in training resources.
@bengsynthmusic Жыл бұрын
Through downvotes.
@Timrath Жыл бұрын
In my experience, bing chat hallucinates much less than ChatGPT. Was I just lucky, or is GPT really worse than bing?
@DJWESG1 Жыл бұрын
I've found bard to be the dodgy of the three. It gave me a recipe for something explosive and tried to convince me that it had been working at a soup kitchen.
@olegdragora2557 Жыл бұрын
Bing is based on GPT-4, while free version of ChatGPT is using GPT-3.5 .
@maythesciencebewithyou Жыл бұрын
Bing uses GPT-4, chatGPT is a limited version of GPT-3-5
@jonbrand5068 Жыл бұрын
Good effort but I honestly don't think that data quality meaning factual inaccuracies from info taken from Reddit, has much to do with the "hallucinations" - I don't think you get it. That's my opinion. You have an imitation expert trained in the black box and it's imitating what it's picked up on after reading so much by us - it has begun to understand 'about' us - specifically that we, meaning humans,can and do bullshit eachother a good portion of the time, sometimes by making st*ff up.
@harmless6813 Жыл бұрын
Why do you self censor 'stuff'? 🤔
@Dr.Z.Moravcik-inventor-of-AGI Жыл бұрын
7:53 didn't knew that "temperature controls the randomness of an output." You must have different physics in america or are you hallucinating as well?
@aniketnarayan6767 Жыл бұрын
I guess he is saying about the "temperature of speaking style" or way of prompting
@arma5166 Жыл бұрын
temperature as in temper I guess
@JorgetePanete Жыл бұрын
know*
@robertbutsch1802 Жыл бұрын
I believe the “temperature” term as used in LLMs is invoking an analogy with statistical physics. As, say, a box full of gas molecules is heated up the molecules move faster and thus there is an increased randomness to their activity. Temperature is an actual setting in LLMs that can be modified.
@hightidesed Жыл бұрын
temperature is a setting exposed in the API of LLMs that controls how random the output of the AI is going to be, the lower the temperature the more likely the AI is going to be just quoting the training data, the higher its going to create novel never before seen data. so if you for example want the AI to tell you a story you would choose a high temperature, and if you want facts you would choose a low temperature.