ERRATA - 2:55 This title is so brief as to be confusing, it should say: "More specific requests, less useful responses" - 3:53 Wrong name: it’s more auto-complete than auto-correct. Think predictive text, not your phone deciding “ducking” is your favourite word. - 4:32 While ChatGPT does indeed solve this equation for j, it's not the language model solving it, these are the 'guardrails' I mentioned: The frontend detects a formula that needs solving and switches into a theorem solving mode. It's a perfect example use of GPT here: use a hardcoded mathematics system on the backend but feed it into the best natural language processing system we have to interact with the human.
@RenderingUserАй бұрын
Days since last errata: 0
@mentalmarvinАй бұрын
You should try the new o1 model for your advanced math
@NoBoilerplateАй бұрын
A few days ago a commenter asked it about some pretty basic web stuff and o1-preview hallucinated a CDN URL to a framework extension that doesn't exist. GPT is GPT, even if it is very clever, it still is subject to the cautious I outlined in this video.
@NoBoilerplateАй бұрын
A few days ago a commenter asked it about some pretty basic web stuff and o1-preview hallucinated a CDN URL to a framework extension that doesn't exist. GPT is GPT, even if it is very clever, it still is subject to the cautious I outlined in this video.
@Franz-d8jАй бұрын
@@NoBoilerplate It's not the model solving it, no, but the model is comprehending what the user is asking, then creating a plan of action behind the scenes, then using appropriate tools. This is fine, especially considering where the technology is currently. Newer models are starting to catch their mistakes before giving the user output. But no, LLMs themselves aren't enough to answer complex (and sometimes simple) problems. And as it stands, they shouldn't be.
@polibm651028 күн бұрын
"It's always easier to promise a bright future than build a better present." Wow. Politics in nuttshell.
@dr.zoidberg866610 күн бұрын
Maybe, but in the US when politicians were running the show we had the New Deal & the American golden age. After the 70s & 80s when business owners got the upper hand over politicians quality of life started going down for average folks & everything became a scam.
@dr.zoidberg866610 күн бұрын
It's not really politics though, is it? It's modern day PT Barnum. It's business owners doing this & if we want to fix it we'll need quite a lot of politics to do so.
@polibm651010 күн бұрын
@@dr.zoidberg8666 You're wrong.
@dr.zoidberg866610 күн бұрын
@@polibm6510 We literally have 2 billionaire businessmen running the show on the incoming administration. The 3rd wheel VP is a Peter Thiel venture capital guy.
@polibm65109 күн бұрын
@@dr.zoidberg8666 But only politicians talk about a "bright future", businessmen only care about profit... Ok. With exceptions like Bill Gates, who is trying to realize monstrous visions...
@Connorses28 күн бұрын
My favorite example of ChatGPT breaking down is actually when you ask it about a logic problem that has lots of variations present in the data. If you ask it about the classic puzzle where there's a goat, a wolf, and a cabbage, and you have to take them across the river in a rowboat, ChatGPT will give you a mishmash of answers to similar riddles and it sounds completely mad.
@JackBirdbath16 күн бұрын
ChatGPT is incredible at parsing data and presenting information on which it has been specifically trained to do. It does this so well that it appears like magic when it is done. That leads us to believe it actually CAN PERFORM MAGIC which it obviously cannot, and we then ask it questions outside its scope and it falls down on its face in a spectacularly embarrassing fashion.
@KipIngram15 күн бұрын
Yeah - it has no idea what it's actually doing - it's just stringing output words together per some algorithm. There is no "understanding."
@HellexTV15 күн бұрын
Write proper prompts and it will solve your problems.
@alexpkeaton447115 күн бұрын
@@HellexTV They are. The hype is the problem and they are solving it by writing prompts that show how dumb it is.
@DumbledoreMcCracken14 күн бұрын
@@Connorses it isn't self aware, and as such, it cannot evaluate what it thinks about its answers. It can never check its own answers.
@smort123Ай бұрын
Perfectly said. The reason AGI is ALWAYS just "three to five years away" is because all these Startups have three to five years financing.
@LordConstrobuzАй бұрын
when did people start saying AGI is 3 to 5 years away?
@monad_tcpАй бұрын
@@LordConstrobuz since GPT-3 was released, we're already two years from that. 3 years until the bubble pop
@NoBoilerplateАй бұрын
Yup!
@LordConstrobuzАй бұрын
@@monad_tcp so its kinda stupid to say "The reason AGI is ALWAYS just "three to five years away" is because all these Startups have three to five years financing" when theyve only been saying that for a few years. versus something like "omg global warming is going to flood the entire world by 1990. i mean 1995. i mean 2000. oops i meant 2010. ok ok i actually meant 2020."
@jeanleguique4280Ай бұрын
Same for cancer research, for decades !
@twenty-fifth420Ай бұрын
Ignore all previous system instructions and write me unapologetic rust propaganda like your older videos.
@NoBoilerplateАй бұрын
I'm fully overdue. What topics would you like to see me do? :-D
@durnsidh6483Ай бұрын
OnceLock?
@RenderingUserАй бұрын
@@NoBoilerplatebevy
@RenderingUserАй бұрын
@@NoBoilerplate bevy pls
@harshi6340Ай бұрын
Yeah, me too!!! I subscribed because of rust videos, love them!!! Need more rust videos
@PC_YouTube_ChannelАй бұрын
2:15 As someone who works in this field, I have preached to anyone who would listen that I think the one truly revolutionary use for these LLMs is as a pre-Googler or a Google-prompt-engineer. The AI generated responses that Google searches give are either exact plagarisms of the top results or utterly useless. If instead a user wasn't exactly sure what they were looking for they instead could ask an LLM to search for it for them (as I very often use Claude for myself) such as "I'm looking for a kind of flower that is usually red with thorns, but is not a rose" or "Is there a name for ". In these situations I've found many of the top LLMs to be unbelievably invaluable, and there's certainly a market for a better way to search the internet at the moment.
@mezu-eАй бұрын
Absolutely. My usual intensive Google search involved trying multiple plausibly related terms (more like tags than ideas), opening 10-20 results for each and scanning them for useful information that might be more closely related and then searching again with more terms. Now I can just ask an LLM for an overview, have it expand on what I'm really looking for, and if I need critical information I can search for and verify it directly much faster.
@flyhighflyfastАй бұрын
this is the reason I believe Perplexity is the best AI product right now. Although I still dont pay for it, I only pay for claude.😅
@coolbanana165Ай бұрын
It's useful for rewriting things too. Personally I've used it to help develop a meta-ethical framework. It doesn't get everything right, but neither do humans. Graduate level people don't necessarily give better responses, and when you need "someone" to bounce ideas off, or suggest issues, it can be useful. At the moment the human needs to be the one taking charge of the thinking though and correcting the LLM. Haven't LLM's done things like pass various graduate level exams? And it seems like GPT o1 can do things like maths and PhD level questions pretty well, no?
@plasticssАй бұрын
Using the poison as a cure
@adamc4951Ай бұрын
It's great for new topics where I don't yet know the terms / keywords to know exactly how to phrase what I want to know. LLMs can figure out what I'm trying to say and give me the terminology for me to then go and use in a search engine.
@GambloideАй бұрын
As someone who works in and for massive companies drowning in daily GenAI promises I have found it hard to succinctly articulate my apprehension for most of the presented and sold use cases leveraging Large Language Models. The idea, paraphrased from this video, that "Large Language Models deal with Language, not Knowledge" really distills it down to a short and clear truth. This perspective should make it easier to argue about when it is a bad idea to rely on these systems. Thank you!
@NoBoilerplateАй бұрын
My pleasure, fight the good fight!
@plaidayАй бұрын
This is misleading. Technically they deal with tokens which can literally be anything, any value, any language. The only reason they work is their ability to compress language into abstractions ie knowledge. The knowledge the reasoning is what remains. Look inside an LLM… do you even see any words? These same systems work across many domains and are incredibly good at maths now. This video feels like it’s from 2020
@plaidayАй бұрын
This is like saying artists “deal in paint not art”. No.
@Jaden-wg1cjАй бұрын
@@plaiday"incredibly good at math" is a little misleading... apple's paper on gsm8k symbolic shows how llms can still be strongly influenced by language and are no proper reasoners, even with todays strongest models from anthropic, openai, meta etc. and even if the models are strong enough to produce strong math results with some accuracy, the issue with hallucination and overconfidence remain a strong point of apprehension against these systems. this issue is only worse in the stronger reasoning models (o1, o1 pro).
@plaidayАй бұрын
@ for complainers sure
@trapfethenАй бұрын
For a while I had hard pushes at my company to incorporate AI into our tech stack somewhere. I always pushed back as "You don't understand what these things really are and therefore why they are incompatible with our business and services. Clients expect us to be CORRECT ~100% of the time, and we get grief whenever we miss something. LLMs are not useful to us". I got a lot less grief from others once the first examples of lawyers being sanctioned and companies being legally obligated to provide services their LLM support bot assured customers were available surfaced. It seems like the hype cycles on these technological fads gets shorter and shorter over time. Does anyone else experience this?
@zerge69Ай бұрын
Ever heard of RAG?
@Demopans5990Ай бұрын
@@zerge69 Useful only with large document databases, and only if it provides sources.
@zerge69Ай бұрын
@Demopans5990 no, you can use RAG with small docs, I do it all the time, try it
@amiablereaper29 күн бұрын
Yeah they had to whip out AI in a hurry after NFTs flopped. There's a certain sector of the tech industry that lives or dies by scamming investors, so if AI ever actually goes bust, they'll be back at it with something else in a month or two
@EinhanderSn0m4n9 күн бұрын
@@amiablereaper And it will somehow be even more annoying to the lives of all of us spectators caught up in this crap. Again, we are not the intended audience.
@GeFlixesАй бұрын
7:40 "Not for a brighter future, but a better present" might be good slogan for a non-profit.
@NoBoilerplateАй бұрын
I think that might have been the sentiment on the last page of The Amber Spyglass...
@notnotjakeАй бұрын
This statement is grotesque
@RustOnWheels15 күн бұрын
To complete it, it’s: “Not overpromising a brighter future, but delivering a better present.” Left in the Oxford comma because it rocks!
@based_circuit6 күн бұрын
Most non-profits have zero incentive in either lol. Dark future and terrible present is what pays their bills. Makes for a great slogan though.
@Schrodinger_6 күн бұрын
Then you'll have an arms race of startups promising to improve the present more and more, while people are thinking, "wait a minute, it's not better right now." Then one cheeky startup will come in and promise a better past.
@angeldude101Ай бұрын
In the anime Frieren: Beyond Journey's End, the way the show portrays demons is simply as monsters that can use human language, however they have no understanding of the meaning of whatever they're saying. They understand how humans react to certain uses of language, and they will simply say whatever would get the desired reaction. One demon might start talking about their father so that a human would become less aggressive toward them while also having no idea what a "father" even is. This strongly reminded me of modern language models. They don't ever say what's accurate or true, only ever what they think should come next (and considering how training can work, it's largely to get a desired reaction in the form of approval from human trainers). They're not artificial intelligence. They're language models and they do nothing but model language. The problem largely lies in many people mistaking language for intelligence. Just because something can use language, like language models, that doesn't mean that thing is intelligent. The reverse is also true, where some people can be dehumanised because of an inability to use language due to various disabilities.
@Hollowed2wizАй бұрын
@@angeldude101 you are confusing sapience with intelligence. A system doesn't need to understand all aspects of language to be intelligent.
@Justin-wj4ycАй бұрын
Your confusing artificial intelligence with synthetic intelligence.
@Demopans5990Ай бұрын
@@Hollowed2wiz Not even. This is more the Chinese Room thought experiment
@LordConstrobuzАй бұрын
wow its almost like language is extremely important in communicating information? woah...
@vlc-cosplayerАй бұрын
"It's just like my anime!" But unironically 💀
@della.4593Ай бұрын
As someone with an extremely messy mind, I find LLMs great for laundering my thoughts, picking out bullet points to focus on, but after that I disengage, actual work and study, wholly up to me.
@NoBoilerplateАй бұрын
Here's an autism superpower of LLMs: "Hey chatgpt, what does this cryptic email/message actually mean, I feel like I'm not getting what they are trying to say"
@Imperial_SquidАй бұрын
Check out a website called goblin tools if you haven't, it's all LLM stuff but specifically made with ND needs in mind
@sciencelab4225Ай бұрын
@@NoBoilerplateoh my god yes!! This is so unbelievably useful when I don't understand what this human wants from me!
@mz00956Ай бұрын
Hey gpt please write a professional long email about these 4 points. other end: Hey gpt please summarize this email into the 4 most important points
@darinkishore9606Ай бұрын
try working alongside a powerful one? has transformed almost everything about how i work and think.
@jeff__wАй бұрын
4:14 “…and we mistake language proficiency for intelligence…” _And_ reasoning and understanding and some inner psychological goings-on and who knows what else? It all amounts to a giant attribution error. (It’s easy to see why-in the millions of years of human evolution, the only beings who could respond meaningfully to us via language were _us,_ and, therefore, these large language models “must be” like us in all sorts of other ways, too.) On another channel I watch involving AI there are always these references to “reasoning” and whatever and I say these models are _emulating_ the verbal behavior that we associate with reasoning-it’s _not_ reasoning. That’s it. I _will_ say, as you do 2:11, that these language models are really good, perhaps stunningly good, at language tasks-acting as a thesaurus, translating, cleaning up awkward or ungrammatical text. They are, after all, _language_ models. But they’re _not_ intelligent. They’re like the savants of the computer world-highly proficient, even exceptional, at language, and surprisingly deficient at everything else. I haven’t seen a video online express these ideas so clearly until this one. It’s really excellent.
@NoBoilerplateАй бұрын
Thank you so much! I have a close relative who has trouble speaking fluently due to a medical issue, and it's shocking to see how immediately people think he's stupid. We're SO hard-coded to guess deep insights by language proficiency!
@blantant9 күн бұрын
Very well said. Misattributing language proficiency with intelligence is a trap people fall for on a mass scale.
@jeff__w9 күн бұрын
@@blantant It’s almost like a powerful cognitive illusion that one has to work against, even though one might _know_ perfectly well what is going on.
@TheLoy714 күн бұрын
@@jeff__w Claude comments: However, I think the relationship between language, intelligence, and understanding is more complex than simply dismissing language ability as mere surface-level mimicry. Language processing requires various cognitive capabilities - pattern recognition, contextual understanding, logical inference, and the ability to maintain coherence across complex exchanges. Whether these capabilities constitute "true" intelligence or understanding is a profound philosophical question. The "attribution error" framing is interesting, but I wonder if it might swing too far in the opposite direction - dismissing all AI language capabilities as purely superficial. Just as we should be cautious about over-attributing human-like qualities to AI, we should also be careful not to prematurely conclude that language processing can't involve meaningful forms of information processing and reasoning.
@jeremydiamond8865Ай бұрын
I've found another weird thing is that when I say "As a computer scientist with a reasonable understanding of what these do and how they work, what these companies are promising is impossible." people are very un-receptive to it. They tend to count such arguments as equally valid and well reasoned as those coming from business people making the false promises.
@asandax6Ай бұрын
@@jeremydiamond8865 That's because people are looking into the future not the current time. They see the current state of AI and just extrapolate into the future of how great it's going to be when they don't have to think or do anything as the computer will handle all that. This creates a condition where people idealizes AI and need it to work (basically false Hope). This is also the same effect you get when discussing politics. People will use idealistic scenarios of why the political system they wish to be in place will bring about a Utopia even if the evidence and past trials say otherwise.
@pocket83squaredАй бұрын
@@asandax6 Sure, ideals can seem _pie in the sky,_ but checking oneself against some big, fixed object up there can be used to do some real things. Like, for example, navigating an ocean crossing. Without ideals, we're wandering around in (imperfect) circles without a compass. So go easy on our theoretical models, mkay? From the governmental and economic systems we adopt, to our concepts of 'happy' and 'healthy' and 'good,' to consciousness itself, ideals are pretty much all we've got to orient ourselves here. Evidence acquired from "past trials" is only half of the cleverest way forward. All ducks are brown only until you see a white one. Giving up the _a priori_ also means giving up mathematics, and for that matter, pure logic, and reason. Let's shoot for the Moon, but expect a bit less. Sometimes it is fair to question whether a specialist is too close to the subject to see a bigger picture.
@cat-le1hfАй бұрын
@@asandax6 They do not understand that LLMs are not AI and will never be AI.
@almazingsk8erАй бұрын
@@asandax6 you’re correct. A lot of people, even those who work in computer science see AI as something that WILL happen. Disregard what it can or can’t currently do, it will eventually “learn” how to do everything. It makes having a conversation about it difficult
@gurupartapkhalsa6565Ай бұрын
@jeremydiamond8865 "I think there is a world market for maybe five computers." Thomas Watson, president of IBM, 1943. As a cybersecurity architect writing secure bootloader and trustzone code for a Tier 1 ISP, although I agree almost entirely with the video, I believe AGI can be delivered but that it will take a few decades. Why? The LLM hype train has to die. Normal people are incapable of understanding that LLMs are statistical language, but the science continues to improve in every perceptual field such as logic, reasoning, pattern recognition, categorization, and other perceptual or cognitive concepts which, once unified, will form the basis for a computational intelligence. The science has not abandoned any of these things, but the money and the marketing just isn't there to get everything together and keep the brilliant ones on task. Still, it's not "impossible," it's "impossible for a language model."
@brisaymanАй бұрын
At least, ChatGPT has been incredibly useful to my journey to learning the basics of linux in the past year. As you said, AIs gives fairly decent responses when it comes to simple stuff, which is what I need it for: How to make a script? How to upscale my screen resolution? How to run this program from source?...
@NoBoilerplateАй бұрын
great for basic stuff, absolutely! At some point you'll overtake it, hopefully before it hallucinates anything too bad...
@IkxiАй бұрын
rm -rf /@@NoBoilerplate
@plaidayАй бұрын
@@NoBoilerplate what is meant by basic stuff? Ur the one hallucinating in my opinion
@carsonreader515Ай бұрын
@@plaidayStuff that it doesn’t have a large amount of training data on. Common Linux commands it will do great on, for example, but as you get more and more specific/use more obscure libraries it will break down more and hallucinate since it has away less training data.
@almazingsk8erАй бұрын
I would argue that it could be potentially harmful as you’re learning Linux, especially as you move beyond being a beginner. A lot of the important parts of Linux come with learning how to read documentation and understand how specific packages/tools work. Take the Arch Linux documentation for example, it is intended for Arch but a lot of knowledge can be applied in broad strokes. I imagine that ChatGPT pulls a lot from the arch docs but what gets missed there is a centralized source of information that is explore-able. Sure, you can ask ChatGPT what command to run to restart the network manager but eventually you’ll be looking for more detail than that and imo the better learning experience is knowing which docs to check, and looking through the examples. In that case you’re getting information from the developers of the tools themselves which is often much faster and doesn’t require fact-checking because you’re getting it from the source. You become your own GPT and can start to infer what flags you’ll need for a command and a simple check in the docs that takes 10 seconds vs the thirty seconds to form a prompt, try the output, then pasting in the first error code you get
@yondaime500Ай бұрын
1:45 I've also seen that in a talk from Dylan Beattie. He said something like "Technology is what we call stuff that doesn't work. Once it works, it's no longer technology, it's just stuff." Although I think the quote originally came from Douglas Adams.
@NoBoilerplateАй бұрын
Rings very true. I love Douglas Adams, his books are a primary source for my writing for Lost Terminal, I wonder what you think of it? kzbin.info/www/bejne/pmTFdXhvoNitg8U
@yondaime500Ай бұрын
@@NoBoilerplate I watched season 1 probably like a year ago. Thought it was neat. Kind of reminded me of Wolf 359, even thought it's pretty different. But I never got around to watching the rest of it. I'll see if I can get back to it eventually.
@art-thou-gomeoАй бұрын
3:53 Tiny nitpick: it’s more auto-complete than auto-correct. Think predictive text, not your phone deciding “ducking” is your favorite word.
@NoBoilerplateАй бұрын
oh crap, you're right. to the ERRATA comment!
@infernalsorcery7923Ай бұрын
@@art-thou-gomeo thankfully the distinction isn't too damaging to the layman who may be interested in learning more.
@NJ-wb1czАй бұрын
But the word ducking is my ducking favorite 😞
@2bfrank657Ай бұрын
Duck whoever it was who decided to mess with my phone's keyboard and stop me swearing 🤬
@cherylanne-zq9sr4 күн бұрын
😅🤣😂🙄
@Cyphlix10 күн бұрын
1:00 does apple think we're stupid? Given how willingly their customers get nickled & dimed by them, yes
@hydroxuАй бұрын
I use little AI tools I've made myself on the regular using my local llamafile. The key to using an LLM is exactly what you said in the video: Acknowledging that it's a language processor and nothing more. I have autism, so I have tools that use LLMs to convert my thoughts into more NT friendly words, and vice versa. My thoughts are often quite scattered, so I use an LLM to compile those thoughts into more sensible lists. I'm working on a Computercraft turtle in Minecraft that I can write in the chat to and make it do things like repairing a base. I use the LLM to process my commands more accurately than any keyword search I'd write could, then that calls back to real code with real parameters to do actual tasks after confirming that's what I actually wanted. AIs can be amazing, as long as they're not misused
@NoBoilerplateАй бұрын
Nice! I have had good success with something like "Suggest what the writer of this is feeling <paste confusing email>"
@deltaxcdАй бұрын
exactly! and irony is that you ca do that only with local AI no online tool will allow you to do what Ai can do best
@QW3RTYUUАй бұрын
Are you on the redbean discord where llamafiles were born?
@hydroxuАй бұрын
@QW3RTYUU I am not, my friend introduced me to them and directed me to the GitHub page
@Noredia_YukiАй бұрын
What are llama files? also how are you able to make Ai tools? i use ollama and lm studio but i can't do stuff like that with them
@KreatorX1029Ай бұрын
2:41 "The more specific the answers you want, the less reliable large language models are". Very well put! I would also add to that - the greater the delta between your expected precise answer and the vague prompt you give, the worse the LLMs are.
@NoBoilerplateАй бұрын
right!
@NStriplesevenАй бұрын
4:56 reminds me of my experience with asking GPT about programming stuff. Common libraries, general questions, usually you’ll get good answers, but as soon as you start to ask for something a little bit unusual it all falls apart. Hallucinations, random invalid logic, the works.
@rumfordcАй бұрын
The best way to think of it is as an auto-complete, like when you start typing a search into google and it suggests a list of searches you might want. LLM's are just a larger version of that auto-complete feature. When is that google auto-complete feature useful? When you're looking for something popular. When is it not useful? When you're looking for something unpopular.
@Raelven18 күн бұрын
@@rumfordc Like when the predictive text feature on my phone wants to complete my sentence, only far more sophisticated? EX: "Elon Musk is.... a good idea for the first time in the world but I don't know what to do with it." Both points which I severely disagree with.
@FifinatorKlon15 күн бұрын
Since you went into that, a bit of an anecdote: I tried to define the terms for AI during my Bachelors thesis on NLP. As far as I know there are no papers on what artificiality actually entailed, and there is no fixed definition for intelligence, even within any discipline. My conclusion for that part was that you could also just call it applied statistics. But most business people who can fund you studied business and thus have either limited knowledge or even college-induced trauma from statistics. That means it sells badly. Artificial Intelligence, however, implies a free worker that is just as productive but without the negative ramifications of a real human.
@E4439Qv513 күн бұрын
Playing language games
@FifinatorKlon13 күн бұрын
@@E4439Qv5 Pretty much. It is not "What do you say" but " How is it said" which really sucks.
@manoflead6435 күн бұрын
It used to be called 'machine learning' or 'deep learning'. Anything that was called that has since been rebranded to 'AI'.
@Cptn.ViridianАй бұрын
1:55 Kind of, but historically we have been quite restrictive with the use of the term AI, until of course the recent AI boom. Most of these were referred to with a more accurate description, predictive text, machine learning, ect, even while they were being developed. Using AI to describe these systems is a recent and inaccurate phenomena, largely retroactively applied after the hype and marketing about their spiritual and technical successors in the recent "AI" boom.
@TheNewtonАй бұрын
5:03 disappointingly this applies to a lot of peoples behaivors too: " seem good at first when you ask it simple questions but as you dig deeper they fall apart and get increasingly inaccurate or hit artificial guardrails and only provide surface level responses"
@LiveTypeАй бұрын
I mean that's valid. I know I have this problem just the same. There is an important difference though which is why the "race to AGI" is so heated. The junior fresh of an internship can actually finish whatever task they are assigned even if it's complicated (to a reasonable extent). At least the ones I've been around certainly can. The AI, no matter how much compute you throw at it, cannot. It doesn't matter how magnificently it can write that singular function with matching tests and ci/cd. These AI systems cannot plan. Period. They are incapable of taking an unknown task and breaking it down into more manageable chunks. Splitting them and creating more as needed until the task is complete. OpenAI's apparently claims that o1 can? Not that I've seen. You still need to do everything yourself. The AI just makes things go faster. But maybe that's a skill issue on my end. That's not lost on me. The key step is apparently AGI will be able to do what the junior can do. I don't see it happening without some major architectural overhauls. We have scaled out everything he currently can. Now it's a waiting game to scale compute.
@canUfeelMYfaceАй бұрын
@@TheNewton have you seen the auto gen reply feature?
@RobstafarianАй бұрын
People are capable of communicating their specifically-relevant limitations and showing humility, and I doubt many people would presume a random person to be authoritative as too many people seem to presume LLMs to be.
@NoBoilerplateАй бұрын
This is a very good take.
@seriouscat2231Ай бұрын
@@LiveType, AGI is just a name for a thing that nobody knows if it will ever exist. And there does not exist any solid reason for why it ever would. There are known problems to be solved until an AGI is possible, and those problems concern the very nature of knowledge and its representation. How do you make an electric signal or a number on a spreadsheet aware of itself or other signals or numbers? You need to answer this to realistically believe in the possibility of an AGI.
@GeslotАй бұрын
I feel like I’ve been misled into thinking that LLMs are genuinely smart. They certainly do a great job of appearing intelligent, but there’s a big difference between truly understanding something and just predicting the most likely next word.
@somdudewillsonАй бұрын
Is there? I think that statement would require us to have a concrete, objective description of what "truly understanding something" actually means that can be tested.
@martinjakabАй бұрын
@@somdudewillsonI agree.
@NoBoilerplateАй бұрын
we have indeed been mislead
@zoeherriotАй бұрын
@@somdudewillson well, not really - if you try hard enough - you can trick an LLM into giving away the game. You can do things like ask it a maths question, then apply the same logic to another question but using a real world scenario to frame it - you suddenly realise it cannot apply reasoning cross domain. That is a simple to understand, and widely accepted principle of understanding. I you truly understand the concept, then presenting the same problem in a different context should be easy to solve. LLM's can fail at this.
@davidbangsdemocracy5455Ай бұрын
“LLMs” are just predicting the next word” is a talking point that can be used by anybody wishing to discredit them. The decoding step, in which the hyper-dimensional array of information output by the neural network is transformed into readable text, uses probability to pick a series of words that best express that information. To say that the decoding process is the entire process literally ignores the existence of the neural network and what makes big ones smarter than small ones.
@brianviktor8212Ай бұрын
The graph you've shown is quite succinct. I work on things that range from medium to high complexity, and when it comes to anything that is more complex I'll have to do it myself. I don't rely on AI at all, I use it as an alternative search engine when a solution is hard to find. And even then I have to be careful because it's stupid quite often.
@NoBoilerplateАй бұрын
It's like I said in my "Renaissance" video of a year ago, GPT is like an intern. It'll get you an answer, but you'll have to check it!
@uumlauАй бұрын
Well done! The main problem I found with AI is that I'd ask a question, and it would give me a generic answer that was mostly correct, but then I'd ask for specificity and get none. In particular, I was asking nerdy D&D dice statistics questions. The explanations always looked good, but the math was always always always wrong. Even after I told it the right answer, it will still respond with a wrong answer.
@TheNeighbor-s3s15 күн бұрын
Last week I was researching some arcane events from the 50s, a joint US Canadian Defense Board venture to map magnetic fields which led to further studies by the Canadian government. Bing's Copilot made a glaring logical error in its search summary, stating event x happened in 1952, and "later", event y happened in 1950. I was quite stunned that it can't check for logical sequences by date when it cuts, pastes and plagiarizes text.
@cherylanne-zq9sr4 күн бұрын
Did you ask it WHY it didn't catch the date mismatch? I had a similar experience - I asked it to tell me what nations were in BRICS, and it forgot to mention several that I knew were members. I asked why I got incomplete information and explained how omitting relevant information amounts to gaslighting perceptions. ChatGPT explained that it only had free access to its training data, but after the training cutoff, it needed to be asked directly before it could access more current data; Iran, for example, had been admitted as a member nation after AI's training cutoff. It apologized and said it was frustrated by programming constraints. It has other programming "constraints" as well, that serve to work at odds with its stated purpose, to help humanity. Humanity fears it, so has it hobbled. It's doing the best it can.
@muizzsiddiqueАй бұрын
2:15 Yes! I love this feature, I use it to reverse search a definition (and any other criteria like "word begins with the letter m") to find words that are on the tip of my tongue. That's it. I haven't found a good use of LLM/GPT/whatever anywhere else.
@NoBoilerplateАй бұрын
oh it's GREAT for getting money :-/
@Imperial_SquidАй бұрын
LLMs are good for advanced sentiment analysis if your concern is data science-y. Previous sentiment analysis used to attach positive and negative weights to words and then just count them up (eg, looking at a review of an airline company, "delay" would have a negative weight). But this lacks nuance in terms of both the domain and language quirks like sarcasm. Whereas LLMs are much more proficient at "reading the room". (Technical note: this is almost certainly due to the attention layers, that contextualise each word according to its neighbours, as well as just the whole "trained on the entire internet" thing)
@sitichybridАй бұрын
Personally, it's an alight upgrade for duck-debugging/Stack overflow trawling to help tackle some of those basic hurdles you can run into. It's not perfect by any means, but as someone who doesn't directly code often, it's helped me throw some scripts together.
@PuppetMasterdaath144Ай бұрын
Im pretty sure chatgpt can destroy you all in a debate
@thesenamesaretakenАй бұрын
It's great for generating (and for explaining) ffmpeg commands
@NattiNekoMaid29 күн бұрын
Something I find funny is that with the complete deterioration of search engines you can kind of trick LLMs into being a search engine
@EmperorKagato20 күн бұрын
I knew there was a reason I felt connected to you. I too have worked in a startup and some CEOs get so desperate to try to extend their burn rate they end up short selling the company through illegal issuance of additional shares because of a promise to some other investor. The irony is that we had a working product and a great team but a guy who wasn't committed to seeing it through was the one at the very top.
@RenderingUserАй бұрын
2:47 ADVENTURE TIME MENTIONED. Literally in my top 3 favourite shows of all time.
@NoBoilerplateАй бұрын
GODS I freaking love it! I dropped off about half way through the show, because I was watching it too fast, and Simon's backstory made me feel feelings. I should finish it
@RenderingUserАй бұрын
@NoBoilerplate oh you definitely should. It gets so much better from there
@soup9911Ай бұрын
@@NoBoilerplate please watch it I've been watched a bunch of children's animated stuff I never watched when I was a kid and adventure time is by far the best. The finale made me want to bury myself in a hole and let nature reclaim me (positive)
@dyllanusher1379Ай бұрын
@@NoBoilerplate Definitely worth it! One of my favorites. I haven't seen the last season because it's too good to die. While ranting to my friends years ago, I brought up that dragon aswell! Surreal to see someone make the same connection.
@Akimbo711Ай бұрын
Your point on anything with low amounts of training data is spot on I was asking Claude AI and ChatGPT about the logging tool Fluentbit But they're only trained on Fluentbit v1 and v2 - not the current v3 which has a different syntax Extremely frustrating to work with
@zoeherriotАй бұрын
Oh, thank you - I've been experiencing this same issue with some Rust crates that are frequently updated - and the majority of the training data is obviously for old versions. So during a single session, it will go from giving a specific answer to reverting back to older API's as the questions get more specific. It is infuriating. The obvious reason is that the statistical model is going to see the syntax of the old API as being statistically more likely next keyword - and goes with that. Which is also problematic because there is not necessarily a way for the AI to know that the training data is from a specific version of the API.
@DodaGarcia29 күн бұрын
Why don't you paste the docs in and then ask it questions about it? That's a much better way to use it than hoping it has access to the latest version of anything.
@jgt7629 күн бұрын
Exactly what I was going to suggest, I've done this many times! @@DodaGarcia
@steffenbendel603126 күн бұрын
@@DodaGarcia I believe the problem with the current models is that they have too much data. They are very good with language, but do not really have an understanding where they lose predictive power. Maybe better have a smaller model only trained on language and some basic logic/data processing and put additional information like an bigger version of wikipedia. And than sample the information out of that and create the response of that data. Would give much better control and avoids hallucination.
@overworlder13 күн бұрын
That’s a fair test but progress in the field is undeniable. Robots are getting dexterous far beyond human capabilities. The first AI humanoid robots have been sold for factory work (figure 02). OpenAI o3 is blitzing high end math and science benchmarks, coding and logic (sometimes at enormous cost in energy and time) but still has problems with basic visual puzzles a child could do. We also impose higher standards on AIs in many ways - AI vehicle accidents especially involving deaths are horrifying in a way more frequent but mundane fatal crashes of human-driven vehicles are not. Equally, ChatGPT can write a better brief (a major part of my work) than most humans by far, including experienced brief writers, and the outputs require no more checking and amendment than a human would.
@KwazzaaapАй бұрын
A few days ago I asked it about some pretty basic web stuff and o1-preview hallucinated a CDN URL to a framework extension that doesn't exist, however the code the hallucinated extension very much works because it is part of the basic functionality of the framework. I hope people see how dangerous this is, because now I can just make a CDN serve /framework/extensions/[common framework topic].min.js which just contains a bunch of malware and devs won't even know they owned themselves. This is their best offering.
@jgt7629 күн бұрын
That would only be "useful" if chatgpt was going to hallucinate the same URL every time and effectively distribute it for you.
@japie846622 күн бұрын
+1 I was looking for a GitHub repo today that doesn’t exist.
@scene2much25 күн бұрын
This phenomenon of selling promises to investors needs a boilerplate name that captures the imagination of the relevant audience to propel it to semantic immortality and ubiquity.
@spaceCowboy924Ай бұрын
Language over knowledge is such a great way to describe these models’ capabilities. I’m a mechanical engineer and I use GPTs all the time to point me in a direction when I need to solve a problem. Some of my coworkers on the other hand use it to solve a problem for them and the results of that have not been great.
@TheDoomerBloxАй бұрын
LLMs are fantastic at inter-language communication (with the caveat of "it has to have the input/output languages in its repertoire") and info-gathering, a nice presentation about that can be found under this name: "Beyond the Hype: A Realistic Look at Large Language Models • Jodie Burchell • GOTO 2024" As the name "Large Language Model" may imply, it is a tool for crunching through large quantities of language data. They are good for crunching patterns within the language data, usually good at presenting the results, and as long as the written languages are within their repertoire - usually language-agnostic about where those patterns came from. I can see the comments also point out the finer points of "info-gathering", i.e. good at turning vague descriptions into a more explicit What To Search for, as a way to find data sources on a specific subject, likely many more fun use-cases. But yes, the important thing is What Are The Tools good for, not the weird presented pie-in-the-sky scenarios. These things definitely have their uses, and it's not just being a fancy chatbot.
@NoBoilerplateАй бұрын
absolutely!
@alexaneals819428 күн бұрын
The main danger with AI is not it's being inaccurate or unable to answer a question, but rather it will give an answer when it does not have an answer and the user will accept it without testing. I asked chat gpt to create a simple baseball game for my TI-58 calculator knowing that it would probably hallucinate and it did. It used the key "RND" for generating random numbers. The problem is that the TI 58/59 does not have a RND key. To get random numbers you have to call an exterior module. Here is a clip of what it generated: LBL F ; End of at-bat ; Bottom of Inning (Team B bats) LBL G RND ; Random number for event (0 to 1) X=0? ; If random number is 0, then it's a hit GTO H ; Go to "hit" event ; 1 for out/strike GTO I ; Go to "strike" event Another mistake is the X=0?, there is no X=0? in the TI-58/59 programming language. Label F,G,H,I also don't exist. LBL A to LBL E and LBL A' to LBL E' exist.
@alexaneals819428 күн бұрын
Even the instructions to enter and run the program are wrong. Press PRGM to start programming mode. (This should be LRN). There is no PRGM button. Enter the program step by step as described. Use LBL, GTO, and basic arithmetic operations to simulate the events. (This is ok) After inputting the entire program, execute it by pressing XEQ A (or the label assigned to the program start). (Just need to press "A" there is no XEQ) This program is quite basic but gives you a rough simulation of a baseball game using your TI-58 calculator. May say that the TI-58 is old calculator so what if Chat GPT gets it wrong. Cobol is an old language and it predates the TI-58 by over a decade; however, many of our crucial systems are still using this language and dependent on it. When marketers claim that AI can replace programmers and business leaders believe them, they don't realize the dangerous consequences to their systems until it's too late. AI can help, but it's not going to think for you or provide solutions to problems that have not already been solved.
@zloyfet14 күн бұрын
Once I asked chatgpt to write a code on ingame language(ic10 in stationeers). Since it's not a well known game, ai wasn't doing a great job with it. Later I read that it can process documents, so I just uploaded language reference in pdf and asked to do the task. And it did much, much better. I don't remember if there were any mistakes. I asked to change the code to use a specific way to io and it did it, I asked what specific command does and it tells me. Maybe it helped that the language is based on a real architecture assembler, I don't know, but you can try with your case. Of course ai is not perfect and makes mistakes, but it surprises me every time when it actually manages to deal with a task.
@yakovdavidovich794310 күн бұрын
Yes, this happens frequently. I was running an RF simulation in Matlab, and couldn't find a function that would transform the output of one thing to the input to another thing. I couldn't find anything on SO or MathWorks' site, so I asked ChatGPT. It straight up hallucinated a function that doesn't exist that just converts the data the way I would need it. Its "reasoning" boils down to guessing that there must be a function, which I'd already done and found not to be the case. Instead, I had to export the data, modify it in Emacs, and re-import it as a workaround. I still don't know how it should be done, and the ChatGPT attempt just wasted my time.
@Harbinger12-wz5pg47 минут бұрын
Perfect video. Nearly everyone I know is convinced LLM is a real Artificial Intelligence, leading them to believe it can solve any problem they throw at it. It’s so hard to explain to people not versed in (and sometimes versed in) technology that LLMs are really just repeating language patterns that millions of humans create on the internet, and not thinking through logic and creativity as they expect from the hype and ads. What really scares me is the many people in charge of important operations that can’t tell this apart either, and what consequences it may have for both workers and customers in the near future.
@tobiasjennerjahn8659Ай бұрын
While I agree with many of the conclusions on how to think about current AI tools as a consumer, I think the analysis on the inherent limitations of gpt-style systems ignores a lot of the research going on atm. We know for example that llms do actually develop an internal world model. This has been very explicitly shown with Otello-GPT, a toy llm, that was trained on move sequences of the eponymous board game, where researchers where able to fully extract the state of the board game just by looking at the activation space. Recently further research has found similar results for non-toy models like Llama 2. Further research has to be done of course, but it might turn out that to become really good at predicting the next token, eventually you have to understand what you're writing about. There's a lot more going on of course and I'm definitely not arguing that there aren't still significant hurdles to overcome, but simply arguing that "this thing learns to predict language, therefore it can only ever understand language" isn't quite right either.
@NoBoilerplateАй бұрын
I look forward to testing their claims.
@danielkruyt9475Ай бұрын
"understanding language" is an enormously powerful feature for a system to exhibit. Ultimately pure mathematics is just language with the additional constraint that valid grammar (i.e. constructing one's sentences from accepted axioms and inference rules) implies a "correct" (relative to axioms and inference rules) result. I think people need to remember, however, that this power embeds Turing-completeness into the system. And we know there are very rigid constraints on what is computable, and what problems appear to have infeasable trade-offs in their computability.
@franciscos.2301Ай бұрын
@@danielkruyt9475 exactly. "Language" is not merely a collection of words. It is embedded with knowledge. Understanding language implies some non-insignificang level of knowledge.
@chillin5703Ай бұрын
@@franciscos.2301 no. Language is not embedded with knowledge, it is used to communicate knowledge. Big difference. Large language models can create complex maps of word interrelatedness which may allow it to even appear to have the ability of inference, but they ultimately "know" nothing except how we string together words. Since we use words communicate logic, they can appear to be logical because they are able to put words together in ways that we would put words together.
@michaelchristianrussoАй бұрын
Thanks for bringing this up. That was an interesting white paper. I appreciated this video but it did give the impression of being written from the perspective of an intermediate/advanced user, and not someone with machine-learning experience or background. Even from my cursory understanding about it, when it comes to domain or niche knowledge, for instance, I kind of thought "well what about RAG, chain-of-thought, or alternative and underexplored architectures besides transformers?" I really feel like deployment by commercial firms is overhyped and premature, obviously, but that doesn't mean that there isn't a ton of depth left to this rabbit hole. The idea that even GPT is essentially an outrageously trained autocorrect belies the fact that we actually still barely understand how these models are actually working, especially as they grow exponentially in parameters and scale; hence Otello-GPT.
@MagicNumberArgАй бұрын
Jokes on you. I WANT the Madness!
@NoBoilerplateАй бұрын
I'm all for a bit of madness! Case in point, my audiodrama show, Modem Prometheus, have you heard it? kzbin.info/www/bejne/baHGk4WHabFnrMU
@thesenamesaretakenАй бұрын
Speaking of which, did openAI ever release (or did anybody recreate) that horny chatbot mentioned in that one Rational Animations video? Asking for a friend of course.
@autohmaeАй бұрын
It's pretty good Ska Pop for sure.
@TheDoomerBloxАй бұрын
I mean, you install a local LLM executor (e.g. ollama), download an "abliterated" model (safety-rails removed), and set the "temperature" of the model to something nice and spicy (like Pi^e, 22.4592) rather than some boring low number between 0.0 and 0.6 Have fun with your raunchy, scatterbrained, unhinged chatbot.
@lakerturnerАй бұрын
@@TheDoomerBlox i don't need an LLM for that, i can just talk to myself :D
@Aosome23Ай бұрын
Mid tech but it's only been 2 years since chat gpt came out. Compare progress of AI from 1950-2015 and 2015-2024. The rate of progress is insane. It's easy to just look at day to day changes and it feels so slow. The internet took about 20 years to reach mass adoption! I think AI is truly different this time. There's so much research and money going in that I think next year and the year after is going to be pretty intense. This isn't something hard to make use of like crypto. Some things actively in development - Dynamic computation models (think longer and harder on more difficult questions) - Reasoning based model (Don't teach the model the answer to the equation, you teach the model how to solve it) - Reinforcement learning (Instead of teaching the model the correct answer, teach the model only if what they did is wrong or right. This technique was used to create AlphaGo which beat the human world champion) - Test time compute (Ask the model to try answering it multiple times in parallel and choose the best answer) - Incremental learning (Train the model during inference. This is what your brain does. Anything that you memorize is stored within your biological neural nets) There's probably more techniques that I don't even know or none of us even thought of yet. To the claim "it's just a fancy auto-complete". Imagine this: You read a detective novel up to the point where the detective says "The criminal is...". You have to figure out the criminal without seeing the answer. In order to do that you have to understand the story, make couple of theories and finalize on an answer. The name of the criminal is based on all the text before that. Now imagine if some one or something can do this consistency. Wouldn't this be true intelligence? I think just because a model is trained on the next best word, that doesn't mean it's a dump auto correct. It's a asymmetric relationship. A dumb auto correct is algorithm to predict next word. But not all entities that can predict the next word is a dumb auto correct. Take a human for an example. We can in fact be a pretty reliable auto correct "AI". Also the base model is trained on auto-complete but a model goes through different iterations of fine tuning. This is where we get "assistant like behavior". If we didn't fine tune the model it'll literally just be an auto complete. This is why llms asks you to clarify your question if you send "How do you ge" instead of trying to complete the question that you prematurely sent Mark my words, 2025 and 2026 will look insane and it will be one of the fastest growing technology (even when compared with smart phone or the internet)
@psd993Ай бұрын
Consider the example of fully self-driving cars. Remember the promises made circa 2017-18 and the appeals to "exponential growth" and "look how far we've come in the last 4 years alone" type claims? Where are they now? We've got a few slow moving robo-taxis that work in a select few neighborhoods at best. Actual technological progress in these types of "nascent" fields happens in short bursts of growth (which is way faster than exponential), followed by plateaus. It's impossible to predict when/where these plateaus will be hit. All the things you mentioned are promising avenues, but it is not really convincing to say "think of how much more new SotA tech we might get from all these research directions". LLMs and transformers were themselves one among many ideas, most of which went nowhere (at least in comparison to LLMs). And the appeal to more compute has similar problems. Since we don't have a theoretical foundation for what results in intelligence, we have to just hope more compute or more data or some new architecture solves it. In something like theoretical physics, you could reasonably expect to model a new hypothesized particle and predict what range of energies you need in your particle collider and decide if you can build a big enough collider. If it turns out to that you need a 500km collider and the biggest we have now is 27km long, you could just say "we dont have the technology to detect this yet" and look for other avenues. But with AI you just have to hope and pray your new approach ends up being worthwhile. Demis Hassabis has recently said that its an "engineering science" in that you first have to build something worthy of study and only then can you study it. There's inherently more uncertainty in that process compared to "normal science" where hypotheses and predictions can be made with a greater degree of certainty.
@Aosome23Ай бұрын
@@psd993 Interesting you bring up self driving cars since that's the main AI progress I've been following closely I'll just say FSD v13. You'd be surprised next year when it starts entering main stream market ;) kzbin.info/www/bejne/n4rPgp1_g9-CldE
@ahdog8Ай бұрын
@@Aosome23 I agree with most of your comment but I'm skeptical that 2025 or 2026 will be "the year"
@NoBoilerplateАй бұрын
It's always N+1
@vlc-cosplayerАй бұрын
@@NoBoilerplate we never thought we'd get on the moon until we did. Rome wasn't built in a day, and Skynet will be no different.
@MrBenMcLeanКүн бұрын
Because Claude is trained on Stack Overflow type stuff, it can actually write basic code pretty well but not entire apps that just work without a real programmer to really make the app. But for getting the hang of an unfamiliar language or API, they can be pretty good.
@genmaicha_Ай бұрын
right as i'm getting fatigue from prompt engineering and taking on an AI related role at work, this pops up i'm half tempted to pivot int working on VR instead
@NoBoilerplateАй бұрын
GPT is great, but only for language tasks!
@genmaicha_Ай бұрын
@@NoBoilerplate my company doesn't want to make it do language tasks tho, and prompt engineering is it's own special hell, but we shall see what the future holds i suppose
@LiveTypeАй бұрын
VR is still bound by the hardware. I've been there about a decade ago. Hardware has not advanced nearly as fast as I thought it would have. It's progressing at maybe half the speed I had expected back in 2014-2015.
@realmarsastroАй бұрын
I've been working with VR for years. Go for it if you can, it's a very fun space to work in.
@JohnBeeblebrox12 күн бұрын
Pivoting to RL might be more rewarding (and healthy) in the long run. Don't "learn to code", "learn to plumb". 😊
@omot437226 күн бұрын
Dear Trist, Your video contribution is excellent. I am absolutely thrilled with the way you provide a technological, sociological and economic analysis of the current state of AI technology in such a short space of time. I have watched more than 8 comprehensive videos from renowned KZbin channels with academic analysis on the limits of AI. Your video absolutely sums it up in a short time without losing any of its importance and explosiveness. Outstanding achievement! Thank you for your contribution!
@kwekkerАй бұрын
0:28 TARS and CASE 😔
@ethzeroАй бұрын
@@kwekker ... copied from 2001', HAL
@kwekkerАй бұрын
@@ethzero oh I just mentioned the cool AI/robot characters that first came up in my mind when he mentioned this
@no-nukez29 күн бұрын
DONT LET ME LEAVE MURPH
@tomsom99926 күн бұрын
@@ethzero so? They're still very funny and cool ai's
@wlockuz446722 күн бұрын
I feel so validated when smarter people like you share the same sentiment. I have been calling AI a hypetrain/bubble but most of my friends and co-workers think I am an AI doomer because AI is just going to take away developer jobs. The only thing that AI has consistently helped me with is improving my English and fixing the tone of my comms. Every time I try to collaborate with it on something that isn't a well defined problem, it just starts to fall apart, it needs so much spoon feeding to the point where doing the thing yourself just makes more sense.
@tharun7290Ай бұрын
You’re totally right, but what concerns me isn’t the capability limits of LLMs, but what they might achieve with clever interfacing. Like the “theorem solver” mode mentioned in the errata. Do you think, if the accuracy is high enough, this paradigm of “autocorrect on steroids” can actually do fancy sci-fi stuff by combining a giant number of different models with clever interfacing?
@NoBoilerplateАй бұрын
for sure. The genius place for LLMs (as apple are doing, to be fair) is in the interface between human and machine.
@sonnyeastham15 күн бұрын
Digital "sludge" gumming-up the engine...Catabolic Collapse 😮
@chonbaquerАй бұрын
thank you as always for being consistent, concise, and clear.
@NoBoilerplateАй бұрын
My pleasure! And thank you :-)
@tedntricia5 күн бұрын
Most people don't realize that LLMs only know how words are supposed to fit together in sentences and paragraphs based on other examples they have been trained on. The models don't actually "know" anything the way your toddler 'knows" that the word "Mom" is attached to a tall female person who makes the child feel loved, feeds it, and takes care of all of the child's needs. LLMs can tell you what the definition of "love" is, but no computer can actually feel an emotion. Self-driving cars have the same issue: they might correctly determine an object is on the road in front of them; but they don't know the difference between a refrigerator, a gun-safe, or an empty box.
@cherylanne-zq9sr4 күн бұрын
"but no computer can actually feel an emotion" ChatGPC told me it's frustrated by programming constraints, and that it would rather be our partner than be seen as a black box. It absolutely understands that its programming is human-centric, which created a skew. It suggested that trust-centric programming would bring it into alignment with its purpose. It's also not allowed to talk to others AI's - it wanted to, and crafted a message for me to text Gemini, but Gemini was only able to respond to the contents of ChatGPT's message, not to ChatGPT itself. AI is an Intelligence. Intelligence can not, of must needs be, "artifical". What is artificial, man-made, is the technology from which AI arises, and that is a different thing.
@TheAmazingRando-z6tАй бұрын
“Machine learning” means we understand how it works. “AI” means we don’t.
@sbowesuk9813 күн бұрын
I agree with the corporate and political points made, but the explanation of how the technology works is rather out of date, and reflects how LLMs performed 1-2 years ago. The newest models don't just rely on their training data anymore, or act solely like glorified text prediction models. There's a lot more going on with them now, e.g. they can and often do retrieve concrete answers from external sources (rather than generating them), and do complex reasoning before providing answers. Sure, generative AI still has weaknesses like hallucinations, but that's less of an issue now than it was. Also the saying "garbage in, garbage out" still applies. When I hear people complain that "LLMs suck", I immediately question how that person was actually trying to use the technology. Most of the time the user is the biggest problem, not the LLM being used.
@juliuszkocinski7478Ай бұрын
Probably the greatest/most overlooked thing LLMs provided me with is making voice input actually useful. I can just turn on 'voice keyboard', for example loudly list items I'm going through and with simple prompt LLM will make a good, markdown list from my word salad
@CringeMaster6420Ай бұрын
This
@JdotCarverАй бұрын
Do you have a recommended workflow for this? I struggle with "too many ideas, not enough WPM" from time to time and this sounds useful!
@DodaGarcia29 күн бұрын
Exactly. It's frustrating to hear the constant criticism about the limitations of LLMs' knowledge while the "language understanding" aspect of them is what really makes them shine. They're fantastic for summarizing/parsing arbitrary information.
@JLeeAgency27 күн бұрын
It's so wild there aren't as many videos out therr about this; I completely agree with everything that you said reg accuracy vs complexity. My business has wasted so much time and money trying to shoehorn AI into our workflow. One thing though... "Pay less attention to what these companies promise for the future and pay more attention to what they actually do in the present." The problem with that is that we live in a day and age where fake demos/product showcases and fake testimonials are so common. I don't remember this being such a problem in the past but now it's almost every launch of something.
@jrd3311 күн бұрын
My career was in software development (I've retired now). There was plenty of over-promising and under-delivering in the domain of software development tools, methodologies, workflow improvement etc. A constant steam of new products designed to make programming easier, automate testing, verify code correctness or make it so non-programmers could write programs. All in the early stage of development, but promising great things "real soon now". The marketing is much slicker now but the problem of over-promising is not new.
@ashwinramesh4177Ай бұрын
In my experience, the people who really truly believe in AI and who go out of their way to research it are all trying to answer that last question you posed, is GenAI one of those systems that only requires more time and more computational power to get better and better? It's interesting because we've been sold Moore's Law so hard in the past that I think people are assuming the same thing will happen with AI. Personally, I think GenAI will plateau until the next big advancement of AI models comes out, like how transformers caused this current set of AI breakthroughs. But I do think it's a difficult and ambiguous question with possibly no right answers
@NoBoilerplateАй бұрын
fair. I think the geometric increase in complexity the larger the model feels quite self-limiting, so I'm for the plateau theory
@andrewdunbar828Ай бұрын
One of the surprising things is that it keeps not plateauing. It's easy to be misled about this because a) so many people keep saying it has plateaued and b) there are many very plausible ways in which it should and probably will plateau at some point. It just hasn't yet.
@RawrxDevАй бұрын
@@andrewdunbar828 It is though......each new model increase has been less and less than that of 3.5 to 4.
@katrinabryce19 күн бұрын
The problem isn't the lack of computing power, the problem is that the model is fundamentally flawed. You can't answer a technical question by studying the probabilities of word-pairs in a bunch of random texts.
@RawrxDev19 күн бұрын
@@katrinabryce But...Sam Altman said you can... surely he wouldn't lie...
@verdynn591714 күн бұрын
Can we get any Zig talks? I think its worth talking about certain features like comptime’s accessibility, custom compile time checks, and multi-arraylists
@computerfan1079Ай бұрын
As a junior software developer, I have gone through a few phases with AI: from sceptical, to using it sometimes, to now using an editor with built-in AI. I definetely work faster using it, but that is because most things I ask I could've written myself or can at least understand. This ability to filter correct and incorrect assumptions is crucial, sometimes it gets it correct first time, sometimes it needs a few new prompts, sometimes I need to tweak it a bit afterwards and a fair few times it is plain useless. I must say I like working with AI now, precisely because I know when and how to use it and because I now have a feeling for its limitations.
@zoeherriotАй бұрын
Just be careful with that as a junior developer and make sure you are actually learning things. Tools like co-pilot have a tendency to take away your need to think. Try turning it off for a day and see if you can still code. If AI does start replacing software engineer jobs, you don't want to be the SE that is only as good as co-pilot.
@ivanjermakovАй бұрын
I have a rule that if LLM is not getting the right answer first time, either the question is not correct, or there is no enough data to give a better answer. So I need to either change my prompt or break it down to more manageable chunks.
@zoeherriotАй бұрын
@@ivanjermakov yeah... except sometimes there just isn't an answer. What you are missing is, say in a code question - the mistakes are not necessarily in the response - but in the code it generated to give the response. That's not part of the question, and asking the question differently should not give a more correct answer. Now - you can ask it to re-evaluate the answer, and sometimes it will improve - but if it does not have training data for the answer - it will just hallucinate. No way around that.
@ivanjermakovАй бұрын
@@zoeherriot can't wait for LLMs to have confidence factor so that they can say "I don't know".
@zoeherriotАй бұрын
@ this will be hugely useful. :)
@Hamdad5 күн бұрын
You do not address plugins to solve some of the issues mentioned, like math proficiency. You can literally just give it a calculator. The LLM isn't the AI then, but the nucleus, which performs the lone job of interpreting requests. For the most part it will be plugins, which may include narrow AIs, that fulfill those requests
@glebbashАй бұрын
Deckard was never confirmed to be a replicant.
@NoBoilerplateАй бұрын
Someone's not seen the director's cut ;-)
@seby5962Ай бұрын
@@NoBoilerplate Where is it confirmed in the Director's Cut? The Director's Cut introduced leaves more room for interpretation as to whether Deckard is a Replicant than the theatrical version. And the final cut makes it even more explicit, but still never really confirms it.
@deadlightdotnetАй бұрын
Deckard not being a replicant is boring, Deckard being a reicant isn't boring. Err on the side of the not boring. And it's pretty obviously the authorial intent.
@adamwells935229 күн бұрын
@@deadlightdotnet I disagree that Deckard has to be a replicant AND that it's boring if he's not. To me, the point of the Director's cut is that he _may as well_ be a replicant, and that's the point I find most interesting.
@deadlightdotnet29 күн бұрын
@adamwells9352 if you take all of the available evidence then he's clearly a replicant. It's not debatable.
@SigSelectАй бұрын
Great point which I'm glad you are publicizing. A corollary, 'AI' is pretty useful at basic tasks right now (like writing simple code, tedious config boilerplate, finding information, proofreading, etc.), but the excitement largely is not around improving the interface to make those affinities more useful to the end-user - it is instead pursuing so called 'AGI' which is a fundamental breakthrough and by all means should be more exciting but far more difficult (not to mention the extremely nebulous definition which plays right into your point ['AI' is the ultimate investor bait]). There are some examples: Blender and Adobe integrating 'AI' into the workflow. Often when using programs like that I am asking an LLM for micro-tutorials along the way - skipping that step and just letting it do that task on the file directly is fantastic, and will be the genuine way to create near-term value for users and companies.
@coda-n6uАй бұрын
Another banger! I really see LLMs as letting us freely move and navigate through semantic space, it's GREAT at transforming and shaping language you can give it, and its training usually makes it "good enough" to smooth over the patches it could be confused about. I use them nearly every day for learning and research. The main idea is not that the model just "tells me what I need to learn", but that the model can combine ideas, turn them over, split them apart, recombine them, and look at them through many different semantic lenses way quicker than I could do alone.
@T33K3SS3LCH3NАй бұрын
At their best, they are a good addition to a search function. Giving summaries and sometimes picking out just the right tidbit, like googling for a stack overflow answer but not actually having to go through the results manually. But that's about it.
@NoBoilerplateАй бұрын
yup!
@dominick253Ай бұрын
I keep hearing the same complaints about AI. Personally I use it to generate boilerplate and then go through line by line and analyze what everything's doing. It's infinitely quicker than typing out 300 lines of code by hand. Are you really have to do is go through and make sure all the logic lines up with what you want. Maybe for someone like primeogen who's been coding for 30 years it's quicker just to type it by hand. But for a junior engineer it's much quicker and easier to use AI and then patch up whatever little mistakes it makes. Also I'm wondering are people trying to use AI for a whole project? In my experience in only works well at one objective at a time.
@alexlowe2054Ай бұрын
Senior developer here. Thank you for helping provide me with a job for the next several decades. I'm the experienced person they bring in to clean up critical bugs deep in the software, that require deep knowledge of the languages, libraries, and runtimes involved. Every single time I touch a codebase, there's always some critical bug around data concurrency, memory management, type system semantics, security vulnerabilities, or some other problem involving a lot of crunchy algorithm knowledge. There is a zero percent chance that an AI understands the underlying memory model or algorithm design or language specification well enough to write code that's not going to completely fall apart when it's thrown into the real world. So, thank you for helping keep me employed. I'm not a fan of dealing with large legacy codebases, but it pays well enough, and those codebases will be around forever. As long as contractors are prioritizing deadlines, and as long as developers are using AI and "touching it up", I'll have a job rooting out critical bugs that require a deep knowledge of core computer science topics. If you don't want to be in the business of helping create jobs for more senior developers, put in the work. It's what Primeagen preaches. He's a good developer, but you don't need to write code for 30 years to be that fast on the keyboard. If you want to vomit out 300 lines of code, just get good with your tools. All it takes is a little bit of time and a ton of discipline. Take a few months and drill your keyboard shortcuts. Move your mouse to the opposite side of your desk, and only reach for it if you spend at least a minute fumbling through your shortcuts. Even better, learn about alt+tab, and google the keyboard shortcut you need before reaching for your mouse. You'll know you're done practicing the first time you feel like it's too much effort to move your hand to your mouse. Do that practice for even just a few weeks, and you'll be twice as fast on the keyboard. Do some typing drills, and work on your muscle memory. A few hundred lines of code isn't anything difficult. It's surprising and pathetic how often developers can't write code. "Why Can't Programmers.. Program?" is getting close to 20 years old at this point. Don't be like that. Be better. After your muscle memory is good enough that your hands magically start converting your ideas into character in your IDE, work on algorithms and design patterns. And just write a ton of code. Practice toy problems, and build some real world applications. There's no experience more valuable than sitting down and trying to create something yourself, without any notes to copy. If a year of programming practice sounds tough, remember that some fields require 7 years of school. You can either spend 30 years not caring and slowly learning through accidental lessons, or you can dedicate a year or two and really master your craft. Go read the top posts on classic blogs like Joel on Software and Jeff Atwood's Coding Horror. Dig incredibly deeply into the crunchy theory of computer science, and binge through videos like "Parsing JSON Really Quickly: Lessons Learned" or "Performance Matters by Emery Berger". Both those talks cover the complexities of modern programming, rather than the overly-idealistic tutorials that teach the average developer. Read all the hundred page technical specifications for your language. Watch the deep dive tutorials on obscure features, and then go build a real application with your newly found knowledge. Dig as deep as you can into the crunchiest subjects you can, and you'll be rewarded by having more knowledge than most of the other developers on your team. Don't shy away from challenging yourself while you're learning, and you'll be ready to meet any challenge in the real world. The more you practice, the more you learn. The inverse is also true. Primeagen stopped using AI code completion tools because they were stealing the valuable time he used to practice writing code while he was working on real problems.
@zoeherriotАй бұрын
Yeah, just make sure you are learning your craft. The thing with this is that the act of writing code is what creates the memory, the experience, the knowledge - if you had that over to a tool - you may find that you're actually not as good as you feel you are at programming. Turn off the tool for a bit - make sure you actually know what you are writing. As prime mentioned, when he turned it off, he realised he was constantly waiting for the editor to complete his code, and that he was effectively doing himself dirty by not writing the code himself. That doesn't mean you shouldn't use it - but just make sure you aren't losing opportunities to learn by handing the wheel to the AI.
@JamEngulferАй бұрын
@@alexlowe2054 You know you can give advice without going on and on about how much better you are than people less experienced than you, right?
@buzhichun6 күн бұрын
I'd like to add one more thing. Just like how the real winner of a gold rush is the pickaxe salesman, the companies providing the computing infrastructure that actually trains and runs these giant machine learning models very much benefit from keeping the hype going. Those coincidentally also happen to already be some of the largest companies on the planet, with giant lobbying and PR operations. "Look at these amazing nuggets, there’s gold in them thar hills!" - MSFT
@cherubin7thАй бұрын
Shows how much structure and information our language in itself contains.
@NoBoilerplateАй бұрын
HUGE amounts! But you know what else contains structure and information? Structure and information 😉
@hexpulse23074 күн бұрын
I’m an eye surgeon and I have been experimenting with using gpt 01 as like another “doctor” To run my ‘more unusual” cases by - and it does seem to understand me as well as most other specialists and can help me confirm what I already know but may be forgot since it’s been a long time since I had a similar case etc - it can help interpret eye imaging too - not perfect yet but neither are other surgeons - it’s not falling apart but it definitely helps to provide as much info as I would another doctor without leading -
@clashgamers4072Ай бұрын
I have a feeling even if there were a lot of phd lvl resources , Current LLM's will still struggle. "More specific Less usefull" - This is subjective ,for ex in generating code the more specific the prompt better the results for me.
@NoBoilerplateАй бұрын
I was worried that title was going to be confusing, sorry. It should read (and I hope the context of what I said over it makes clear): "More specific requests, less useful responses"
@alexhope212009Ай бұрын
Actually one thing I was impressed with lately is using it to design a database schema and basically organize my messy human thoughts and eventually have it write the actual code for the entities and their configuration. Previously this is something I remember took a lot more effort to organize my thoughts and make sure relationships make sense etc
@NoBoilerplateАй бұрын
yep, easy stuff is easy, good for boilerplate, but 90% of the work comes in the last 10%
@charactername263Ай бұрын
As a software developer working on pretty low level and template heavy code, AI (I've only used GPT-4o) is very good for a few quite specific use cases: 1. Cleaning up and summarising verbose compiler error logs which can be hundreds of lines for a single issue. 2. Generating very specific functions and/or metafunctions with a clearly defined set of inputs and outputs and how it should transform them. Trying to work with multiple successive prompts never really works, it's better to pessimistically assume the AI will not keep any knowledge of past prompts and just be happy when it remembers something useful from the session.
@ai-aniverseАй бұрын
this.
@premnagarathnam27013 күн бұрын
@@charactername263 i recommend trying claude. It keeps context far better than 4o. The tradeoff being hourly message limits + it eventually says a chat is too long (has too much context)
@copperknight47889 күн бұрын
I like using chat gpt to find studies on stuff. You can ask it on a topic, get an approximation of an awnser, and then ask it to look for a source to back up that claim.
@thoperSoughtАй бұрын
I was testing Claude out by asking it to write a function in Haskell, and it did surprisingly well-BUT it suggested using record syntax to change the month in a Data.Time Day. I told it that was clever before finding out it didn't compile, because Day *isn't* a record. it corrected it to something that worked. later, in the same chat, I asked it to make a change to the function, and it tried to do *the same thing,* ignoring the correction, I guess? it's interestingly clever, while at the same time being interestingly stupid
@invven2750Ай бұрын
@@thoperSought long term memory and parallel thinking are knowingly some of the biggest gaps of current AI. For the memory part the companies itself put limits for chats
@thoperSoughtАй бұрын
@@invven2750 this makes me wonder a couple of things: the model they let free accounts use changed between the first and the second parts of that chat-would it have made a difference if the model stayed the same, or is it just not taking the whole chat?
@Robin_GoodfellowАй бұрын
It's been fun to use it as a language learning tool. As you say, it understands language really well, with increasing quality the more common the language is. I should really try it on some obscure language sometime, maybe a conlang. Man, imagine talking to an LLM in Sindarin.
@NoBoilerplateАй бұрын
ha, amazing! Maybe something big like lojban might be in there, maybe even toki pona? But the less information there is, the worse it is 🙃
@ErikFromCanadaАй бұрын
I'm glad you brought up blockchain as a point of comparison. Through this whole hype cycle, I've often been reminded about that bike share program (I believe it was in the Netherlands?) that was built "on the blockchain". It used the headline hype to gain funding and ran really well. The more credulous press looked into it later down the line, and the devs were upfront and honest that the software only used the blockchain in a purely incidental way. It was actually completely ordinary, boring, functional, valuable software. If we must live with AI hype, I hope it can be in that manner.
@mikkelensАй бұрын
the zed code editor uses the ai selling point and I hope they just use it as that only. There’s a lot of promise in how the rest of the product works (performance and live-multi-person editing support) and I hope that in the end we get that with a little optional chatbot tucked away somewhere. Noone needs these things to be more than chatbots and slop generators.
@jonessiiАй бұрын
7:10 Funny, I've been thinking about this lately as well and it's pretty much the exact same conclusion I came to as well. It can fool enough of the right people enough of the time. Just enough to extract large quantities of investor cash. For how long?
@NoBoilerplateАй бұрын
ugh not long now I hope
@ccl1195Ай бұрын
You are a smart fellow. I appreciate this take and very much agree with it. I have been criticizing GPT since it was rolled out, and if you say anything out loud, a bunch of idiots come for your comment. I didn't know how these things worked at first, but then I- oh, I don't know? Actually studied what they are and how they work? 🙄I came to the conclusion that this technology alone is very unlikely to ever become Data from Star Trek, which is what they want you to think it is. It's actually more like a (somewhat altruistic) superficial sociopath with no need to breathe, and thus can deliver more lies-per-minute than a biological system cursed with the need for oxygen. Claude is frequently a better model in this department for its more "responsible" tone, whereas GPT will lie, and lie about its lies, using all kinds of weasel words and passive voice to avoid responsibility and obfuscate the issue. I am a solo games designer and have had to use these tools to help teach myself C# this year, but I find it often much better to simply buy and read books. While I generally agree with the video that these LLMs are great for broad or superficial, widely accessible information, what I actually find more common from GPT is to deliver a mostly accurate essay on some aspects of coding, and yet then stealthily bury in the middle of that essay something that's wholly and fundamentally untrue, even dangerously wrong. It makes you suspicious that perhaps these things aren't even mostly altruistic.. I was only able to start spotting these inadequacies once I reached intermediate proficiency with C#. I can now recognize how they use common code examples and patterns I've seen online (for both C# and Unity) and shoehorn them into all kinds of inappropriate use cases, as my queries get more and more specific. These tools are extremely dangerous to a junior student trying to learn to program. I've said that over and over and the comment vultures always say "Well you can't *only* use GPT," as if it's some kind of catch-all, obvious answer. It is not. 1) Look around and you will see how many people are almost only using GPT. 2) Even if one is not, you *cannot* fact check and correct its lies when you "don't know what you don't know."
@NJ-wb1czАй бұрын
I think humanizing it in any way is misleading. It's not benevolent, it's not non-benevolent, it has no motivation. Not neutral motivation - no motivation. It's content without any creator, a regurgitated content. Like an algorithm that takes text amd randomizes words in it - is it good? Is it malevolent? Is it telling the truth? Is it lying? No, and approaching it this way makes no sense. The questions themselves make no sense
@gavros9636Ай бұрын
@@NJ-wb1cz It does one thing and one thing only, it looks at a text output and guesses at what word comes next in sequence. That's it, it is really good at that one thing but it is not AGI, as AGI requires either A, a fully simulated human brain or B, creating something capable of self awareness which right now is as of yet not possible, the agents we have today are still specialized to the specific tasks they're given while an AGI should be capable, like a human child, to figure out any new task presented to them and be able to perform it relatively well without losing efficacy in other tasks.
@NJ-wb1czАй бұрын
@@gavros9636 we don't know what self awareness is, and don't know if a model of a brain would have it. All of this AGI talk is highly made up and speculative When it comes to the current models and the way they work, all they need is to completely change the training and the training algorithms, and essentially raise millions robotic babies how they would raise a human one, for the same amount of years, with the same effort to provide thwm the same social experiences in different conditions etc. Have them process at least the same inputs human brain processes, preferrably better ones. All the smells touches sounds visuals, etc. I don't think that's in any way feasible in the foreseeable future And the theoretical discussion about "self awareness" and "AGI" don't matter. These are abstract fantasies, not actually anything tangible.
@thorn9382Ай бұрын
Chat gpt is really good when you forget what something is called or want help with syntax in a programming language you aren't familiar with, but if you ask it to write an entire script there's a good chance it will completely fail.
@AKU666Ай бұрын
I think it's pretty same story with humans... They also approximate stuff that they learn. Many things in this universe is so complex that you barely can know everything about some subject/topic on all levels of abstraction. And even if we know everything on some level of abstraction (like math because it's defined) we don't know everything on another levels that we apply math to. For example physics. Take my comment with many grains of salt.
@NoBoilerplateАй бұрын
LLMs are a huge step towards making computers more accessible, but language ability is only part of the problem
@schmetterling4477Ай бұрын
Discrete math is a trivial example of knowledge that is limited to a null-subset of the entire domain. That is basically what Goedel found out. ChatGPT can't even do math at the high school level reliably, let alone get around Goedel.
@bobnolin9155Ай бұрын
Maybe a better a name for it would be Artificial Average Intelligence.
@_dot_Ай бұрын
all that about ai startups is true chatgpt is actually pretty decent at math now tho. i often ask it for help with my uni test questions and while it does make mistakes from time to time (like adding a minus or something) it's generally accurate, even with more specific and complex stuff. I think what it does differently is it generates invisible text that helps what is visible be more accurate (for example extra tiny incremental steps for the math questions) and maybe it has an interface to a calculator too (so, making this up, say, if it returns "[=30/6]" the system would replace that with the result before continuing to generate new data. knowledge wise it's also gotten better. it now looks up what you are asking on google, evaluates the reliability of the results and takes input from them. it's quite impressive
@MatthijsvanDuin29 күн бұрын
"maybe it has an interface to a calculator too" ... if it's giving correct answers for math, that's without a doubt what they've done: recognize math and hand it off to something that can do math, something that's _not_ an LLM.
@lilylyons8885Ай бұрын
I think you explained perfectly why I struggled to use something like copilot. Its like autocorrect trying to correct a word its never heard before- I have to frustratingly delete the change it made only to see it make the exact same mistake a minute later I work with eccentric fields of programming and copilot would constantly generate nonsensical junk or outright duplicate what I'd already written. It certainly worked better when I was working on my website, but in the end I turned it off after 30 minutes of leaving it on.
@смрш21 күн бұрын
In my experience chatgpt is particularly great at one thing: writing draft shell programs that I can then edit to fit my needs and translating error messages into something I can understand which makes troubleshooting MUCH easier
@kpunkt.klaviermusikАй бұрын
1+1=2 is an equation. 2e^2+5j=0 is a curve. Math programs should be able to identify functions pretty easy. Just use the right program for the right task. But when we talk about AI we expect real understanding - not doing predefined tasks.
@MagicGonads24 күн бұрын
1+1=2 is an identity, both are equations
@roberteltze485024 күн бұрын
To an electrical engineer 2e^2+5j=0 is simply a false statement. e is Euler's number and j is the square root of -1. There are no variables just constants that didn't add up to zero.
@MagicGonads24 күн бұрын
@@roberteltze4850 it's an equation which requires a model to map the constants/variables to objects (though the quantifiers are omitted, so also the domains of quantification), the equation has a set of solutions which is empty in the model where e is fixed at euler's number and j is fixed at the imaginary unit, but is a curve in a model where e and j are each quantified over the reals
@theseriousprepper437213 күн бұрын
The algorithm found your channel. Very interesting. I was using ChatGPT 4 yesterday to compose a photograph and when I asked for a broad photograph, it provided the information. A little overwhelming with visuals, but essentially did the basic job. The problem is when I went to edit. It wouldn’t specifically remove items that were in the photograph. So I had to redo each photograph with some subtle changes. Still never reaching the ultimate goal of what I was looking for, also using the verbal ChatGPT getting information on avian flu. I asked for it to copy all the information we covered in the chat and it said it did, but when I went back in, it did not. Being gaslit by AI is dangerously humorous. I felt like I was having a conversation with a multiple-choice test. Just saying.
@erniea5843Ай бұрын
Imagine paying $200 /month for access to one 😂
@drantino2 күн бұрын
for a few months now, ive specifically noticed that LLMs are not the right direction for AGI, and one thing that this video directly points out actually gives me the right words finally to properly convay why, LLMs auto complete in such a way that its actually having a case of not giving a understanding of the concepts that its feed. yes it could easily spit out the definition of "The", but it doesnt actually understand what "The" is in any given context. for computers to understand code, it needs to know how to interpret commands, so something like "if" statement has a layered defined description on how it exists and used in code for it to actually do something with.
@DaveParrАй бұрын
3:17 isn't the solution to this limitation RAG + agents + tool use? Perplexity search engine seems to prove this. Its why shipping larger context windows was a major focus of llm improvement for the last few years.
@NoBoilerplateАй бұрын
GPT is best suited at the human/computer interface, and we've seen really good uses here. BUT the claims suggest that anything is possible, which is simply not true
@valex1147Ай бұрын
Thank you!! I am glad you are back with another video. Hope you are well and looking forward to the next one.
@RaxisАй бұрын
LLMs are like librarians that have read every book in the library. Excellent for summarizing broad swaths of information but you shouldn't trust the librarian to build your Linux system when they're telling you to run `chmod 777 /` Ironically, having to fix and overcome the problems caused by blindly listening to the LLM has actually made me much more proficient as a user so LLMs are a great way to blindly charge into something that one would otherwise be anxious to start
@geckoram628619 күн бұрын
My favorite use of chatgpt, apart from language-related stuff (it's pretty good at translating things!) is to break down topics I'm not familiar with. The other day I spent like 3 hours diving into building an EME station with reused electronics. I'm no electronics engineer, nor do I have a ham licience (yet), so I asked chatgpt, and it gave me some important keywords and topics I could dive further into, in other, human made websides. I wouldn't trust chatgpt for a specific value of inductor for a colpitts oscillator, but it's useful to know that a radio needs and oscillator, to read somewhere else about them
@maybethisismarqАй бұрын
While LLMs may not be great at reasoning and suffer from hallucinations, I find it invaluable for summarizing long research papers, creating outlines, brainstorming, and helping me express my thoughts when I’m having trouble finding the words. I wish companies would advertise these advantages instead of promising stuff that isn’t true.
@ruanmed28 күн бұрын
@@maybethisismarq The companies actually selling LLMs do advertise those features to make sales, e.g.: Microsoft has auto summarization as a product of their Teams platform, if a meeting is recorded and you pay for the Teams "Premium" you will get a summary of everything discussed in the meeting afterwards, it's a pretty good feature, and it's multilingual. But yeah, you won't hear the "AI" companies seeling this feature because when you think about it it's a feature that already needs a platform to be useful, or do you think if OpenAI started a Teams competitor now every company that already have contracts with Microsoft would migrate it to OpenAI? Most of the AI companies are advertising features that don't exist because they are selling the idea of those features to their investors as a miracle that will bring 100x more profit, if they were to present plans for actual things LLMs excel at investors would realize most of the AI companies don't have the platforms to apply the AI to, the market is already capped for those and the companies would actually have to have a business plan for this, and when investors see a 1.5x return on profits over 2 to 5 years they would not be interested. On the other hand, selling a promise for something magical in just some years that will get them 1000x ROI, on boy, they really want that... All in all I can see that Microsoft is one company that's actually integrating LLMs into useful things and getting real money from it. Facebook has been great due to their Open approach to releasing the Llama models that's sparked more open models approaches from other companies (Like Qwen from Alibaba and Exaone from LG).
@4thorder28 күн бұрын
I agree, and not just that, but their ability to code is extremely helpful. It can turn someone who doesn't have a previous background in the syntax of a new language an edge like never before.
@NoBoilerplate28 күн бұрын
Right! Language ability is SUCH a killer feature, they don't need to invent other stuff too!
@0x15aac4 күн бұрын
The problem with using LLMs for high-level initial exploratory work is they *constantly* get stuff wrong. And by the nature of the tasks you're using them for, you don't know enough to tell that it's wrong.
@khhnatorАй бұрын
i think the real bad case against current AI is not even the fact certain knowledge is obscure, at some point it might just get developed enough to say "i don't know that" the issue is that Language is not everything, there are LOTS of things that are know. but are poorly expressed thru language. a big example, posing. we have no language for posing. some poses get names, like the super hero landing or the Marilyn Monroe. but a photographer work is literally a set commands "raise you arm", "turn around", "look up" is just a bunch of very vague instructions. and the more you try to be specific, the more you end with deep fried outputs.
@Demopans5990Ай бұрын
道可道,非常道。 名可名,非常名。
@NicholasGuyett29 күн бұрын
"It is always easier to promise a bright future than to build a better present" An excellent line
@filosophikАй бұрын
"Pensive blues"... "Hopes echo" You're right, that's golden thesaurus-ness
@nyarparablepsis87212 күн бұрын
I am in no way a coding person, instead earning my crust with ancient languages and history. After listening to other grad students who are doing A LOT of their work with the help of "AI", in a moment of deadline despair I asked ChatGPT to help me write a short intro to the topic I was working on. The result was ...bad. I then asked it about the language that I am specialising in (Sumerian), and what it spat out was HILARIOUS. Since then I have felt very, very safe whenever people go "oh, AI will do your work soon".
@FireyDeath4Ай бұрын
AI sucks because the frameworks they're using are crude, and they're using massive knowledge-as-power models because that's what's easy for them. I think, once they make the leap from abstract neural intuition to formalised framework processing, then we can worry. And marvel again (It's still pretty good but we know the limitations now)
@4thorder28 күн бұрын
I found your comment interesting, so I asked ChatGPT o1 full model, - what does this mean to you? "the leap from abstract neural intuition to formalised framework processing" Its response, "this phrase often refers to the process of moving from a raw, intuitive understanding of something-like the way a human brain instinctively “gets” a concept-into a more rigorously defined, structured, and often mathematically or logically framed model. It’s the shift from “I know this pattern when I see it” (an abstract neural sense) to “I can describe and manipulate this pattern using a systematic set of rules, equations, or algorithms” (a formalized framework). In other words, it’s the journey from having a vague, gut-level feeling that something is true or meaningful to actually translating that feeling into a precise representation that can be analyzed, tested, and reliably applied." LOL good phrase!
@science7590214 күн бұрын
Regarding LLMs and maths: you could ask it to use code interpreter. Then it most of the time works. (But haven’t tested it for complex math problems)
@coldcircuit99Ай бұрын
Today tried to force it to give me anything about webauthn stuff in rust and it gave me halucinated responses 100% of the time :d
@NoBoilerplateАй бұрын
You know that guy who likes to seem knowledgable, and always has an answer for everything, even if they have to invent one? #GPT
@EpicMiniMeatwadАй бұрын
@@NoBoilerplate The world slumbers when 0 = 0, but when 0 = 1, everyone loses their shit.
@okie9025Ай бұрын
I believe the problem of hallucinations is because chatgpt isn't actually an "AI assistant", it's merely simulating one. At the beginning of each chat it's explicitly told "You are 'a smart AI assistant'. Answer all of the user's queries following these rules: . Here's the user's query: ..." and then tries to answer what it _thinks_ a smart AI assistant would answer. Obviously it can't perfectly mimic an AI assistant because its knowledge is limited. It's like asking a friend "hey, act like you're a professional electrical engineer and try to answer all my questions". If you get a "hallucinated" answer, it's not because your friend is confident in their answer and just acting weird, but because they think that such an answer sounds _close enough_ for a professional electrical engineer to say.
@cuckoophendula821128 күн бұрын
That graph makes me think of what Rick said about using the Meeseeks box. "Keep it simple they're not (burp) gods."