Has Generative AI Already Peaked? - Computerphile

  Рет қаралды 1,072,244

Computerphile

Computerphile

Күн бұрын

Bug Byte puzzle here - bit.ly/4bnlcb9 - and apply to Jane Street programs here - bit.ly/3JdtFBZ (episode sponsor). More info in full description below ↓↓↓
A new paper suggests diminishing returns from larger and larger generative AI models. Dr Mike Pound discusses.
The Paper (No "Zero-Shot" Without Exponential Data): arxiv.org/abs/...
/ computerphile
/ computer_phile
This video was filmed and edited by Sean Riley.
Computer Science at the University of Nottingham: bit.ly/nottsco...
Computerphile is a sister project to Brady Haran's Numberphile. More at www.bradyharan...
Thank you to Jane Street for their support of this channel. Learn more: www.janestreet...

Пікірлер: 3 600
@Computerphile
@Computerphile 7 ай бұрын
Bug Byte puzzle here - bit.ly/4bnlcb9 - and apply to Jane Street programs here - bit.ly/3JdtFBZ (episode sponsor)
@worldofgoblins
@worldofgoblins 7 ай бұрын
Could you explain what “There exists a non-self-intersecting path starting from this node where N is the sum of the weights of the edges on that path” means? Is the end node for the “path” one of the purple nodes?
@MacGuffin1
@MacGuffin1 7 ай бұрын
Humans do fine with less data, volume of data is clearly not the issue. 'Has generative AI already peaked?' : Not even close...
@mr_easy
@mr_easy 7 ай бұрын
@@worldofgoblins Yeah, same doubt
@squirlmy
@squirlmy 7 ай бұрын
B​ut things like LLMs don't at all work like human intelligence. That's like saying "people make ethical decisions all the time, so my pocket calculator should have no problem with issues of morality."
@paulmichaelfreedman8334
@paulmichaelfreedman8334 7 ай бұрын
@@MacGuffin1 There's still ways to go, but I do think there is an asymptote. But not before this type of AI has become as useful and intelligent as the droids in Star Wars. But not much beyond that. It remains an input-output system, defeating any chance of ever evolving to human-like emotions, for example. For human-like AGI, I think Ben Goertzel is much in line with that and he says that is still quite some time away as such an AI is radically different from transformer and generative AIs.
@miroslavhoudek7085
@miroslavhoudek7085 7 ай бұрын
As a sort of large trained model myself, running on a efficient biological computer, I can attest to the fact that I've been very expensive over the decades and I certainly plateaued quite some time ago. That is all.
@ExecutionSommaire
@ExecutionSommaire 7 ай бұрын
Haha, I totally relate
@avi7278
@avi7278 7 ай бұрын
fr fr
@SteinGauslaaStrindhaug
@SteinGauslaaStrindhaug 7 ай бұрын
@@avi7278 Apparently youtube thinks "fr fr" isn't English so it offered to translate it… it translates to "fr fr" apparently 🤣
@AnimeUniverseDE
@AnimeUniverseDE 7 ай бұрын
I get that you were just making a joke but the current version of AI could not be further from humans.
@salasart
@salasart 7 ай бұрын
XD This was hilarious!
@marcusmoonstein242
@marcusmoonstein242 7 ай бұрын
You've just described the problem being experienced by Tesla with their self-driving software. They call it "tail-end events", which are very uncommon but critical driving events that are under-represented in their training data because they're so rare. Tesla has millions of hours of driving data from their cars, so the software is better than humans in situations that are well-represented in the data such as normal freeway driving. But because the software doesn't actually understand what it's doing, any event that is very uncommon (such as an overturned truck blocking a lane) can lead to the software catastrophically misreading the situation and killing people.
@bened22
@bened22 7 ай бұрын
"Better than humans" (X)
@andrej2375
@andrej2375 7 ай бұрын
It's better than humans AND we'll keep working on it
@James2210
@James2210 7 ай бұрын
​@@bened22It's like you commented without actually reading what it's saying
@bened22
@bened22 7 ай бұрын
@@James2210 I read it but I don't even believe the softened claim.
@ids1024
@ids1024 7 ай бұрын
In theory, an advantage of self-driving cars *could* be that the software has "experience" with many of these uncommon situations that few human drivers would, which could save lives when a quick reaction is needed, or the best response is something most people wouldn't think to do. But that may be decades away still.
@Posiman
@Posiman 7 ай бұрын
This is the computational side of the argument for AI peak. The practical side is that the amound of existing high-quality data in the world is limited. The AI companies are already running out. They theorize about using synthetic data, i.e. using model-generated data to train the model. But this leads to a model collapse or "Habsburg AI" where the output quality starts quickly deteriorating.
@ShankatsuForte
@ShankatsuForte 7 ай бұрын
This isn't even remotely true. Microsoft has already trained a GPT 3.5 tier model entirely on synthetic data.
@chja00
@chja00 7 ай бұрын
I absolutely love the fact that someone named it Habsburg AI.
@ash7324
@ash7324 7 ай бұрын
CPT 8 is going to have a big weird jaw and a limp
@funginimp
@funginimp 7 ай бұрын
The only reason synthetic data works in practice but not in theory is because it makes up for the transformer algorithm not being optimal and other training data being worse quality.
@nathanshaffer3749
@nathanshaffer3749 7 ай бұрын
Personally, I believe that we have to improve the quality of data. But we are no where near the quality that humans train on. We percieve the world in 4 dimensions (3 spatial + time) and have additional context layered into it such as sound and touch. The difference between observing pictures of trees and being able to walk around a tree every day as it grows. Being able to watch the way it moves in the wind, seeing the difference in lighting throughout the day is orders of magnitudes of data quality vs seeing static 2d images of trees. This of course completely ignores all of the human to human interactions involving direct instructions as well as being able to ask questions about your world to to get specific information. imagine living your whole life floating in empty space and being taught through a series of pictures paired with words. No instruction on how to read. Just here's a picture and here is it's description, you work it out. Maybe you have continuous access to all of the data to be able to compare and contrast so you don't have to rely on the human memory to analyze it. Even given an immortal human, I think we would plateau pretty quickly. Any critical thinking that this human acquired, I feel would be attributed to a base level biologic programming that we are born with. Which is not much I think.
@tommihommi1
@tommihommi1 7 ай бұрын
generative AI has destroyed internet search results forever
@vincei4252
@vincei4252 7 ай бұрын
Nah, Google did that because of their greed.
@priapulida
@priapulida 7 ай бұрын
@@vincei4252 .. and because they are woke (edit: I thought this is obvious, but apparently not, use the simple prompt "Google and DEI" to get a summary)
@Alice_Fumo
@Alice_Fumo 7 ай бұрын
Come on, any 100IQ+ human with half an hour of time could figure out how google or whomever could largely fix those issues if they really wanted to. Also, the grapevine says that OpenAI search gets announced next monday so maybe there'll be some competition finally. Take buckets of salt with this though, I don't know where I heard it, but I'm quite sure it wasn't a trustworthy source.
@tommihommi1
@tommihommi1 7 ай бұрын
@@yensteel the point is that trash generated by AI has flooded the results
@no_mnom
@no_mnom 7 ай бұрын
Add to the search before:2023
@Rolox01
@Rolox01 7 ай бұрын
So refreshing to hear grounded academics talk about these sorts of things and take realistic look at what’s happening. Feels like everyone wants to say anything about generative AI
@Dino1845
@Dino1845 7 ай бұрын
It's only now I feel we're learning of the monstrous cost & technical limitations of this technology now that we're past the initial hype.
@harrylane4
@harrylane4 7 ай бұрын
@@Dino1845I mean… people have been talking about that since the start, you just weren’t listening
@the2theonly672
@the2theonly672 7 ай бұрын
@@harrylane4exactly, you can’t clickbait that like you can with “this crazy new AI will take your job”
@snickerdoooodle
@snickerdoooodle 7 ай бұрын
People can still talk about the ethics and ramifications that AI has on the human element without your permission, just saying.
@the_mad_fool
@the_mad_fool 7 ай бұрын
It's because all the crypto grifters jumped onto the AI bandwagon, so there's just a ton of scammers and liars flooding the air with their phony claims. Reminds me of back when stem cell research was the next big medical thing, and suddenly people were coming out with "stem cell anti-aging cream" made from "bovine stem cells."
@LesterBrunt
@LesterBrunt 7 ай бұрын
People just completely underestimate how complex human cognition is.
@BrotherTau
@BrotherTau 7 ай бұрын
It's complex because we don't understand it.
@Chris-fn4df
@Chris-fn4df 7 ай бұрын
@@BrotherTau It is complex because it is full of variables. That is what complexity is. There are many things described as "complex" that we have excellent understanding of.
@BrotherTau
@BrotherTau 7 ай бұрын
@@Chris-fn4df Point well taken. However, my point is rather that human cognition and consciousness could have a very simple explanation. We think it's "complex" because we don't understand it.
@murathankayhan2312
@murathankayhan2312 7 ай бұрын
@@BrotherTau nah. human cognition and conscousness will never have a "simple" explanation. even though they will have an explanation in the future, it will not be a "simple" one. you are underestimating our brains.
@patrickirwin3662
@patrickirwin3662 7 ай бұрын
​@@BrotherTau "We"? The most unexamined term used by the true believers in scientism. What's trying to understand human consciousness and cognition is human consciousness and cognition. When your screw turns itself without a driver, when your cornea sees itself without a mirror, when your tastebuds know what they taste like, you will understand it all.
@leckst3r
@leckst3r 7 ай бұрын
10:37 "starts to hallucinate" I recently heard it expressed that AI doesn't "sometimes" hallucinate. AI is always hallucinating and most of the time its hallucination matches reality/expectation.
@sebastiang7394
@sebastiang7394 7 ай бұрын
yeah but the same could be said of humans to some extend. We always have our own model of the world that is flawed and doesn’t match reality perfectly.
@fyang1429
@fyang1429 7 ай бұрын
AI is just a very very cleaver Hans so that does make sense
@drew1564
@drew1564 7 ай бұрын
If a hallucination matches reality, it's not a hallucination.
@fartface8918
@fartface8918 7 ай бұрын
​@@drew1564untrue
@Brandon82967
@Brandon82967 7 ай бұрын
Not true. Someone told GPT 3, which is way worse than GPT 4, that it can call the user out if they asked a nonsense question. It was able to answer sensical questions like "who is the 40th president of the US" and called the user out when they asked nonsense like "when is the spatula frighteningly spinning"
@Spnart
@Spnart 7 ай бұрын
You don't understand, a bunch of people with finance degrees on reddit told me General AI is just around the corner because ChatGPT responded in such a way that it "held a conversation with them therefore its alive". It's almost like the less familiar you are with computer science and fundamentals, the more you treat technology as a form of magic.
@mrosskne
@mrosskne 7 ай бұрын
If something seems intelligent to you, it is. There is no experiment you can perform to show otherwise.
@benjaminloyd6056
@benjaminloyd6056 4 ай бұрын
"Any sufficiently advanced technology is indistinguishable from magic"
@Darca1n
@Darca1n 4 ай бұрын
​@@benjaminloyd6056And amusingly the less you understand technology on a basic level, the lower the bar for "That's not technology, it's magic" gets.
@jared_bowden
@jared_bowden 4 ай бұрын
I got a STEM degree not too long ago and I'm genuinely surprised how, eh, "not technical" alot of people in finance and business are, even though their jobs often demand technical thinking. However, the news I've heard is that even Wall Street is starting to see "through the smoke veil" and realize a) generative AI might reach a plateau, and b) even if it doesn't, how exactly are we going to make money off of this? Like, they don't just need to make money, they need to beverage a Ton of money to recoup costs, and it turns out generating money is harder than cat images.
@TehIdiotOne
@TehIdiotOne Ай бұрын
its almost as if LLM are literally designed to give you the illusion they are intelligent.
@3dartxsi
@3dartxsi 7 ай бұрын
Years ago, all the talk about AI was that anything resembling proper "strong" AGI was likely decades off, if we ever managed to achieve it. This is largely due to A.) limitations imposed on computer hardware(as currently envisioned) by the laws of physics themselves, & B.) the fact that we didn't have a full understanding of how a human brain works, limiting our ability to replicate it in any functional way Suddenly, silicone valley is selling AI as if we've reached that level, even though neither of the previous considerations have been addressed. This would be like someone you know discussing how they want to take a year off to go to Europe, despite not having the money to pay for the trip, or having a valid passport and then suddenly they are saying how they've bought plane tickets and will be leaving the country next week, even those previous issues were never dealt with.
@thiagopinheiromusic
@thiagopinheiromusic 6 ай бұрын
You make a very compelling point about the current state of AI development versus the hype surrounding it. The sudden surge in AI capabilities and the way it's being marketed can indeed seem premature, given the unresolved foundational challenges.
@bluedistortions
@bluedistortions 4 ай бұрын
I remember about 12 years ago, the cover of some big tech magazines were boasting that computer scientists had modeled a snail's brain. Reading for more details, 1. A snail doesn't have anything we can really call a "brain," it's a bundle of nerve fibers, very limited in functionality. 2. It took the full resources of one of the biggest computational super centers in the world to do so, and consumed a town's worth of electricity to do so. 3. And it could only do it at 1/4 speed. Really should make us a bit more humble about our technology and more in awe of nature.
@granienasniadanie8322
@granienasniadanie8322 13 күн бұрын
@@bluedistortions NOw they made a fly brain on a laptop, 6x more neurons but it took 10 years so still no revolution.
@michaelujkim
@michaelujkim 7 ай бұрын
even if you took the whole internet as a dataset, the real world is orders of magnitude more complicated.
@B.D.E.
@B.D.E. 7 ай бұрын
A simple but very important point that's easy to forget with all the optimistic ambitions for AI.
@mharrisona
@mharrisona 7 ай бұрын
I appreciate your comment sir
@dahahaka
@dahahaka 7 ай бұрын
Which is part of why people are working on merging robotics and ml However nobody is trying to let these things train on the real world, quite the opposite, turns out that training on the "whole internet" is vastly more efficient and transferable zero shot into the real world without any problems It's funny how you assume humans have perfect models of the world... You just need very very rough approximations
@Cryptic0013
@Cryptic0013 7 ай бұрын
Yup. Look at the behaviors and beliefs of human beings who are, as they say, chronically online. Getting 100% of your information from the internet's not a great anchor to reality
@freezerain
@freezerain 7 ай бұрын
But an AI should not be perfect, just be better then an average human. If dataset will contain all image recording, all books and movies and music all news and tiktoks and comments this is could be enough to be better then human in some tasks
@RobShocks
@RobShocks 7 ай бұрын
Your ability to articulate complex topics so simply with very little cuts and editing adlib is amazing. What a skill.
@philippmillanjochum1839
@philippmillanjochum1839 7 ай бұрын
Facts
@Pabliski577
@Pabliski577 7 ай бұрын
Yeah it's almost like he talks to real people
@bikerchrisukk
@bikerchrisukk 7 ай бұрын
Yeah, really understandable to the layman 👍
@RobShocks
@RobShocks 7 ай бұрын
@@docdelete Teaching you say?! What a cool concept. Thanks for sharing.
@tomsenkus
@tomsenkus 7 ай бұрын
He’s obviously AI 😮
@alan1507
@alan1507 4 ай бұрын
Your point about under-representation in the training set is spot on. In an experiment to chatGPT I gave it a. poem I'd written to analyse, and I asked it to identify what type of poem it was. It stated that it was a Shakespearean Sonnet and quoted the rhyme scheme ABABCDCDEFEFGG. Well it was a Sonnet but not a Shakespearean sonnet - a Petrarchan sonnet, which has an entirely different rhyme scheme. The reason why it got the wrong answer was because if you Google "Shakespearean Sonnet" you get over 7 million hits, but if you google "Petrarchan Sonnet" there are only about 600,000 hits. Therefore it's over 10 times as likely to plump for Shakespearean than Petrarchan. The only way to get this right would be to have a balanced training set, or to have the Bayesian prior probabilities of the two classes. Given the massive number of different classes it could be asked to identify, I do not see how this problem could be solved.
@golfrelax9795
@golfrelax9795 Ай бұрын
AI is just a better GOOGLE. It summarize it for you.That's it. If you asked something common, it'll sum up better. If it's something less common, it'll struggle. If it's something new, no search results!!
@alan1507
@alan1507 27 күн бұрын
@@XenoCrimson-uv8uz it's not quite like that. The softmax function yields a probability for each possible answer. It would be possible for it just to pick the one with the highest probability, but (I'm told) this leads to repetitive cycles. So having got the posterior probabilities of all the words in the dictionary (as it were) it samples that probability distribution using a random number generator. In this way it focusses on maybe 20-30 most probable words. The trouble is that in the example above "Shakespearean" is 10 times as likely as "Petrarchan" so 10 times as likely to be picked by the random number generator.
@XenoCrimson-uv8uz
@XenoCrimson-uv8uz 27 күн бұрын
@@alan1507 oh.
@ekki1993
@ekki1993 7 ай бұрын
As a bioinformatician, I will always assume that the exponential growth will plateau sooner rather than later. Sure, new architectures may cause some expected exponential growth for a while, but they will find their ceiling quite fast.
@visceralcinema
@visceralcinema 7 ай бұрын
Thus, the exponential curve becomes logarithmic. 🤭🤓
@arthurdefreitaseprecht2648
@arthurdefreitaseprecht2648 7 ай бұрын
​@@visceralcinemathe famous logistic curve 😊
@typeins
@typeins 7 ай бұрын
Imagine Blackrock created Aladin the super computer back in the days in the desert protected by the American military with full force. And now we are talking about small companies (compared to that giant monster)
@ekki1993
@ekki1993 7 ай бұрын
@@typeins Were you responding to a different comment there?
@jan7356
@jan7356 7 ай бұрын
The curve he drew and explained was actually logarithmic in the amount of data (sublinear).
@djdedan
@djdedan 7 ай бұрын
I’m not a child development. Specialist so take this with a grain of salt but What’s interesting is that you can show a child one image of a cat. Doesn’t even have to be realistic and they’ll be able to identify most cats from then on. What’s interesting is that they may mistake a dog for a cat and they will have to be corrected but from then on they will be able to discern the two with pretty high accuracy. No billions of images needed.
@bened22
@bened22 7 ай бұрын
Yes. Computer AI has nothing to do with human intelligence.
@joaopedrosousa5636
@joaopedrosousa5636 7 ай бұрын
That brain was in fact trained with a vast amount of data. That baby brain was created with a DNA that guided the development of those nervous structures while it was in the mother's womb. The ancestors of that baby going millions of years in the past interacted with the world through visual means and the more successful at reproducing the genes encoded those useful visual perception and classification tasks
@dtracers
@dtracers 7 ай бұрын
What you are missing is the 3-4+ years of "training" the human AI has taken up to that point of different cats and dogs. +Millions of years of evolution that has made the best sub networks that has learned. That second piece is hard because it's like running Google's automl over every possible data set and possible network architecture for an LLM
@ParisCarper
@ParisCarper 7 ай бұрын
Like someone else mentioned, evolution has selected for brains that can learn and adapt new concepts quickly. Especially very tangible concepts like different types of animals and how to easily recognize them. For the AI, you have to start from scratch. Not only do you have to teach it cats, but you have to teach it how to understand concepts in general
@mrosskne
@mrosskne 7 ай бұрын
@@ParisCarper what does it mean to understand a concept?
@ohno5559
@ohno5559 7 ай бұрын
Everyone remembers the people who wrongly said the internet wouldn't transform our lives, no one remembers the people who correctly said the segway wouldn't transform our lives
@isiahs9312
@isiahs9312 7 ай бұрын
I can still remember people telling me how social media would fail.
@ohno5559
@ohno5559 7 ай бұрын
@@isiahs9312 yeah and you remember because social media succeeded. it's survivorship bias
@morgant.dulaman8733
@morgant.dulaman8733 7 ай бұрын
@@ohno5559 My guess: we're looking at the same situation as all booms, busts, and normalization we've seen for everything from Tulips to Cryptocurrency. New product gets introduced->it's viewed as the next big thing with infinite possibilities and a money printer attached-> people overinvest-> something goes wrong and the bubble pops->it collapses->later, quietly, we see its actual potential...and its usually less than suspected, but still has its uses and slips into mundane use.
@vincemccord8093
@vincemccord8093 6 ай бұрын
"Victory has a thousand fathers, but defeat is an orphan."
@squamish4244
@squamish4244 3 ай бұрын
@@ohno5559 Clicking on this video itself is confirmation bias.
@supersnail5000
@supersnail5000 7 ай бұрын
Im surprised 'degeneracy' wasnt also mentioned in this - basically that as more AI generated content leaks into the dataset, further training could actually lead to worse results. There are ways of encoding the output to evidence that the data was generated, but that likely wont hold up if edits were made to the data prior to it entering the training corpus.
@Raccoon5
@Raccoon5 7 ай бұрын
AI generated data is frequently used for training of AI and it has pretty decent results. I doubt what you say is true, but having real world in the data set is always important. But that's not really a problem since we are taking more and more videos and photos of the real world.
@TheManinBlack9054
@TheManinBlack9054 7 ай бұрын
I do not think that the so called "model collapse" presents an actual danger to AI advancement as was shown by the Phi models. The models can be trained on synthetic data and perform well.
@existenceisillusion6528
@existenceisillusion6528 7 ай бұрын
@@TheManinBlack9054 That synthetic data was carefully curated. The problem still exists if someone isn't careful enough.
@monad_tcp
@monad_tcp 7 ай бұрын
@@Raccoon5 no, those are called adversarial models, they don't work that well
@pvanukoff
@pvanukoff 7 ай бұрын
Think about pre-AI, when we just had humans learning. What "dataset" did they learn on? They learned on things created by humans before them. If humans can learn from humans, and generate new, interesting, innovative results, I don't see why AI can't do the same by learning from and building on data generated by other/previous AI.
@tommydowning3481
@tommydowning3481 7 ай бұрын
I love this content where we get to delve into white papers with the convenience of a youtube video, not to mention with the genuine enthusiasm Mike always brings to the table. Great stuff, thanks!
@xX_dash_Xx
@xX_dash_Xx 7 ай бұрын
same here, +1 for paper review. and I appreciate the pessimism-- nice change of pace from the autofellatio that two minute papers does
@skoomaenjoyer9582
@skoomaenjoyer9582 7 ай бұрын
@@xX_dash_Xx i’ve had my fill of the generative AI hype too… “no, im not worried that my quality, new work will be replaced by a machine that moves heavily-represented, average work.”
@blucat4
@blucat4 7 ай бұрын
Agreed, I love Mikes videos.
@brianbagnall3029
@brianbagnall3029 7 ай бұрын
​@@blucat4The videos are all right but you can tell he really wants to pick his nose.
@lobstrosity7163
@lobstrosity7163 7 ай бұрын
That paper will be displayed in the Museum of Human Civilization. The robotic visitors will shaken their headoids at the naïvité of their late creators.
@OneRedKraken
@OneRedKraken 7 ай бұрын
This is sort of confirming my suspicions of where this is all heading atm. When I understood the reason why AIs are confused about human hands and limbs. It made me understand the biggest flaw with AI. It doesn't 'understand' anything. Which is why even though its' been dumped tons of reference images and photos of humans. Still doesn't understand that the human hand has no more than 5 fingers. Why is that? Because in it's test data it has pictures of people holding something with two hands, but where one hand is hidden by the angle/perspective. And so the AI only sees one hand, but a bunch of extra fingers. It's conclusion "Human hand can have up to 10 fingers". That's a really big hurdle to climb over.
@picketf
@picketf 7 ай бұрын
Fingers problem has been fixed in Dall-E 3 also you can ask it for 3D Models now and it will output in Blender script which means it's being trained to link concepts to shapes
@nbboxhead3866
@nbboxhead3866 7 ай бұрын
A lot of the problems caused by limbs and fingers in AI-generated images happen because there isn't any pre-planning, so duplicates of something there's meant to be only one of happen easily. For example, if you have it try and generate an image of a man waving at you, there are several different positions his arms and hands could be in that would make a coherent image, and because the AI doesn't start out thinking "One arm here, fingers positioned like this..." and instead just generates an image based on a function that returns pixel values independent of each other, you get separate parts of the image thinking they're the same arm. I guess Dall-E 3 must have some sort of pre-planning that happens, which is why it doesn't fail limbs and fingers as badly. (I say "as badly" because it still won't be perfect)
@kirishima638
@kirishima638 7 ай бұрын
Except this has largely been fixed now.
@TheMarcQ
@TheMarcQ 7 ай бұрын
"fixed" by programmers adding phrases to your prompts making sure to include appropriate number of fingers. That is a hotfix at best
@picketf
@picketf 7 ай бұрын
@@TheMarcQ apparently polydactyly is a real condition that is statistically not that rare. Will Smith slurping noodles is not that long ago and current AI is really leaps better.
@peterisawesomeplease
@peterisawesomeplease 7 ай бұрын
I think a key issue is we are actually running out of high quality data. LLMs are already ingesting basically all high quality public data. They used to get big performance jumps by just using more data. But that isn't really an option anymore. They need to do better with existing data.
@jlp6864
@jlp6864 7 ай бұрын
theyre also now "learning" from ai generated content which is making them worse
@Alan.livingston
@Alan.livingston 7 ай бұрын
I worked on a system a while back that used parametric design to convert existing architectural plans and extrapolate them out into 3d models needed to feed into steel frame rolling machines. The hardest part of what we did was accomodating the absolute garbage architects would put in the plans. Turns out when a source isn’t created with being machine readable in mind it’s often difficult to do anything about it.
@peterisawesomeplease
@peterisawesomeplease 7 ай бұрын
@@Alan.livingston Yea the problem of garbage or at least badly structured data is really clear in LLMs. Probably the most obvious example is they never say "I don't know" because no one on the internet says "I don't know". People either respond or they say nothing. So the LLMs don't have any idea of uncertainty. Another related issue that comes up constantly is that LLMs will give a popular answer to a popular question when you actually asked a question with a slight twist from the popular one. For example ask an LLM "what is an invasive species originating on an island and invading a mainland". They all answer the much more popular reverse question. Its a data problem the LLMs can't escape the overwhelming larger amount of data on the reverse question because all they see is the text.
@quickdudley
@quickdudley 7 ай бұрын
⁠@@jlp6864there are machine learning algorithms that aren't really affected by that kind of thing but adapting them to text and image generation would be pretty tricky to do right.
@picketf
@picketf 7 ай бұрын
​@@peterisawesomeplease I asked for ChatGPT 4 to calculate the distance to Gamma-ray burst 080916C one of the most violent time distortion events ever on record using a trifecta of 1. the cosmological formula for luminosity distance in a flat universe, 2. redshift, specifically to compensate for the time dilation effect, 3 Emission energy calculations. It responded 3 times the first 2 answers it concluded right after filling the page with formulas that they were incorrect and restarted the answering process presenting new numbers. I'd say it is a rather peculiar case but for sure those 2 wrong answers and the fact that it had become aware of their fallacy AFTER formulating everything twice is an attest to its indecision😅
@PristinePerceptions
@PristinePerceptions 7 ай бұрын
The data we have is actually incredibly limited. We only use mostly 2D image data. But in the real world, a cat is an animal. We perceive it in a 3D space with all of our senses, observe its behavior over time, compare it all to other animals, and influence its behavior over time. All of that, and more, makes a cat a cat. No AI has such kind of data.
@swojnowski8214
@swojnowski8214 7 ай бұрын
you can't recreate a 3d matrix using a 2d matrix. it is about dimensions. Development is about going from lower to higher dimension, you can do it if you are at the higher dimension, but not at the lower, umless you maku up some stuff to plug holes, that's why llms hallucinate, that's why we dream ...
@burnstjamp
@burnstjamp 7 ай бұрын
However it's also true that humans can easily and accurately tell animals apart from a young age, even if shown only static images (or even abstract representations) of them. The fact that we have more senses and dimensions with which we can perceive input seems less important than the fact that the human mind simply has far more capacity for pattern recognition. I also don't think that introducing more variety in input would solve the issue presented in the video-only delay it. If 2D image data fails to produce exponential or linear improvement in a generalized model over time, I fail to see how 3D data, or sonic data, or temporal data, or combinations therein would substantially change the reality of the problem
@joelthomastr
@joelthomastr 7 ай бұрын
Has nobody done the equivalent of plugging a large neural network into a Furby or an Aibo?
@celtspeaksgoth7251
@celtspeaksgoth7251 7 ай бұрын
@@burnstjamp and born instinct
@picketf
@picketf 7 ай бұрын
Well, AI is currently being trained on the billions of minutes being uploaded to youtube. Imagine you could say at any point of 2023 or 2024 that you watched every single cat video ever uploaded.
@Benjamin1986980
@Benjamin1986980 7 ай бұрын
As a chemical engineer, this reminds me of a warning from my numerical methods professor. More data add a more complex model does not mean better prediction. Eventually, you will get to the proverbial curve fitting of the elephant. This is where new model everything absolutely perfectly, but you have zero predictive power. Because you have now modeled in all the inherent chaos in the system. More complicated does not necessarily mean better
@adaelasm6467
@adaelasm6467 6 ай бұрын
That’s just basic bias-variance trade off though. Literally taught in Machine Learning 101. Overfitting is just as bad as underfitting.
@balasubramanianah969
@balasubramanianah969 4 ай бұрын
I don’t understand how what you said is different from basic overfitting.
@Benjamin1986980
@Benjamin1986980 4 ай бұрын
@@balasubramanianah969 It's not. Just two ways of saying the same thing
@srikanthan1000
@srikanthan1000 4 ай бұрын
U just described the problem with String theory. At least according to popular opinion.
@Reydriel
@Reydriel 7 ай бұрын
5:35 That was clean af lol
@squirlmy
@squirlmy 7 ай бұрын
I've never seen a better right angle drawn by hand!
@NoNameAtAll2
@NoNameAtAll2 7 ай бұрын
@@squirlmy tbf, it's on a celled paper
@dBradbury
@dBradbury 7 ай бұрын
@@NoNameAtAll2 I certainly couldn't do that that quickly and smoothly even it were gridded!
@drlordbasil
@drlordbasil 7 ай бұрын
he's an AI
@BillAnt
@BillAnt 7 ай бұрын
AI is more like crypto with diminishing returns. There will be incremental improvement, but less and less significant than the original starting point.
@TheGbelcher
@TheGbelcher 7 ай бұрын
“If you show it enough cats and dogs eventually the elephant will be implied.” Damn, that was a good analogy. I’m going to use that the next time someone says that AI will take over the Earth as soon as it can solve word problems.
@tedmoss
@tedmoss 7 ай бұрын
We haven't even figured out if we are at the limit of intelligence yet.
@WofWca
@WofWca 7 ай бұрын
1:00
@Brandon82967
@Brandon82967 7 ай бұрын
That's not true but if you show it enough Nintendo games and images of supervillains, it can put Joker in Splatoon.
@SimonFrack
@SimonFrack 7 ай бұрын
@@Brandon82967What about Darth Vadar in Zelda though?
@Brandon82967
@Brandon82967 7 ай бұрын
@@SimonFrack its possible
@natey313
@natey313 7 ай бұрын
Another factor that I think plays a huge part in this decline of AI art generators specifically, is the decline of recular artists... As AI replaces and kills off artistic jobs, you reduce the data set even further, as generative AI images require other 'look alike' images to generate something... At some point, the data set available will be so small that AI art programs won't be able to make anything unique anymore... It will just be the same handful of images generated with slightly different features... At that point, the AI developers will have to make a choice; to leave it alone and watch their program get progressively more boring and stagnant, or use other AI generated images as part of the data set... But, AI art using AI art is similar to inbreeding... If it takes another AI generated image as 'acceptable data' it will take and add its imperfections with it, which, overtime, will pollute and corrupt the data set, and produce increasingly inferior images...
@walkingtheline1729
@walkingtheline1729 7 ай бұрын
You are missing the fact the professional artist are putting invisible squiggles over there art that ai will pick up on to protect there are for being used for learning with out their concent. It's actively being lead to failure. They have been doing this for years
@童緯強
@童緯強 6 ай бұрын
Until Open AI fires their entire Content Management Team, there will always be a job for real artists.
@thiagopinheiromusic
@thiagopinheiromusic 6 ай бұрын
You've highlighted a critical issue in the long-term sustainability of AI art generation. The potential decline of human artists and the reliance on AI-generated images for training could indeed lead to a degradation in the quality and uniqueness of AI-generated art. Here’s a closer look at the implications and challenges: Decline of Regular Artists Reduced Diversity in Training Data: As fewer human artists produce original work, the pool of diverse, high-quality training data shrinks. This limits the variety and richness of inputs available for training AI models. Homogenization of Style: With a smaller dataset, AI art generators might start producing images that are increasingly similar, leading to a homogenization of artistic styles. The unique, personal touch that human artists bring to their work becomes harder to replicate. Challenges of Using AI-Generated Data Data Quality Degradation: Relying on AI-generated images for training can introduce errors and imperfections that compound over time. This can lead to a decline in the overall quality of the generated art, akin to the genetic defects seen in inbreeding. Loss of Innovation: Human artists innovate and push boundaries, creating new styles and techniques. AI models, however, are primarily imitators. Without fresh human input, the innovation in AI-generated art stagnates, leading to repetitive and uninspired creations. Potential Solutions Hybrid Approaches: Combining AI with human creativity could be a way forward. AI tools can assist artists, enhancing their productivity and providing new avenues for creativity, rather than replacing them entirely. This synergy can help maintain a diverse and high-quality dataset. Regular Infusion of Human Art: To prevent data degradation, it’s essential to continuously infuse the training datasets with new human-created art. This can ensure that AI models remain exposed to novel and diverse artistic expressions. Improved Feedback Mechanisms: Developing better feedback mechanisms where AI can learn from critiques and preferences of human artists and audiences can help maintain the quality of AI-generated art. This involves more sophisticated algorithms capable of understanding and incorporating nuanced feedback. Ethical and Sustainable Practices: Encouraging ethical practices in AI development and usage is crucial. This includes fair compensation for artists whose work is used to train AI models and promoting the importance of human creativity in the arts. Long-Term Implications The trajectory of AI art generation depends significantly on how these challenges are addressed. Here are some long-term implications to consider: Cultural Impact: The decline of human artistry could have a profound cultural impact, as art has always been a reflection of human experience and emotion. Maintaining a vibrant community of human artists is essential for cultural diversity. Economic Considerations: The economic ecosystem surrounding art and artists might need to adapt, finding new ways to support and integrate human creativity with AI advancements. Technological Evolution: AI technology itself will need to evolve, finding ways to better emulate the creativity and innovation inherent in human artistry. This could involve breakthroughs in how AI understands and generates art. In summary, the future of AI art generation hinges on finding a balance between leveraging AI’s capabilities and preserving the unique contributions of human artists. Ensuring that AI complements rather than replaces human creativity will be key to sustaining a dynamic and diverse artistic landscape.
@JohnnyThund3r
@JohnnyThund3r 6 ай бұрын
I don't think any real artists jobs are in danger. I've been messing around with this A.I. generated stuff for awhile now and it's clear what you're talking about has already happened. The A.I. cannot create any new art styles like a human can, it's mostly just mimicking the styles of other human artists to create basically knock offs of their work. This is great for fast prototyping and early concept work, but it's not a real replacement for real designers who knows how to make things work in a game or movie. Any studio getting rid of their art team and replacing it with A.I. is shooting themselves in the foot, AI is a billion years behind us humans when it come to creativity and abstract conceptualization.
@madrabbit8722
@madrabbit8722 5 ай бұрын
@@thiagopinheiromusic I think you just ironically highlighted a critical issue in generative AI by the fact that this entire essay you dumped here is instantly recognizable as being AI-generated
@marcbecker1431
@marcbecker1431 7 ай бұрын
This is totally off-topic, but the quality of the X and Y axes of the graph at 5:43 is just stunning.
@johnpienta4200
@johnpienta4200 5 ай бұрын
Thought this was sarcasm when I read it before getting to that point in the video - expecting wonky labels, honestly, but man, I don't think I could draw that well to save my life.
@michelleblit6526
@michelleblit6526 7 ай бұрын
my boyfriend is your student in the university of nottingham and he loves your videos!! you're the reason he came to study here!
@MaddMoke
@MaddMoke 7 ай бұрын
I love the idea that not only are these companies investing millions of dollars into something that might be for naught but also convincing clueless board of directors in several industires to cut jobs and invest in a product that might never get better. Add into this the legitimate issues of stolen art or apps using personal data from its users who are forced to sign very unclear TOS and you get a morality bomb of human wasteful greed
@msclrhd
@msclrhd 7 ай бұрын
I've seen image generation gets worse the more qualifiers you add as well. Like asking for "a persian cat with an orange beard" -- it will start including orange coloured things, the fruit, or colour all the fur orange, give it orange eyes, or make the background orange. I think this is a fundamental limitation of the way transformer models work. Specifically, transformers are trying to reduce the entire thing into a single concept/embedding. This works when the concepts align ("Warwick Castle", "Medieval Castle", etc.) but not when the concepts are disjoint (e.g. "Spiderman posing for a photo with bananarama."). In the disjoint case it will mix up concepts between the two different things as it mushes them to a single embedding. A similar thing happens when you try and modify a concept, e.g. "Warwick Castle with red walls". The current models don't understand the concept of walls, etc., how things like cats or castles are structured, nor can they isolate specific features like the walls or windows. If you ask it for things like that (e.g. "Warwick Castle with stained glass windows") it is more likely to show pictures focused on that feature rather than an image that has both.
@Pravlord
@Pravlord 7 ай бұрын
nope
@justtiredthings
@justtiredthings 7 ай бұрын
Image generators are empirically getting better at this sort of coherence and prompt-adherence, so 🤷
@inverlock
@inverlock 7 ай бұрын
Most people just work around this issue by doing multi step generation where they clearly separate concepts between steps and allow the later steps to generate content in the image with the context of the previous step as a base image. This doesn’t actually solve the issue but is a reasonably effective mitigation.
@roxymigurdia1
@roxymigurdia1 7 ай бұрын
that's not how transformers work and also i dont know where u got the idea that models don't understand the concept of walls
@msclrhd
@msclrhd 7 ай бұрын
@@roxymigurdia1 If you are just talking about walls, transformers can encode that concept as a vector sure. Similarly they can encode variants of those given enough data like "brick wall" or "wall with graffiti". My point is that if you ask the model to point out the walls on warwick castle it cannot do that as it does not know how to relate the concept (feature vector) of a wall to the concept (feature vector) of castles. Thus, even if it can encoded "red brick walls" and "warwick castle" correctly (which it can), it does not necessarily know how to draw "warwick castle with red brick walls" as it does not know how to align those two concept vectors, nor where the walls on the castle are in order to style them differently. That is what I meant. I've just tested this on Stable Diffusion 2.1 and that does a lot better job with these combined wall/castle examples than 1.5 does. It still messes up my "Spiderman in a photo with bananarama" example, though (ignoring the weird faces) :shrug:!
@ohalloranpeter
@ohalloranpeter 7 ай бұрын
I was having exactly this argument during the week. Thanks Computerphile for making the point FAR better than I!
@emerestthisk990
@emerestthisk990 7 ай бұрын
I'm what you call a 'creative' that uses Photoshop and After Effects and in my professional work and personal time I cannot find a single use for AI. After a playing around period with ChatGPT I haven't gone back to use it. I don't even use the new AI features in PS. I think the corporations and big tech really want AI to be the next revolution and are trying to sell it to us as that so they can line their collective shareholder pockets'. But the reality is far from the monumental shift being sold.
@francisco444
@francisco444 7 ай бұрын
Lol not a single use for AI? I've seen so many creatives make this mistake it's actually enraging. Creatives are about pushing the boundaries, fostering connection, and mastering their tools. If AI is not at all for you, that's fine. But you'll find yourself in the back of the line eventually.
@G3Kappa
@G3Kappa 6 ай бұрын
I've been using Runway to automatically do masking and it works nice. of course it has its limits, but it does 80% of the work I would need to do by hand in seconds. And that's fine, because I can always retouch it by hand.
@LordConstrobuz
@LordConstrobuz 6 ай бұрын
so basically youre a boomer and you dont know anything about AI or the AI tools available, lol. this is no different than being like "what? iPod? I dont need that, i have a CD walkman"
@newerstillimproved
@newerstillimproved 4 ай бұрын
A not very creative creative, perhaps?
@squirrelzar
@squirrelzar 7 ай бұрын
It’s interesting because animals, and specifically humans as the prime example of what a “general intelligence” should be almost proves it’s not a data problem. It’s a learning problem. I’d argue “the something else” is a sufficiently complex system that is capable of learning on comparatively small data sets. And we probably A don’t have the right approach yet and more importantly B don’t yet have access to the sort of technology required to run it
@fakecubed
@fakecubed 7 ай бұрын
Our brains are constantly learning in massively parallel operations every minuscule fraction of a second every single day for decades at a time, across all of our senses simultaneously. It's pretty hard to compete with that.
@squirrelzar
@squirrelzar 7 ай бұрын
@@fakecubed agreed - it’s a totally different system than “just some random numbers in a black box”
@annasofienordstrand3235
@annasofienordstrand3235 7 ай бұрын
It's not a learning problem either, it's a metaphysical problem. If the senses gather "information," then how do neurons in the brain know what that information represents? Try to answer without using a homunculus argument. That is seemingly impossible, and so neural coding has failed.
@squirrelzar
@squirrelzar 7 ай бұрын
@@annasofienordstrand3235 I think that’s what I’m getting at by saying a sufficiently complex system. The brain and its senses do ultimately boil down to a system of inputs and outputs. It’s just extremely well tuned for pattern recognition to the point where you only need to see a handful of cats to sufficiently identify all other cats. Hence my argument it’s not a data problem but a learning problem. You need a system that can operate on a small subset of data and be able to distinguish the important pieces of detail while chocking the rest up to attributes of that specific thing. And that’s only for classification problems. Just a tiny subset of what a general intelligence should be capable of
@mrosskne
@mrosskne 7 ай бұрын
@@annasofienordstrand3235 What does it mean to know something?
@t850
@t850 7 ай бұрын
..."pessimistic" (logarithmic) perfomance is what economists would call "law of diminishing returns" and is basically how systems behave if you keep increasing one parameter, but keep all other parameters constant...:)
@fakecubed
@fakecubed 7 ай бұрын
The thing is, the other parameters aren't constant. I also don't think we're close to maxing out on dataset sizes either.
@ekki1993
@ekki1993 7 ай бұрын
And exponential performance is what any scientist outside of economics calls "unsustainable".
@t850
@t850 7 ай бұрын
​@@fakecubed ...that may be so, but each parameter contributes to the outcome to some degree, and even those have their limits. In the end it's only a matter of where exacty are we on the logarithmic curve. At firs curve may look as if will be rising indefenitely, but in realitiy it always reaches the "ceeling" before it flattens out. It's like driving a car. At first it seems as it will keep on accelerating forever, but in the end it reaches top speed no matter how hard or how long you floor it, how many speeds you have, how much power in the engine there is, or how low of a drag you can reach. If you want to keep on accelerating you need a new paradigm in propuslion (technology)...
@t850
@t850 7 ай бұрын
@@ekki1993 ..."stonks" curve...:P
@Baamthe25th
@Baamthe25th 7 ай бұрын
@@fakecubed What others parameters can be really improved, to the point of avoiding the datasets diminishing ROI issue ? Genuinely curious.
@marcuswiesner9271
@marcuswiesner9271 7 ай бұрын
OK, just some practical experimentation to back up what this video is saying. I looked up a bunch of pictures of rare birds, trees, and pictures with a lot of stuff going on in them and headed over to GPT4 to test it out. This is what I found: The AI did perfectly on identifying all of the rare birds, trees and plants. It perfectly described the scenes I sent it, down to the last detail, identifying all of the objects, even in abstract images. However, this is where it failed: When I asked it to identify the type of cloud in the background of one of the images, it guessed cumulus when they were clearly stratus. An easy question to get right. Any person can tell the difference between cumulus and stratus with 5 seconds of education on the subject. When I asked it to identify the state or city where one of the shots was taken, it wasn't able to do so even with the word "Miami" written in the scene. So here's what I think is happening. Being trained on the whole internet, it obviously knows the topic of any picture I could pull up and put into it. But in the case of asking what the clouds in the background of this image were, it failed because the image was not tagged with that data specifically. So in its training data, it knows images by their features and what they are tagged by. If the researchers want to create more comprehensive understanding, they'll have to label every part of every photo, which is obviously not even practical. This probably means: AI will be amazing at clear categorization tasks involving any image on the internet - especially the main subject of the image, like a specific image of a bird that was labeled with the species of the bird, or an image of a tree labeled with the type of tree. It may even become really good at new categorization tasks if specifically fine-tuned on some data, like cloud types and types of animals. But fresh pictures taken "in the real world" may stump the AI. And even the same image of the same tree in the background of a photo that appears to feature something else entirely may stump the AI, because the image was never tagged as having that tree in the background. That's essentially what I'm thinking. Because these stratus clouds are as clear as day, and it has doubtless been trained on millions of pictures of stratus clouds. But it couldn't identify them in the background of this racing track. So, seems pretty clear to me what's happening here. It doesn't seem to be developing a real ability to generalize beyond a certain point. So yeah, just to wrap up - this may be fixed if the researchers painstakingly tag every piece of every image they train it on, from the clouds to the trees in the background, at least for images on the internet. But it may struggle with real world, "fresh" imagery (that was not in its training data) forever, making hilarious mistakes for a long time to come.
@Onyx-it8gk
@Onyx-it8gk 3 ай бұрын
This is very true. If you want to look into this further, all of these labels AI was trained on were provided by humans, specifically thousands of underpaid people in India.
@marcuswiesner9271
@marcuswiesner9271 3 ай бұрын
@@Onyx-it8gk Oh 100%
@TheNewton
@TheNewton 7 ай бұрын
4:12 "it didn't work" , too right , afaik the emergent behaviors of large language models(LLMs, big data sets) as you scale up are plateaus and have not lead to any consistent formula/metrics to indicate emergent behavior can be extrapolated as a trend Meaning we don't know if there are more tiers to the capabilities a LLM could have even IF you have enough data. And there's a paper that such emergence is humans doing the hallucinating with a 'mirage' due to the measurement metrics. [1] , and also see the emergence part in the standford transformer talk[2], [3]. The other issue in the here and now with scaling up to even bigger data is that most of the data after ~2023 is just gonna be garbage, as every source: the internet , internal emails, ALL content sources get polluted by being flooded with AI generated-content and no true way to filter it out. AKA model collapse[4] , though I don't know of much published work on the problem of LLM's eating their own and each others tails, probably more stuff if you view it as an attack vector for LLM security research. Bringing us again and again to realizing authenticity is only solvable by expensive intervention of human expertise to validate content. [1] arXiv:2206.07682 [cs.CL] Emergent Abilities of Large Language Models [2] youtube:fKMB5UlVY1E?t=1075 [Standford Online] Stanford CS25: V4 I Overview of Transformers , Emily Bunnapradist et al speakers [3] arXiv:2304.15004 [cs.AI] Are Emergent Abilities of Large Language Models a Mirage? [4] arXiv:2305.17493 [cs.LG] The Curse of Recursion: Training on Generated Data Makes Models Forget
@SlyRocko
@SlyRocko 7 ай бұрын
The polluted data could definitely be alleviated if generative AI had functionality similar but opposite to the Nightshade antiAI tool, where generated AI works could have injections to exclude themselves from learning models. However, there are still the other limits to AI that can't be solved without some novel solution that we probably won't even find anytime soon.
@Hagaren333
@Hagaren333 6 ай бұрын
I think I've seen 2 studies on model collapse, one applied to inverse diffusion models and one to LLM, I have to look them up because they both also come to the conclusion that synthetic data is harmful and model collapse is inevitable
@thunkin-ai
@thunkin-ai 7 ай бұрын
10:17 the blue marker pen _still_ doesn't have the lid on...
@harryf1867
@harryf1867 7 ай бұрын
I am that Analyst who says this in the meeting room to the person at the whiteboard :)
@redmoonspider
@redmoonspider 7 ай бұрын
Where's the sixth marker? Bonus question..what color is it?
@sergey1519
@sergey1519 7 ай бұрын
​@@redmoonspiderBlack.
@sergey1519
@sergey1519 7 ай бұрын
Last seen in stable diffusion video.
@redmoonspider
@redmoonspider 7 ай бұрын
@@sergey1519 great answer!
@theronwolf3296
@theronwolf3296 7 ай бұрын
There is a big difference between pattern matching and comprehension. At the first level, comprehension allows a mind to eliminate spurious matches, but further on, comprehension allows coming to conclusions that did not exist in the data. This is what intelligence really is. (GenAI could not have independently conceived of this study, for example). Essentially it's regurgitating whatever was fed into it. Actual comprehension goes far beyond that. Nonetheless this could be very useful for finding information though. An AI trained on millions of court cases, for example, could help a (human) lawyer track down relevant information that is critical... but it would require the application of human intelligence to determine that relevance as well as eliminate the material that does not apply.
@user-qm9ub6vz5e
@user-qm9ub6vz5e Ай бұрын
I’d push back on this. Modern neural networks do have some level of learning that they do by definition. You can see this when you look at specific layers and what they are paying “attention” to. Cough cough attention is all you need. But the modality in what they learn is limited. This is why to have true comprehension of the world around us these system need to be embodied in a physical 3d form like a humanoid robot that has more than just vision and language also known as embodied ai. The true challenge is ensuring the network is able to understand an edge in an image and then be able to draw and reason over the edge if that makes sense. We are just at the beginning of what ai could truly bring into this world.
@petermoras6893
@petermoras6893 7 ай бұрын
I think people mysticize Machine Learning and Generative AI far more than it needs to be. At the end of the day, ML is just an arbitrary function. It can be any function as long as we have the right input and output data. The obvious problem is that the possibility space of any problem balloons exponentially with it's complexity, so you eventually reach a point where you don't have enough resources to brute force the solution. However, I don't think we've reached the peak of generative AI, as there are other avenues of improvement other than more training data. One solution I think we'll see employed more is using more complex algorithms that help bridge the gap between the input and output data. For example, we don't train a Neural Net on pure images. We use a convolutional layer at the start in order to pre-process the image into data that is easier to find correlations with. But these layers can be anywhere in the NN and still be effective. (Personal opinion) For Image based gen-AI, I think future algorithms will use pre-trained algorithms that show understandings of 3D objects and their transposition onto 2D planes. The general image classifiers could then use the pre-trained 3D transposition as a basis for understanding 2D images, which would in theory give them an understanding of 2D object representation that is closer to our own.
@LesterBrunt
@LesterBrunt 7 ай бұрын
I’m a musicologist so my expertise is music and people truly, and immensely, underestimate how incredibly complex our ability to make music is. Sure AI can produce convincing tracks, but the ability to play in real time, in response to a human playing, is way beyond anything computers are capable of. The way we perceive music is actually extremely unique and complicated. We don’t see music as a series of frequencies in a rhythmical pattern, we see them as abstract entities. If I play Happy Birthday in another key, 2x as slow, and change a few notes, everybody would, somehow, still recognize the song. By every objective measure they are not the same song, but somehow every human perceives the music as some kind of abstract entity that can change its appearance without losing its identity. This seems so easy and trivial to us that it is incomprehensible to think that a computer could never do this. A computer reads the data and concludes that it can’t be happy birthday because the frequencies don’t match, the tempo doesn’t match. We all intuitively understand it is the same song, but to be able to quantify why is incredibly complex, and something WE don’t even fully understand.
@Aircalibur
@Aircalibur 7 ай бұрын
AIs currently only look at one thing or concept at a time. They can't perceive the relationships between things or concepts. For instance, ask an image generator AI to give you a rhino wearing rollerblades and it will give you the rhino and the rollerblades, but they're not in the correct relationship; the rhino isn't actually wearing the rollerblades. By that I don't mean the image not looking correct because the rhino's legs/feet and the rollerblades don't mesh quite right, I mean that the rhino and the rollerblades are nowhere near each other because the AI can't figure out the precise relationship between the two objects. Obviously the dataset doesn't have a rhino on rollerblades, but it doesn't matter even if it did and even if the image looks correct because the fundamental issue remains: the AI doesn't even really know what something simple like a tree is. It doesn't know how the leaves relate to the twig, how the twig relates to the branch and how the branch relates to the trunk. It has just memorized a bunch of pictures and it had to be told what a tree is beforehand. The human assigns the meaning, not the AI. A human knows when a song begins and ends and understands that the song is the relationship between the notes, not a simple set of objective characteristics. The AI just memorizes the song and the definition is sealed right then and there. It's not flexible or creative and it doesn't think critically. It simulates intelligence, but it's not intelligent.
@nuggetpolice2166
@nuggetpolice2166 9 күн бұрын
You just described tempo in the most pseudo science way. People recognize the song because it’s the same pattern, literally the most simple way to identify a song out of everything you mentioned.
@LesterBrunt
@LesterBrunt 9 күн бұрын
@@nuggetpolice2166 Absolutely not true. Go look up modern cognitive research on musical entrainment, relative pitch recognition, pitch/rhythm memory, etc. You clearly don’t know what you are talking about.
@XreezX
@XreezX 8 күн бұрын
​@@nuggetpolice2166 What about a cover in a different genre? What about playing with totaly diffrent instruments. Just thing about "Fugees - Ready or Not" vs "Enya - Boadicea". or even more crazy Winehouse’s cover of "All My Loving” by The Beatles. And think about easter eggs or a hommages in Media. These things are so subtle and need so much context, that even for humans its often difficult.
@nuggetpolice2166
@nuggetpolice2166 8 күн бұрын
@@LesterBrunt the fact you think any of those things are hard to understand only shows that you can’t even comprehend how a fraction works. Imagine thinking there is anything even worth explaining about “relative pitch frequency”…. Hahaha you are genuinely the dumbest person in the room wherever you go.
@programninja6126
@programninja6126 7 ай бұрын
Everyone who wants to understand AI should study mathematical statistics. The Central Limit Theorem with a convergence of 1/sqrt(n) isn't just a coincidence, it's the best linear unbiased estimator of any 1 dimensional problem. (there's technically other distributions like one for min(x) with their own convergence rate, but it's not better than 1/sqrt(n)) Machine learning models may have the advantage of being non-linear and thus can fit many more models than simple linear regression, but they obviously can't outperform linear models in cases where the data actually is linear so to think that a model can have anything other than diminishing returns is yet to be shown and would break statistics at its core (in fact 1/sqrt(n) is the best case, high dimensional linear models have a convergence rate of n^(-1/5) or worse, so if your data has a high intrinsic dimension then your convergence will obviously be slower) On the other side people pay a lot for this kind of research and it's excellent job security to keep trying for something you know probably won't happen
@paulfalke6227
@paulfalke6227 Ай бұрын
AI research has a long history - at least since the 1950s. After some years an AI researcher has to jump from one AI fashion to the next AI fashion. They are used to do it. This is new only for the investors.
@bakawaki
@bakawaki 7 ай бұрын
I hope so. Unfortunately, these mega companies are investing a ludicrous amount of money to force a line upwards that will inevitably plateau, while utterly disregarding ethics and the damage it will cause.
7 ай бұрын
How will it cause ethics and damage, if it plateaus?
@I_INFNITY_I
@I_INFNITY_I 7 ай бұрын
Not gonna happen.
@dsgowo
@dsgowo 6 ай бұрын
It already is. Look at the colossal amount of energy needed to power these algorithms and the underpaid workers in developing countries who label data for the models to train off of.
@george-and-gracie7996
@george-and-gracie7996 6 ай бұрын
This! And don't forget the disinformation it spreads that is destabilizing democracies around the world.
@Tobyodd
@Tobyodd 6 ай бұрын
@@I_INFNITY_I when is the last time you've seen literally anything infinitely shoot improve without a plateau
@SkullCollectorD5
@SkullCollectorD5 7 ай бұрын
Could part of the asymptote argument be that it will become harder and harder to train models on accurate data that *was not* previously generated by (another) model? Oversimplified, basically every written body of work released after 2022 will have to be viewed critically - not just in the prior healthy sense of informing your opinion well, but now that you cannot be absolutely sure it wasn't generated and possibly hallucinated by an LLM. Does generation degrade as if copying a JPEG over and over? In that way, possible training data is limited to human history up to 2022/23 - if you were to be cynical about it.
@tobiasarboe5753
@tobiasarboe5753 7 ай бұрын
I don't know if this plays into the argument from the video, but it *is* a very real problem that AI faces yes
@dahahaka
@dahahaka 7 ай бұрын
Honestly, none of this matters, Human Level Intelligence will always be the lower abound of what's possible with machine learning, if a human has enough data to gain the intelligence that it does, there has to be enough data for the human to gain this knowledge and intelligence... I'm seriously confused how Dr. Pound is oblivious to that :( just like people are blinded by hype, it seems theres a growing amount of people who are blinded by their discontent with hype and trying to disprove hype. Idk what i think of that
@medhurstt
@medhurstt 7 ай бұрын
I think its because the expectations of models is too great. Its already well known its possible to run an answer back through a model to improve it and I think this is what this paper is missing (although I haven't read it!). Its unrealistic to think a model can hold the answer to all questions from training. Many questions simply need multiple passes in the same way we need to think things through ourselves. I think the computerphile issue of insufficient representation of objects in models may well be real but is very solvable even if it become incremental improvement on the knowledge side of the AI.
@dahahaka
@dahahaka 7 ай бұрын
@@medhurstt exactly, just think about if I asked my mum to explain to me what Counter Strike is, you wouldn't consider her to not be intelligent because she's only heard that a couple times in her life :D
@Jack-gl2xw
@Jack-gl2xw 7 ай бұрын
Having enough high quality data to train on is always a concern, but fortunately, I think any (serious) body of work that is published with the aid of a GPT model will be audited by the author to make sure the quality is still high. In this sense, I dont think that this would be an issue because the signal to noise ratio is still positive. If models are trained on random internet content, I would say this is more of an issue as there may be some low quality generated text sneaking into the training data. While training on poorly made synthetic data may cause some issues, I think the main takeaway from the paper is that more data (even quality data) will not be able to get us to new transformative results. We need new learning approaches or techniques to better understand the data and the users intentions. Personally, I think this is where implementing a form of reinforcement learning would help break through this plateau. Supposedly this is what OpenAI's mysterious Q-Star algorithm is thought to be doing
@Shocker99
@Shocker99 Ай бұрын
I like how they called it. I hear ChatGPT's new model is only slightly better than the current model in some cases, but worse in others.
@chinese_bot
@chinese_bot Ай бұрын
If you compare LLM’s advancements all off ChatGPT you’re gonna be real confused when things don’t work out like you planned. Adding more data has diminishing returns, but there are other avenues and algorithms (think AlphaGO’s self learning models, stuff that Strawberry only recently has started incorporating into their LLMs) and it’s already doing things that people said were impossible just months ago. Even if all advancement were to stop and AI were to be forever stuck in its current state, it would still be producing advancements in every single field it’s deployed in for decades and decades.
@poduck2
@poduck2 7 ай бұрын
I'm curious about how AI will train itself as more and more AI content gets disseminated on the Internet. Will it degrade the performance, like photocopying pages that have been photocopied? It seems like it should.
@ItIsJan
@ItIsJan 7 ай бұрын
in the best case, the ai wont et better
@ItsSardine
@ItsSardine 7 ай бұрын
Yeah thats what I’m curious about. A couple years back it was extremely easy for big companies to mass-scrape the internet and take everything but now they would need to fight against themselves
@nadinegriffin5252
@nadinegriffin5252 7 ай бұрын
Most quality information isn't found on the Internet anyways. Quality writing and information is found in books and novels. Paywalls prevent acess to quality material online. AI already has an issue with plagiarism and not citing sources. It's like getting a handwritten photocopy of a photocopy about the Cabbage Soup Diet in the 90s. It supposedly claims developed at a hospital but has noting to back up that claim. It isn't a published paper, it doesn't link to the hospital website or provide you with a contact to verify the information. Infact, because AI has such poor input of information I wouldn't be surprised if I asked it about the Cabbage Soup diet that it would tell me it was developed at a reputable hospital and is great for losing weight. 🤣
@KnugLidi
@KnugLidi 7 ай бұрын
The paper reinforces the idea that bulk data alone is not enough. An agent needs to be introduced into the learning cycle, where the particular algorithm needs to identify what pairs are needed for a specific learning tasks. In a nutshell, the machine need to know to direct its own learning toward a goal.
@davidgoodwin4148
@davidgoodwin4148 7 ай бұрын
We are doing it a different way, via prompts. Our example is SQL for our system. The LLM knows SQL. It does not know our system as we never posted its structure publically (not because it is super secret, it is just internal). We plan to feed it a document describing our system. We then tell it to answer questions based on it (As a hidden prompt). That would work for teaching a model what an elephant is but I do feel you could provide fewer examples of new items once you have it generally trained.
@PazLeBon
@PazLeBon 7 ай бұрын
if all people were he same that could eventually work.... but we aint, im particularly contrary by default :)
@Hexanitrobenzene
@Hexanitrobenzene 7 ай бұрын
AlphaCode 2 has convinced me that LLMs + search will be the next breakthrough. Generate - verify paradigm. At present, it's not clear how to make the "verify" step as general as "generate".
@PazLeBon
@PazLeBon 7 ай бұрын
@@Hexanitrobenzene llms kinda are search already
@siceastwood2714
@siceastwood2714 7 ай бұрын
@KnugLidi isn't this just the concept of synthetic data? And there are actually efforts in creating AI's specialized on creating synthetic data needed for further training. I'm kinda confused why this concept isn't even mentioned here.
@YDV669
@YDV669 6 ай бұрын
So what he's saying, I think, is that cats are going to save us from the AI apocalypse. That's so cool.
@rozukke
@rozukke 7 ай бұрын
Happy to see general sentiment trending in a more realistic and less hype-based direction for this tech. I've been trying to make this same impression on people I know irl for ages now, especially people panicking about their jobs and AI sentience and other similar bs as if there is any likelihood of it happening in the immediate future. I blame KZbin for recommending all those horrible doomer AI podcasts indiscriminately
@Kknewkles
@Kknewkles 7 ай бұрын
Heh, I was worried last year so I learned the basics of "AI". Since then I'm not as worried, but increasingly annoyed at a first tech bubble that I clearly understand is just that. A bubble. A marketing bundle of most promising buzzwords in history.
@dahahaka
@dahahaka 7 ай бұрын
Anti-Hype based sentiment that disregards very very basic mathematics and statistics, isn't a good thing IMO, this is just as bad as Hype
@pedroscoponi4905
@pedroscoponi4905 7 ай бұрын
I think a lot of people worried about their jobs are less worried about the AI actually matching their skill level, and more about their bosses/clients buying into the hype and replacing them anyway.
@riccardoorlando2262
@riccardoorlando2262 7 ай бұрын
@@pedroscoponi4905 Honestly, that's a much more real problem. Just like with all tech bubbles before this one..
@inkryption3386
@inkryption3386 7 ай бұрын
​@@pedroscoponi4905 yeah this isn't a technological issue, it's a class issue.
@astropgn
@astropgn 7 ай бұрын
Every video I watch with dr pound makes me want to do a class with him. I wish his university would record the lectures and make them available on KZbin
@sloaiza81
@sloaiza81 7 ай бұрын
Godel Escher Back came out in 1979 and addressed most issues behind true AI. Almost half a century later and most still don't have a clue what it takes to make general intelligence.
@humanoid60
@humanoid60 6 ай бұрын
Since when did GEB actually propose any solutions to what it takes to create true AI? It seems you fundamentally misunderstood the book...
@djrmarketing598
@djrmarketing598 7 ай бұрын
I think the graph flattens out more with diminishing returns as more examples get added - neural networks are just complex number systems and at some point if you take every image ever made and put it into a computer, I feel like the system just moves into a state where it doesn't really know more, it just is slightly more accurate at the same but still makes errors, while still not being able to identify "anything". One of the issues I think we have is we're applying a "digital" solution to an analog system. Human and animal eyesight isn't pixels. If it was, we'd already have cameras attached to brains for the blind. The data stream of nature I believe is more of a "video" learning algorithm. It's not about the pixels but the change of each pixel to the next pixel. Look at an open field - you see the field, but when a cat runs across, you see the cat because your brain (and animal brains) are designed to see that and hear that, and thus it's much different. AI is not trained that way. We should be training AI from "birth" with a video stream teaching it about shapes and colors, like we do children. Maybe we can rapidly do that, but we're not going to see a breakthrough in AI until we're doing "real live" training and not just throwing a few billion words and images into a neural net.
@hexzyle
@hexzyle 7 ай бұрын
Yeah the reality is that these algorithms, no matter how much data you put into them, are still only reading/understanding the data within the constraints of the data type. E.g. a picture of a cat is just a picture of a cat, not an actual cat with weight, 3 dimensions, and a skeleton. These algorithms will become excellent at classifying or generating pictures of cats, but not being able to process how a cat moves in 3d space, the physical affects it has on the world, or how its skeleton is orientated geometrically. The machine algorithm is still hyper specific to "pictures" even though it has a facade of something more complex than that at first glance.
@thedudewhoeatspianos
@thedudewhoeatspianos 7 ай бұрын
Even if we do that, do we have any reason to believe it can achieve a different result? It might be worth trying but i suspect human brains use a different kind of pattern matching, and no amount of training data will overcome that hurdle. We have to change our approach.
@alexandrebenoit
@alexandrebenoit 7 ай бұрын
What your describing is just a video encoding. At the end of the day, computers are digital. Any video stream, no matter how you encode it is going to be digital. I'm not saying encoding is not important, it's a huge factor in training an effective model, but we are already doing that and have been for a long time.
@EtienneFortin
@EtienneFortin 7 ай бұрын
It's probably always the case for any problem. At some point brute force plateau. Reminds me when the only way to increase the processor speed was increasing the clock speed. I remember seeing graphs where the performance grew linearly and they were planning a 10 GHz P4. At some point they needed to be a little bit more clever to increase performance.
@beeble2003
@beeble2003 7 ай бұрын
Clock speed is a nice analogy. 👍
@ahmataevo
@ahmataevo 6 ай бұрын
It turns out requiring everyone to have liquid helium heat exchangers for basic computing is not viable for consumer-grade economies.
@jcdenton2819
@jcdenton2819 6 ай бұрын
Well, now it's even worse because we are at physical limit and they just add MORE processors on die, calling it "chiplet design", resulting in comically large processors, video chips and their coolers and radiators, with a corresponding power consumption.
@Ptaku93
@Ptaku93 7 ай бұрын
this plataeu isn't a problem, it's a solution. I'm so thankful for it and I hope other papers on the topic repeat these findings
@pedroscoponi4905
@pedroscoponi4905 7 ай бұрын
Reminds me of the story doing the rounds a while back about people trying to use these new gen-"AI" to identify species of mushroom and whether they're safe to eat, and the results were, uuuuh, _dangerously_ incorrect to say the least 😅
@sznikers
@sznikers 7 ай бұрын
Cant wait to see people with baskets of deathcaps identified by some halfbaked istore app as edible 😂
@DonVigaDeFierro
@DonVigaDeFierro 7 ай бұрын
That's insane. Even expert mycologists need to get their hands on a specimen to accurately identify them. Irresponsible to use it, and way more irresponsible to publish it or advertise it.
@sznikers
@sznikers 7 ай бұрын
@@DonVigaDeFierro sillicon valley bros dont care, gotta hustle to pay all the coaching bills...
@calvin7330
@calvin7330 7 ай бұрын
@@DonVigaDeFierro This means you can't ever advertise an AI that can "do anything", like ChatGPT. Even if you build in exceptions for potentially dangerous queries people will get around them, sometimes on accident
@xlicer
@xlicer 7 ай бұрын
@randomvideoboy1i agree with your comments but I think you are missing the point, that AI startups are the moment right now filled with all kinds of bad-faith bad actors grifters that just want to rush AI products with no care for any ethical concerns. And this is coming from someone that is actually supporting of AI and the technology.
@doomtho42
@doomtho42 7 ай бұрын
I can absolutely see there being a plateau regarding AI performance using current data processing techniques, and honestly I don’t think that’s as much of a “cavalier” idea as he suggests - I think the interesting part is how we progress from there and what new techniques and technologies arise from it.
@MichaelLonetto
@MichaelLonetto 7 ай бұрын
I think the future performance is more likely to be a series of stacked sigmoids of decreasing magnitude than either a full flattening or a transition to exponential growth, as engineers figure out tweaks to gain efficiency from the current models until that, too hits diminishing returns.
@thekingoffailure9967
@thekingoffailure9967 2 ай бұрын
Don’t call me a stacked sigmoid!! 😢
@Alex-cw3rz
@Alex-cw3rz 7 ай бұрын
I think one of the most worrying thing is when they start wanting to make the invested money back, over 150 billion has spent already
@mharrisona
@mharrisona Ай бұрын
I think this video will age like fine wine
@radio655
@radio655 3 ай бұрын
Excellent explanation. Its always great to watch Computerphile
@kevindonahue2251
@kevindonahue2251 7 ай бұрын
Yes, data required to train new models grows exponentially. They've already trained it with everything that can get their hands on and are moving towards "synthetic" ie AI generated data. The age of the Hapsburg AI is already here.
@ahmataevo
@ahmataevo 6 ай бұрын
It's already got the body part deformities part down in image generation.
@marybennett4573
@marybennett4573 7 ай бұрын
Very interesting! This reminds me of a post that demonstrated that one of the image generation AI programs couldn't successfully follow the prompt "nerd without glasses". Clearly the model determined that having glasses is an intrinsic property of "nerds" given the majority of it's source images included them. Silly little example but illustrative of the broader issue I think.
@tamix9
@tamix9 7 ай бұрын
That's more to do with the clip encoder being not great at encoding meanings like "without". Adding "without something" can instead make it more likely for that thing to appear.
@fakecubed
@fakecubed 7 ай бұрын
@@tamix9 Yeah, you need to use negative prompts if you want to avoid something.
@Lambda_Ovine
@Lambda_Ovine 7 ай бұрын
considering that, on top of these findings, the internet is being flooded with ai generated garbage, to the point that models are being trained with the generated output of other generative ai resulting in degenerate output that is going back into the datasets, i think is very reasonable to believe that we're going to hit a plateau or even start to have qualitative degradation
@kabosune9097
@kabosune9097 7 ай бұрын
I always hear this for the past 10 months, but I haven't ever seen AI image models degrade. And they can always roll back
@specy_
@specy_ 7 ай бұрын
One thing im mostly worried about LLMs is the fact that we are able to improve them only if more data is found or the architecture changes, lets say the second is doable, we still need the first. Where do we get this data? Well big portion of it comes from the internet, blog posts, articles, open source software,etc. who uses LLMs the most? Blogs, articles, and open source software. We are also polluding the web with low quality LLM generated texts. We all know what happens when you give an AI it's own input... It will be analogous to over fitting. We could find a way to detect and exclude LLM generated content, but pretty much everything nowadays uses a little bit of it. At best we have very few salvageable data to use, at worst, the LLMs actually start getting worse
@thiagopinheiromusic
@thiagopinheiromusic 6 ай бұрын
You raise a crucial point about the sustainability of training LLMs with high-quality data. The feedback loop of AI generating content that's then used to train future models could indeed lead to a degradation in quality, akin to overfitting. Detecting and filtering out AI-generated content is a potential solution, but as you mentioned, it's becoming increasingly challenging as AI use grows. One way to address this might be to create more robust methods for curating and verifying data quality, perhaps by combining human oversight with advanced AI techniques. Additionally, diversifying data sources and encouraging the creation of new, high-quality content by humans can help maintain the integrity of training datasets. Ultimately, innovation in how we approach AI training and data curation will be key to ensuring that LLMs continue to improve without succumbing to the pitfalls of self-referential data. It's a complex issue, but with thoughtful strategies, we can mitigate these risks.
@TheBookgeek7
@TheBookgeek7 7 ай бұрын
This is very interesting. I did something like he was suggesting, once, with the Bing AI. I have this entirely original (and possibly insane) idea in the world of Aesthetics, and I wanted to try it out on the AI; I mean if anything is likely to tell you if you've got something completely insane or not, it would be a device which basically exists to tell you the most popular ideas of the moment- so... that's basically why I did it. All that it could do, actually, was repeat what I was saying back to me, almost verbatim. It was weird! 🤣
@zoroark567
@zoroark567 7 ай бұрын
I think there’s an additional interesting problem here, which is that the more data we add to these models, the more pronounced the issues of underrepresented edge cases will be. If having more pictures of house cats than of rare endangered animals is a problem now, doubling the input data is going to double the size of that problem.
@jorgandar
@jorgandar 6 ай бұрын
The algorithm in our brain is NOT AN LLM!!! We may express thoughts through language, but "thought" is not the same as merely expression of words. We visualize, we have 'feelings', we have fuzzy intuitions, and then we have all sorts of unconscious processes that interact without our consciousness in strange ways. The point is, only the surface of all of these processes is language that comes out. An LLM is not the same as a thinking being. I have no doubt we'll eventually replicate real intelligence as there is nothing magic about our brains, but right now we are doing very interesting surface fakes, but it's not actual thinking.
@johnnylollard7892
@johnnylollard7892 7 ай бұрын
People thought AI would be like a thinking robot from a sci fi movie. In fact, it's a fancier way to outsource labor. Instead of some developing economy, it gets outsourced to a computer program. In a similar fashion, quality declines in a way that's first invisible, and then slowly snowballs.
@realleon2328
@realleon2328 7 ай бұрын
Yet companies try to and wonder why no one likes it
@TheBest-sd2qf
@TheBest-sd2qf 6 ай бұрын
It's still remarkable what it can do. If it never gets any better it's ok, it's incredibly useful as it is.
@ergwertgesrthehwehwejwe
@ergwertgesrthehwehwejwe 6 ай бұрын
So society objectively improves because less labor is needed to maintain it? Cry me a river
@thymenwestrum7011
@thymenwestrum7011 6 ай бұрын
​@@jurycould4275The whole technology concept revolves around replacing human jobs, what are you yapping about? AI has already replaced jobs and will continue to do so.
@Strix2031
@Strix2031 5 ай бұрын
​@@jurycould4275technology replaces jobs all the time
@doctorgibberish
@doctorgibberish 6 ай бұрын
That’s the most wholesome and theme fitting sponsor ad I’ve ever seen
@beaverson
@beaverson 7 ай бұрын
I hope generative "AI" gets sued for all the date they stole with the mask of being nonprofit.. But I also hope it stops getting better. Dont want more dirty money getting pumped into Silicon Valley at the expense of 95% of the working classes data.
@EJD339
@EJD339 7 ай бұрын
So can someone explain to me when I google a question now it Tends to highlight the wrong answer or not even answer the question I was searching for? I didn’t feel like it use to be that.
@addeyyry
@addeyyry 7 ай бұрын
Found bing to be much more reliable actually
@ReflectionOcean
@ReflectionOcean 6 ай бұрын
By YouSum Live 00:00:03 Using generative AI for various tasks 00:00:27 Potential of achieving general intelligence through vast data 00:01:31 Challenge: vast data needed for zero-shot performance 00:02:00 Limitations of adding more data and models 00:02:20 Importance of data quantity for effective AI applications 00:03:28 Difficulty in training models for specific and rare concepts 00:04:44 Performance plateau despite increasing data and model size 00:06:10 Need for alternative strategies beyond data accumulation 00:10:22 Uncertainty about future AI advancements and performance trends By YouSum Live
@joelandrews8041
@joelandrews8041 7 ай бұрын
One potential solution - Has there been any research into developing an AI model which classifies a problem for use by another more specialised AI model? For the plant/cat species case, a master AI would identify the subject, and would then ask the specialised subject specific AI model for further info on that subject. This prevents the master AI from needing all the vast amount of training data of the subject specific AI. Not sure if I've explained this very well!
@zactron1997
@zactron1997 7 ай бұрын
The concept makes sense, but the problem is more fundamental according to this paper. It's not about whether you can train a model to do all these things, it's about the requirement for data at a rate that doesn't match the supply. In your example, the problem is never having enough data to make a good enough "specialist" AI to a sufficient quality.
@helix8847
@helix8847 7 ай бұрын
Yeah its called FineTuning. Companies are doing it all the time right now, but are not just giving it away.
@terbospeed
@terbospeed 7 ай бұрын
Agents, Mixture of Experts, RAG, etc
@joelandrews8041
@joelandrews8041 7 ай бұрын
@@helix8847 thanks for this. I'd like to learn more about this @Computerphile!
@JayS.-mm3qr
@JayS.-mm3qr 7 ай бұрын
Man.... they are developing advancements for everything. Any problem that you can think of, people are coming up ways to address with code. That is the magic of coding. You have a problem, and express the problem and solution in a virtual environment, in coded language that computers understand, and the computer outputs something that we understand. Have you heard of AI agents? Those address the thing you asked about. It turns out that using multiple llm's, and developing the ability for ai to complete tasks, makes them a lot more effective. This is true without increase data size. Yes, models are becoming better, without new data.
@Archimedeeez
@Archimedeeez 7 ай бұрын
the camera work is enthralling.
@Rahvin1230
@Rahvin1230 7 ай бұрын
I’m an economist, color me impressed. The computer guys are finally catching on. Very cool.
@danremenyi1179
@danremenyi1179 7 ай бұрын
This is a very old concept in economics called the Law of Diminishing Returns. And it applies to a lot of different aspects of life.
@lazyraccoon1526
@lazyraccoon1526 4 ай бұрын
There is no intrinsic reason to expect diminishing returns
@demonzabrak
@demonzabrak 4 ай бұрын
@@lazyraccoon1526 study more about entropy
@niello5944
@niello5944 7 ай бұрын
The problem with relying too much in AI in the medical field that I have been worried about is that it'd mean overall less practice for doctors at identifying the problem. If most doctors are lacking in the experience that they should have gain with years of doing the job, then they'll lack experience to identify the more obscure diseases and underlying problems. That isn't something AI can cover, especially when these kinds of things can change without us knowing or deprive us of new discoveries.
@stu8924
@stu8924 4 ай бұрын
Thank you for such an enlightening video. I have been following your channel for some time now and you should be applauded for the way each of your team are able to translate complicated topics into information consumable to by non-academic folks like me. It sounds very similar to when a person first starts a diet or commence a fitness program, at first the person will see fantastic results, they’ll look and feel fabulous, but, as they continue the same routine, the results begin to diminish, as their body adapts to its new norm. It is at this point that the person needs to cycle their routine, make alterations/variations to their diet or fitness program. These Gen AI models need to cycle their routines or make adjusts to their current fitness programs, and, over time, they will once again see fantastic results.
@luketurner314
@luketurner314 7 ай бұрын
2:04 "Always go back to the data" -Larry Fleinhardt (Numb3rs)
@Quargos
@Quargos 7 ай бұрын
Honestly, it sounds like the conclusion you're coming to, about it not being able to infer to new things that aren't very well represented is just a kind of simpler thing that feels like it ought to be obvious: The nature of the dataset informs the nature of the resulting model. If the dataset doesn't have the information to differentiate between tree species (or if the info it has there is highly limited), then of course the model won't either. The model is simply a "compressed" form of the information fed in, and not other info. That "You show it enough cats and dogs and the elephant is just implied" as said at the start can never hold. Because if you've never seen an elephant, how could you ever be expected to recognise or draw one? I do not believe that extrapolation to anything outside of the dataset will ever truly work.
@peterisawesomeplease
@peterisawesomeplease 7 ай бұрын
I don't like the elephant example either. But I think the point of the paper isn't just that if the data is missing the model won't be able to handle it. The point is that you exponentially growing amounts of data to get linear increases in performance. And we are already running out of high quality data to feed models.
@wmpx34
@wmpx34 7 ай бұрын
I agree, but it sounds like many AI researchers don’t. So where’s the disconnect? Either we are wrong or they are overly optimistic. Like the guy in this video says, I guess we will see in a couple years.
@alexismakesgames6532
@alexismakesgames6532 7 ай бұрын
The "elephant" speaks to human creativity. It's the ability to make new things based on base principles. Maybe being trained on only cats and dogs is too optimistic to make an elephant but say it also had some other animals so it knew what "horns" were ect. The hope is you could then give it a very detailed description involving concepts it already knew and hopefully get an elephant. Then you could teach it "elephant" and it basically becomes able to learn anything that can be described. There are a lot of other steps but this is one of the keys to having AGI. I agree though, it is terribly optimistic to think this will happen with the current ML models. Which is my main problem with them, they pretty much regurgitate only what they have a lot of basis for and become very derivative and patchy in areas where the data is thin.
@lopypop
@lopypop 7 ай бұрын
In the same way that it will probably eventually be able to solve math problems it hasn't been explicitly trained on, I think the argument goes that once it "understands" enough fundamentals of the real world, it can extrapolate out quite creatively. This won't work with the example of recalling strict factual data that it hasn't been trained on ("draw me an elephant" ), but it might work with enough prompting to get something reasonable (generate an image of a large land mammal of a specific size, anatomical properties, and other qualities). It's possible that it generates an image that looks like an elephant without ever being trained on elephant photos
@Ylyrra
@Ylyrra 7 ай бұрын
@@wmpx34 Most AI "researchers" aren't asking the question. They're too busy being engineers and looking to improve their AI, not look at the bigger picture. The bigger-picture people at all those companies have a vested interest in a product to sell that has a long history of hyperbole and "just around the corner" wishful thinking. You don't ask the people building a fusion reactor how many years away practical fusion is and expect an unbiased remotely accurate answer.
@AIWorkforceEvolution1
@AIWorkforceEvolution1 26 күн бұрын
These informative KZbin videos are incredibly interesting! I truly enjoy learning from them, and I’m also trying to share similar information with my viewers on my own channel
@KipIngram
@KipIngram 7 ай бұрын
I believe it. Our own human thinking isn't just "a huge amount of computing power." There's something else going on we don't understand - the machines are not ever going to be able to do everything we can do.
@paulfalke6227
@paulfalke6227 Ай бұрын
The creation can not be greater then the creator. Therefore, our AI products will stay forever below our own intelligence capabilities.
@KipIngram
@KipIngram Ай бұрын
@@paulfalke6227 Well, to the extent we do things via step-by-step process (and we do do that sometimes), the AIs will outclass us. They just have so much more SPEED, and they're getting more all the time. What they'll never be able to do is the "flashes of inspiration" that we have - the intuitive "gut instinct" stuff that turns out right way more often than random luck. That's the part of us that isn't "mechanized" and will never be within reach of a machine. And I think all of that is closely connected with how we're able to actually "be aware" of the world around us - machines will never do that either. It will never be "ending a life" when we demolish a machine.
@Cheezitnator
@Cheezitnator 4 ай бұрын
Also literally nobody wants generative AI except the corpos selling it, the corpos who don't want to pay artists/writers, and art thieves. Incidentally, these are all the same entities. So it's only motivated by greed and trend chasing which can only be bad for everyone.
@cefcephatus
@cefcephatus 7 ай бұрын
Logarithm doesn't have limit close to infinity, but rate of improvement (derivative) gets to 0 at close to infinity for sure. That's the technical term of "flatten out". And the reality that causes this growth is usually the very foundation of conditional statistics, Bayes' Theorem. No matter how you improve the learning algorithm, but if the output is determined by summing data distributions, the result will be Sigmoid function. This is worst than pure logarithmic function, because it actually converges and has limit at infinity, so there exists the end of this approach. If you want to improve AI, don't think about applying statistics state machine, you'll have to embrace superposition and code the quantum wave function in to it. Yes, it still uses Bayes' Theorem to discern an output, and the solution is more complicated in traditional computers. But instead of sigmoid, now we have chaos to deal with, which sounds cool if we could stay on the upper branch of each bifercations.
@jessebclimbs
@jessebclimbs 7 ай бұрын
duh. everybody knows this.
@cefcephatus
@cefcephatus 7 ай бұрын
@@jessebclimbs Yeah, everyone knows everything, statistically.
@jonathonreed2417
@jonathonreed2417 7 ай бұрын
This was an interesting video considering public discussion of “power laws”. I hope you do another video about “synthetic data” which is being discussed now, what is it exactly why someone would want to use it drawbacks etc. I’m personally in the skeptical camp but it would be interesting to hear an academic answer to these questions.
@scampercom
@scampercom 7 ай бұрын
Came here to make sure someone mentioned this.
@bilbo_gamers6417
@bilbo_gamers6417 7 ай бұрын
Synthetic data is going to be huge, and will bring a significant performance increase as time goes on. All of OpenAIs data will likely be generated in-house in addition to the data they currently have from before the low bar AI generated dross began to pollute the internet.
@taragnor
@taragnor 5 ай бұрын
The problem is that synthetic data can largely miss the point of finding the exceptions. Synthetic data will be based off some pattern, meaning it's going to create permutations of the existing data, and likely that's going to fall mostly in the most common cases. The real problems with training AI is getting data on edge case scenarios. It needs to see things it hasn't seen before.
@seankelley3619
@seankelley3619 7 ай бұрын
Even other big tech companies are admitting that the only reason their models don’t do as well as OpenAI is because OpenAI has such an ungodly amount of data (a lot of which is stolen). And OpenAI has eluded to the fact that they’re running out… Sounds like a plateau to me.
@Mattje8
@Mattje8 3 ай бұрын
Super interesting. Also worth considering the implications of some of the new results suggesting quantum effects in consciousness may well be “a thing” (as postulated by Penrose et al). I’ll admit to being heavily influenced by Penrose because maths, but the Gödel theorem implications to me suggest that no generative algorithm (at least based on current computational technologies) can ever be as creative as a human, because at the end of the day they can only make “statements” (which of course can be expressed as images) within the rules.
@rayjaymor8754
@rayjaymor8754 7 ай бұрын
AI is autocorrect on steroids. As far as I can tell it's a great tool, I can get it to write some simple javascript stuff for me way faster than I can type it out myself. But it can't create anything "new" it just re-iterates concepts that already exist and are well documented. Now granted, it can do this faster than I as a human can. But if you need it to come up with an idea or a new way of doing something - that is where it falls apart.
@Pabz2030
@Pabz2030 7 ай бұрын
I'm not sure that a "world spanning AGI that can do everything" is actually the correct answer. Surely we are better off with highly specialised (relatively speaking) AI/Bots that can be cheaply manufactured, lowish power consumption and can train using swarm learning techniques? The idea is to make life better for Humans by removing the drudgery not replace them entirely
@dahahaka
@dahahaka 7 ай бұрын
I mean if you replace them entirely we are free to do whatever we want....
@krrishgoyal6567
@krrishgoyal6567 7 ай бұрын
btw many of the generalist AI(s) are performing better than specialist AI(s)
@dahahaka
@dahahaka 7 ай бұрын
​@@krrishgoyal6567 Yup, this! You can't just disregard general information and say "yeah data ain't enough" obviously training on one subset of data is only going to get you so far without generalized information. I do think Vishaal Udandarao's paper is being misrepresented here tho, haven't completely read through it yet though
@user34274
@user34274 7 ай бұрын
I absolutely love this man and the way he explains things. I usually feel so stupid with topics like these when other people are talking about it but this man explains things is very easy terms. This is also a testament to how well he knows what he is talking about.
@clray123
@clray123 7 ай бұрын
I'm not sure why you need complicated papers or long videos on this topic. A language model approximates the conditional probability distribution of the dataset it is trained on. Even if it (ignoring practical memory limits) mirrored the perfect probability distribution of all the words ever uttered by human race, it would be just that, a perfect tool for reproducing words that have been already uttered, while matching frequencies contained in the training set. As for generating anything that has not been uttered or so-called "generalizing", it's up to the researchers to prove that they mean by that word. As far as I'm concerned, I see no "generalizing" or "reasoning" going on whatsoever, just copying (with errors) what has already been put in during the training. That out-of-distribution generations stay close to the in-distribution ones creates an illusion of being "creative" or "able to generalize", while in reality it is more like "idiotically picking the semantically closest words and hoping we get back on track [i.e. in-distribution again]". P.S. Adding more data just means that your model is less likely to "derail" into completely out-of-distribution territory and keep generating garbage. Kind of like these "roly poly" dolls that stand up again after you push them over.
@97DarkSkull
@97DarkSkull 6 ай бұрын
What about models knowing languages they should know or creating their own languages ?
@Real-HumanBeing
@Real-HumanBeing 6 ай бұрын
Great comment. A better answer than one in the dataset is also a deviation from the prediction. This inherently cannot scale beyond the dataset. It’s a new form of contextual storage, not actual reasoning.
@owlmostdead9492
@owlmostdead9492 7 ай бұрын
It’ll plateau because there’s basically no intelligent design, it’s a brute-force comprised algorithm that’ll never evolve beyond the dataset.
@SmileyEmoji42
@SmileyEmoji42 7 ай бұрын
That's not the problem. It's not even "expert" on the entire dataset, only on the common bits.
@taragnor
@taragnor 7 ай бұрын
Well the algorithm itself is intelligently designed, but each algorithm has its limits. Honestly it's hard to believe AI intelligence is limitless. Even biological intelligence is heavily capped by your brain. Sure you can learn things, but your actual reasoning capability eventually hits a wall.
@fakecubed
@fakecubed 7 ай бұрын
It's designed to the level of our current human intelligence.
@menosproblemos6993
@menosproblemos6993 7 ай бұрын
Using a calculation system where numbers are a subcategory of calculations, but the main calculus is based on for example size: atom
@johnarnebirkeland
@johnarnebirkeland 7 ай бұрын
Expecting AGI by increasing LLM size and complexity, sounds a lot like emergence in complex systems theory. I.e. there is precedence for this happening in biology etc. but there is absolutely no guarantee that it will happen, or if there is emergence that it will be anything useful in the context of AGI interacting with humans. But then again you could also argue that the LLM results we currently see already is proof of emergence.
@Qacona
@Qacona 7 ай бұрын
I suspect that we'll develop models that are able to fool humanity into thinking they're AGI long before we actually hit 'real AGI'.
@kneesnap1041
@kneesnap1041 7 ай бұрын
I think it's more likely a new technique / algorithm, or rather a collection of them, rather than a complex system creating unexpected behavior through just sheer data size. Think about how much "training data" a person gets. We only need to see a handful of cats & dogs before we can reliably tell them apart. We don't need dozens, hundreds, or thousands. To imagine that massive datasets are an absolute requirement for AI seems a bit unlikely to me because of this.
@Real-HumanBeing
@Real-HumanBeing 6 ай бұрын
They aren’t. They’re proof of the scalability of contexts from the dataset, but fact of the matter is, giving an answer that is more intelligent than the dataset is also a deviation from probabilities of the dataset, just as a worse answer is. In short, it’s the dataset.
@thiagopinheiromusic
@thiagopinheiromusic 6 ай бұрын
It's a fascinating debate. While the current advancements in LLMs are remarkable and suggestive of emergent properties, relying solely on scaling up models for AGI might be overly optimistic. A more targeted approach, possibly integrating other AI techniques and innovations, could be necessary to achieve meaningful AGI.
@Awould-m5s
@Awould-m5s 7 ай бұрын
This was my first instinctive thought about how useful ai will be in the long run, for this type of thing. Thank you for putting it into words.
@fakecubed
@fakecubed 7 ай бұрын
It won't ever be less useful than it is right now, and it's already very useful. The logarithmic growth may be true, or not, but we're still in the very steep part of it.
@Jesse_359
@Jesse_359 7 ай бұрын
There are few truly exponential progressions in the real universe, and certainly not in the kinds of complex systems we actually interact with. They're all Logistic Curves, which behave exponentially up to a point, and then start hitting progressively more substantial barriers to growth which exponentially *slow* their further growth beyond an inflection point. The end of Moores Law is an excellent example of this - but it applies to nearly everything.
@gloweye
@gloweye 7 ай бұрын
I'm highly confident that AI will plateau. I don't know where, but it will. LLM techniques do not lend themselves to generic AI, let alone self-aware ones.
@jimmy21584
@jimmy21584 7 ай бұрын
Considering the open source community’s achievements with model building, fine tuning and optimisation, I think the really interesting improvements will come from there, and not so much from throwing more data into training and stacking more layers.
@fakecubed
@fakecubed 7 ай бұрын
Yeah, the open source community is where all the best innovations are happening in figuring out how to make the most out of the models that are out there, but big investor money paying for PhDs to do research gets us important mathematical innovations and big volumes of training data. Both are necessary.
@sebastiang7394
@sebastiang7394 7 ай бұрын
Is it really tough? Unix was a commercial product. The first desktops were commercial products. In what discipline does open source lead the technical curve.
@fakecubed
@fakecubed 7 ай бұрын
@@sebastiang7394 The first desktops were home-built kit computers. Open source doesn't mean non-commercial.
@sebastiang7394
@sebastiang7394 7 ай бұрын
@@fakecubed Ok so you're talking about a desktop pc. Not the concept of a gui. Misunderstanding on my part. Sorry about that. Yeah tinkerers definitively do a lot of great stuff. Especially in the early days of computing and still today. But I think it's hard to argue that the majority of innovation happens in the open source world. Just look at all the stuff that happened at Bell labs or IBM. Basically the beginning of Open Source (the GNU project) aimed at reproducing all the commercial tools that already existed. They basically aimed at recreating the UNIX ecosystem. And still today most big open source projects are either backed by large cooperations or are rather small in scale.
@tiffanyl9827
@tiffanyl9827 5 ай бұрын
Thanks for this paper explaination. I do agree we can't just rely on transformers and other statistical method hopping we will reach with it the peak of intelligence. It's working and as it working we are "hallucinate" the fact it can do and mimic us. But Backproagation doesn't happen in the brain, nor Maxpooling as Hinton showed up in 2018 with Capsule Network, nor else.
@johnpwmcgrath
@johnpwmcgrath 4 ай бұрын
It would be really nice if AI companies didn’t need to steal everyone else’s work to be successful
@robinvik1
@robinvik1 7 ай бұрын
An AI needs thousands of pictures in order to distinguish between a cat and a dog. A very small child need about 12 pictures. Clearly they are not learning the same way we are.
@97DarkSkull
@97DarkSkull 6 ай бұрын
That's not true. You can teach stable diffusion how you look from just few pictures
@robinvik1
@robinvik1 6 ай бұрын
@@97DarkSkull Stable fusion was trained on 2.3 billion images
@ericblair5731
@ericblair5731 2 ай бұрын
Clearly this person has never met a human child. Children learn from extensive real world exposure to household pets. Not from isolated images. Even if you don't have a dog at home they meet numerous dogs and have extended interactions with those animals from the comfort of their stroller long before they can speak. So no they don't get just a dozen images. They get hours of real world exposure.
@BiggumsMcHoney
@BiggumsMcHoney 2 ай бұрын
ML models more so create a mathematical imprint of all the data that was fed in, and differences in i.e. raw pixel data are worked out. They don’t “learn” at all. “Learning” was just a helpful analogy that has gone way too far and confused the public. Now even some software engineers are confused and think LLMs can “reason”.
@Eric-gq6ip
@Eric-gq6ip 4 ай бұрын
This matches my observation, "AI" right now is great at regurgitating things from its dataset, but fails miserably as soon as you step outside that and its problem solving abilities are pretty much nonexistent. Throwing more data at the models won't magically make it know how to solve problems since fundamentally the systems have no intelligence or real understanding of anything.
AI can't cross this line and we don't know why.
24:07
Welch Labs
Рет қаралды 1,4 МЛН
AI Deception: How Tech Companies Are Fooling Us
18:59
ColdFusion
Рет қаралды 2 МЛН
Правильный подход к детям
00:18
Beatrise
Рет қаралды 11 МЛН
REAL or FAKE? #beatbox #tiktok
01:03
BeatboxJCOP
Рет қаралды 18 МЛН
XZ Exploit - Computerphile
14:41
Computerphile
Рет қаралды 46 М.
Are AI Advancements Already Slowing Down?
13:29
CNBC
Рет қаралды 375 М.
How AI 'Understands' Images (CLIP) - Computerphile
18:05
Computerphile
Рет қаралды 219 М.
2024's Biggest Breakthroughs in Math
15:13
Quanta Magazine
Рет қаралды 438 М.
Creating Your Own Programming Language - Computerphile
21:15
Computerphile
Рет қаралды 192 М.
Google’s Quantum Chip: Did We Just Tap Into Parallel Universes?
9:34
The moment we stopped understanding AI [AlexNet]
17:38
Welch Labs
Рет қаралды 1,4 МЛН
How Quantum Computers Break The Internet... Starting Now
24:29
Veritasium
Рет қаралды 10 МЛН
Правильный подход к детям
00:18
Beatrise
Рет қаралды 11 МЛН