The Right Way To Train AGI Is Just GOOD Data?

  Рет қаралды 31,866

bycloud

bycloud

Күн бұрын

Пікірлер: 146
@bycloudAI
@bycloudAI 2 ай бұрын
To try everything Brilliant has to offer-free-for a full 30 days, visit brilliant.org/bycloud/ . You’ll also get 20% off an annual premium subscription! (I reuploaded this video cuz there was a pretty big mistake at 6:13, sorry notifications!)
@ghulammahboobahmadsiddique8272
@ghulammahboobahmadsiddique8272 2 ай бұрын
What was the mistake? I watched the original one so I don't want to rewatch this just to know what's changed. So could you please say what the mistake was?
@heys3th
@heys3th 2 ай бұрын
@@ghulammahboobahmadsiddique8272 They had the wrong graphic/text for Class 4 Complex
@bycloudAI
@bycloudAI 2 ай бұрын
@@ghulammahboobahmadsiddique8272 6:13 i showed class 3 chaos twice, i needa catch some sleep lol
@jamesgreen.3271
@jamesgreen.3271 2 ай бұрын
​@@ghulammahboobahmadsiddique82726:16 Here class 3 and 4 had the same captions and images. But now he fixed it
@venadore
@venadore 2 ай бұрын
Your editor still missed it in like two spots lmao
@weirdo8435
@weirdo8435 2 ай бұрын
such a beautiful knowledge, i will not use it anywhere and not talk about it with anybody.
@rmt3589
@rmt3589 2 ай бұрын
I will! Gotta make my own AGI somehow.
@rawallon
@rawallon 2 ай бұрын
​@@rmt3589that would be easier than actually making friends
@weltonbarbosa206
@weltonbarbosa206 2 ай бұрын
ironically I feel the same... There is absolutely no one I know that would even understand the beautiful in this concept comparing to human brain functions, the shear realization of how we perceives or own reality in the middle of caos. It's not total caos , it's just complex order.
@Ikbeneengeit
@Ikbeneengeit 2 ай бұрын
I fell asleep while listening to this video and dropped my phone on my wife's head and now she's mad.
@lionelmessisburner7393
@lionelmessisburner7393 2 ай бұрын
Wow
@Zeni-th.
@Zeni-th. 2 ай бұрын
Wow
@fnytnqsladcgqlefzcqxlzlcgj9220
@fnytnqsladcgqlefzcqxlzlcgj9220 2 ай бұрын
Wow.
@pingpong1727
@pingpong1727 2 ай бұрын
Wow
@skit555
@skit555 2 ай бұрын
Wow
@heys3th
@heys3th 2 ай бұрын
These videos are so nice for someone like me with no technical background in machine learning. Thank you and please keep making more!
@GaiusAnonymous
@GaiusAnonymous 2 ай бұрын
Careful, last time I said "pls never change" he changed a week later.
@OperationDarkside
@OperationDarkside 2 ай бұрын
This feels somehow similar how physics is based on a seemingly simple set of rules, yet creates impossibly complex situations/states. There must be a limited set of core rules a base model needs to learn to become an effective reasoner.
@hugh8709
@hugh8709 2 ай бұрын
It is very similar. These rules are called elementary cellular automata and the naming scheme was developed by the physicist Stephen Wolfram. He has a theory to look for something analogous to explain complex physical phenomena. It has something to do with hypergraphs (I think sabine hossenfelder has a video on it). The connections between complexity theory, physics, machine learning, and intelligence is extremely interesting.
@DeniSaputta
@DeniSaputta 2 ай бұрын
2:31 Lack of words like "skibidi"
@poipoi300
@poipoi300 2 ай бұрын
Multi-step prediction has been known for a while to perform poorly. It's best to either predict probabilities and sample or predict a single timestep, recursing for more. LLMs are doing both.
@ZeroRelevance
@ZeroRelevance 2 ай бұрын
It makes a lot of sense - to predict five steps in advance you’d need to predict one step 5 times in a row, and you only run the model once, so it’d have to make more shortcuts with each step’s prediction given how it has to fit it five times in the same space, accumulating errors in the process.
@poipoi300
@poipoi300 2 ай бұрын
@@ZeroRelevance What you said is true, but I think it's still possible to implement multi-step prediction in a performant manner. It would depend on the specific problem, but generally I can see a lot of instances where timestep 5 does not rely on timestep 4 or maybe even 3, so there is no error to accumulate from those steps. Currently, one of the big drawbacks of predicting multiple steps (this is true of predicting multiple values for one step as well) is that the loss associated with each predicted value is only weakly accounted for and chasing the gradient for an average increase in performance is likely to make some of the predicted values worse. What we need are better feedback mechanisms and more channels. MoE are a sort of rudimentary solution to channels, but we're still relying on SGD of the whole network and in some instances manual freezing, which I don't really like from a technical standpoint. We need to be able to decompose problems into multiple losses, but that might not even be possible depending on the problem.
@adamrak7560
@adamrak7560 2 ай бұрын
You did not mention that rule 110 is Turing complete. It may be not because of the edge of chaos, but because of the Turing completeness. All Turing complete systems behave generally similar to what they define as edge of chaos. Although you can construct some which hides this under apparent noise.
@rmt3589
@rmt3589 2 ай бұрын
GPT-3 is already turning complete. It's a bad test. Edit: I mixed up Turing Test with Turing Complete. The above post makes no sense with it's context.
@owenpawling3956
@owenpawling3956 2 ай бұрын
@@rmt3589 I believe you are confusing the Turing test aka the imitation game with Turing completeness. Turing completeness refers to whether something can be used to simulate a Turing machine, which makes it computationally universal.
@4.0.4
@4.0.4 2 ай бұрын
​@@owenpawling3956 nothing "passes" the Turing test, as it depends on the participants. But LLM are somewhat Turing complete if you assume infinite context to use as "tape".
@rmt3589
@rmt3589 2 ай бұрын
@@owenpawling3956 I was. You are correct. I'm also technically not wrong, but did not convey what I wanted to. I will go fix my post.
@rmt3589
@rmt3589 2 ай бұрын
@@4.0.4 LLMs long passed the Turing Test. Back when widespread conspiracy about Replika, which ran on GPT-3, being real people pretending to be AI. Now with so many fake AIs, people literally cannot tell what's human and what's AI, and keep getting surprised when one turns out to be the other. This is a perfect and natural Turing Test, and is being passed with flying colors.
@4.0.4
@4.0.4 2 ай бұрын
The "long tail" really explains why AI slop is so mid - it is literally the middle of the distribution of language. And you can see it in most models, even if different wording is used.
@CYI3ERPUNK
@CYI3ERPUNK 2 ай бұрын
very carefully , and with compassion and wisdom
@fnytnqsladcgqlefzcqxlzlcgj9220
@fnytnqsladcgqlefzcqxlzlcgj9220 2 ай бұрын
Wolfram would love this
@tiagotiagot
@tiagotiagot 2 ай бұрын
Kinda reminds me of how some fighter-jets are designed to be on the edge of instability allowing for more extreme maneuverability by controlling when to lose control
@AfifFarhati
@AfifFarhati 28 күн бұрын
can you elaorate a little, i have zero knowledge of fighter jets but the concept you described is so interesting.
@tiagotiagot
@tiagotiagot 28 күн бұрын
@@AfifFarhati Basically, airplanes are usually designed so if you let go of the controls under normal conditions they stay pointing the same way or even level off, both side-to-side and up and down, by being shaped such that depending on the angle and speed they're going thru the air, the balance between asymmetrical drag and lift forces automatically push it back to the stable condition; but that means the plane is fighting being rotated in any way away from the stable condition. With some fighter-jets, they're shaped so that effect is very weak (or restricted to a very small range of speeds and angles relative to the motion thru the air), or even works the opposite, and a human would have a really hard time flying it directly, the flight-computer is what keeps it going straight with constant tweaks to the controls; so when the pilot wants to jerk the plane around really hard, they can push it to the edge of control, or even past it and the flight-computer constantly adjusts the control surfaces (and in some cases jet vectoring, the direction of the jet nozzle) with inhuman reflexes to keep the fighter jet under control, or regain control after deliberately causing some brief "out of control" tumbling. TL;DR: They're designed to lose control in a controlled manner
@tisfu17
@tisfu17 2 ай бұрын
It seems almost obvious that just chasing complexity horizons will lead to increasingly complex output potentials also, but to see how this can be done in practice, and related back to OG cell automata is very cool.
@GodbornNoven
@GodbornNoven 2 ай бұрын
Essentially, because the model learns to incorporate past states in its decision making, it becomes capable of better reasoning. AKA, this is just another case where transfer learning is truly an important key. Transfer learning aka generalization is also the reason why sample efficiency improves with training.
@cagedgandalf3472
@cagedgandalf3472 2 ай бұрын
What you mentioned reminds me of curriculum learning from RL. Start off training easy then gradually make it harder.
@ConnoisseurOfExistence
@ConnoisseurOfExistence 2 ай бұрын
Great video! I first heard about the brain being on the edge of chaos by Artem Kirsanov, who has a great channel (of the same name) on computational neuroscience. I'm thinking, those models that are trying to predict 5 steps at once might be ultimately better, but they would require much longer training (and maybe size), and therefore computational resources, to start learning some complex patterns. It could probably be tested with models that try predicting 2 steps ahead...
@MaxBrix
@MaxBrix 2 ай бұрын
If you train a model to predict sin waves from discrete data points it will approximate sin. The more training it gets the closer the approximation. The model does not learn the sin function. The benefit of this is that with missing or incorrect data the model can still approximate the correct sin wave where a real calculation would be completely wrong.
@ryanengel7736
@ryanengel7736 2 ай бұрын
very cool and interesting. I had some similar intuition and I'm glad you discussed this paper. Great work.
@zandrrlife
@zandrrlife 2 ай бұрын
Bro cover “Fourier Heads” or belief state transformer. Fourier head research is interesting, I see a lot of value of integrating Gaussian mixture model principles into LM to better handle complex distributions. To be honest. One of my core principles is disentanglement, there’s a reason why we don’t see expected performance gains with multimodal data and reasoning in general, the model treats it as a single continuous sequence, the solution I’ve been working is multivariate next-token prediction, where each modality is considered, and yes everything can be treated as distinct modality, even reasoning via structured reasoning tokens, instead of for T = sequence length, it would be N x T, where N is the modality count, almost like a time series problem, obviously increases sequence memory for the sequence, I’ve seen clear benefits and think it’s the future. Why I don’t expect legit breakthroughs from any of the top players. No new ideas. Rather no divergent ideas. AGI will be created by divergent thinkers. Someone already releases an entropix I believe it’s called, which recreates o1-preview style outputs lol, just needs DPO to really get that juice out. We need to fund our divergent thinkers.
@nikhilsultania170
@nikhilsultania170 Ай бұрын
Instead of taking the highest probability token , some models, assign the softmax probabilities to the token, and then use a random generator to select the token, so most likely taken is still most probable but not guaranteed, this is controlled by the temperature of the model, and the reason it generates unique outputs every run
@ashish54713
@ashish54713 2 ай бұрын
Would love for you to cover the things going on around Entropix!
@sahilx4954
@sahilx4954 2 ай бұрын
Thank you for your hard work. ❤ 🙏
@Lexxxco1
@Lexxxco1 2 ай бұрын
Great video as always, edge of chaos where understanding ends)
@termisher5676
@termisher5676 2 ай бұрын
What about infinite language model Working on entire output at once and constantly improving it Like it was looking at babel libry with all possible word combinations and looking for the best one using hiking 7z algorithm or smth but it'll need beefy evaluation ai You could then use debuggers or language engines to prune combinations that are certain to fail
@luisfable
@luisfable 16 күн бұрын
Mode collapse is not really because of synthetic data, it is because the model impossibility to learn different modes, it will instead learn a lot less modes, and this can have MANY different causes, not just using low diversity synthetic data
@MathewRenfro
@MathewRenfro 2 ай бұрын
Stephen wolfram a New Kind of Science and itscomputational theory applied to training models is the way to go, me thinks
@stephen-torrence
@stephen-torrence 2 ай бұрын
Read it almost 13 years ago and felt I was accessing some seriously Arcane shit that I was not ready to know. Amazing seeing it applied to transformers.
@ckq
@ckq 2 ай бұрын
Need someone to train an LLM on NBA scores. pretty simple but also trends like scoring effects comebacks run and momentum
@Game99Boss
@Game99Boss 2 ай бұрын
It's all about the "day daa" 😂
@redthunder6183
@redthunder6183 2 ай бұрын
This is basically just information theory…
@rmt3589
@rmt3589 2 ай бұрын
Looking it up. Ty.
@fnytnqsladcgqlefzcqxlzlcgj9220
@fnytnqsladcgqlefzcqxlzlcgj9220 2 ай бұрын
I have a playlist on my channel that has info about this, there are some good lecture sets linked in it, it's not in order so look through to find stuff, it's more of a directory. it's called advanced apaSiddhanta ​@@rmt3589
@fnytnqsladcgqlefzcqxlzlcgj9220
@fnytnqsladcgqlefzcqxlzlcgj9220 2 ай бұрын
@@rmt3589 I made a playlist about information theory, complexity, emergence etc
@warguy6474
@warguy6474 2 ай бұрын
not rly
@redthunder6183
@redthunder6183 2 ай бұрын
@ yeah, it’s exactly the same thing, read through the original information theory paper from like 70 years ago
@smartduck904
@smartduck904 2 ай бұрын
The proper way to train AGI is to put it in the square hole
@lionelmessisburner7393
@lionelmessisburner7393 2 ай бұрын
?
@smartduck904
@smartduck904 2 ай бұрын
@lionelmessisburner7393 everything goes in the square Hole
@warguy6474
@warguy6474 2 ай бұрын
are you stupid
@SkibidiOhioCreeperSigmaGyattRi
@SkibidiOhioCreeperSigmaGyattRi Ай бұрын
2:46 damn, I guess you had a good watch :v
@timeflex
@timeflex 2 ай бұрын
I think, AI should constantly assess the state of play and pick an appropriate strategy, but in order to do that, it should be able to self-reason. Something like o1 but with an extra degree of freedom.
@KingKogarasumaru
@KingKogarasumaru 2 ай бұрын
So in the end, training from small problem to big problem like us humans is the best way to get better? Would the patterns equal to personality in terms of human?
@clearmenser
@clearmenser 2 ай бұрын
Hmmm, reminds me of the Turing patterns on animal skin. I wonder if the human brain does this.
@ginqus
@ginqus 2 ай бұрын
why do i perfectly understand some of your videos, and at the same time get absolutely confused by others?? 😭😭
@robertburton432
@robertburton432 2 ай бұрын
The Pandoras box of programming?
@sgttomas
@sgttomas 2 ай бұрын
fascinating ideas 💡
@JaredQueiroz
@JaredQueiroz 2 ай бұрын
Well, AGI folks. Thats our future doodling chaotically like a baby... Good news: chaos was the answer and emergency is awesome. Bad news: that embryo can beat you at chess...
@warsin8641
@warsin8641 2 ай бұрын
Intelligence is fragile that's why it took so long to emerge
@michelprins
@michelprins 2 ай бұрын
great video thx
@kryptobash9728
@kryptobash9728 2 ай бұрын
great vid
@Zbezt
@Zbezt 2 ай бұрын
The collective human endeavor needs to mature before AGI can emerge in classical data its a simple fundamental that people dont understand
@brll5733
@brll5733 Ай бұрын
Question is how to generate rules for data in higher dimensional spaces. We will probbaly need more thna bit data to get AGI agents capable of operating in the real world.
@md6886
@md6886 2 ай бұрын
I have not seen single AGI yet.
@tendriel
@tendriel 2 ай бұрын
How so
@waterbloom1213
@waterbloom1213 2 ай бұрын
No one has and those that claim so are deluded. Mistaking advanced algorithms for reasoning or sentience is a problem among techbros and data scientists, although for very different reasons.
@agentofenhanced2428
@agentofenhanced2428 Ай бұрын
Agi is impossible, humans dont need an outside input to work, when you're born your parents dont say " cry now" you learn by seeing things, understanding and thinking about things, you have instincts and feelings and everything combined makes you human.and the most important thing we have is a mental framework thats constantly changing, a mental framework is to us what a system prompt for ai is, except we cant change the system prompts during runtime nor can the ai "learn" and change itself, how would you even go about writing a system prompt for an ai to behave like a human? It would be a million lines long " act like this when this happens, say this if someone says that" again ai literally isnt concious it can't ever survive without input from the thing that made it, this doesn't mean its useless as we are getting to the point where ai can train itself while running and thats huge progress and is useful.
@6AxisSage
@6AxisSage 2 ай бұрын
Been there, done that
@telotawa
@telotawa 2 ай бұрын
RLHF is the thing that makes them slop. base models are still way better at good writing
@Nimifosa
@Nimifosa 2 ай бұрын
Give me a pen and paper and i'll teach AGI anything and everything.
@retrofuturism
@retrofuturism 2 ай бұрын
Intelligence is the edge of chaos in a map territory feedback loop
@setop123
@setop123 2 ай бұрын
fascinating
@APozzi
@APozzi 2 ай бұрын
One extra point for the "Critical brain hypothesis", to become a factual theory.
@problemsolver3254
@problemsolver3254 2 ай бұрын
I'm hella suss about Intelligence at the Edge of Chaos some how you make a violine graph for complex rules despite having 2 data points.
@kryptobash9728
@kryptobash9728 2 ай бұрын
what did training on rule 30 do?
@drlordbasil
@drlordbasil 2 ай бұрын
I always say AI is evolving fully backwards, vision is one of basics after brain before language and logic. We are doing it completed in reverse.
@lionelmessisburner7393
@lionelmessisburner7393 2 ай бұрын
We’ll get there though. Not TOO far away I think
@io9021
@io9021 2 ай бұрын
LLMs meet game of life 🤯
@blinded6502
@blinded6502 2 ай бұрын
I believe AIs should be taught similarly to humans. First they need to understand simpler artificial cases to perfection, sometimes giving them realworld cases. Like give them spinning cubes, then switch yo concave meshes, and only then to compound scenes
@jamesgphillips91
@jamesgphillips91 2 ай бұрын
self reflexive knowledge graphs... llms are a piece, the reasoning needs to be in a separate non-black boxed system. Neuro-symbolic ftw
@pajeetsingh
@pajeetsingh 2 ай бұрын
Whichever supports my interests.
@hypervanse
@hypervanse 2 күн бұрын
that's called high level studying
@hypervanse
@hypervanse 2 күн бұрын
Isn't obvious ?
@eRuTAmMi_gNieB
@eRuTAmMi_gNieB 2 ай бұрын
Where's the 34th rule?
@chelol208
@chelol208 2 ай бұрын
Kazuma licks Aqua holes, by the way Colette from brawl stars have tasty legs 😋
@ThangTatNaoNguyenHuuTri
@ThangTatNaoNguyenHuuTri 2 ай бұрын
Or you can just understand intelligence down to a fundamental level.
@ainet8415
@ainet8415 2 ай бұрын
I can't understand
@S8N4747
@S8N4747 2 ай бұрын
What does rule 46 do? 🐾
@vigoworkchannel1681
@vigoworkchannel1681 2 ай бұрын
Keep trying to break the wall set by God. I can’t wait to see the next Magic
@Napert
@Napert 2 ай бұрын
Garbage in, garbage out
@oonaonoff4878
@oonaonoff4878 2 ай бұрын
not sierpinskis triangle
@JinKee
@JinKee 2 ай бұрын
The year is 2029 and the first AGI was raised Catholic
@shodanxx
@shodanxx 2 ай бұрын
IN PUBLIC
@Илья-у9в5с
@Илья-у9в5с 2 ай бұрын
dream
@demonicedone
@demonicedone 2 ай бұрын
🐢🐢🐊🦖
@NLPprompter
@NLPprompter 2 ай бұрын
whoa whoa dude slow down.... need another video please this time slower.... please
@panzerofthelake4460
@panzerofthelake4460 2 ай бұрын
why does this video sound like it's AI generated tho
@float32
@float32 2 ай бұрын
AI reaches bycloud level of intelligence. - 2024
@bokuboke482
@bokuboke482 2 ай бұрын
Am I alone in noticing that A.I. learns a lot like bio brains? Study, reflect and connect, repeat incrementally to make REAL gains in comprehension and knowledge!
@sirtom3011
@sirtom3011 2 ай бұрын
You don’t…😂
@Supermayne555
@Supermayne555 2 ай бұрын
Thumbs down for the singing voice. It’s annoying. Ending with a high pitch like if Becky is gossiping with Amanda all day long
Why "Grokking" AI Would Be A Key To AGI
13:36
bycloud
Рет қаралды 74 М.
Why AI Simulated DOOM Is Actually Absurd
13:20
bycloud
Рет қаралды 128 М.
1% vs 100% #beatbox #tiktok
01:10
BeatboxJCOP
Рет қаралды 67 МЛН
IL'HAN - Qalqam | Official Music Video
03:17
Ilhan Ihsanov
Рет қаралды 700 М.
How to treat Acne💉
00:31
ISSEI / いっせい
Рет қаралды 108 МЛН
It works #beatbox #tiktok
00:34
BeatboxJCOP
Рет қаралды 41 МЛН
The Unreasonable Effectiveness of Prompt "Engineering"
15:12
OpenAI o1's New Paradigm: Test-Time Compute Explained
15:44
1 Million Tiny Experts in an AI? Fine-Grained MoE Explained
12:29
AI Is Not Designed for You
8:29
No Boilerplate
Рет қаралды 354 М.
o3: Smartest & Most Expensive AI Ever… With A Catch
12:40
Is Signal Processing The CURE For AI's ADHD?
11:53
bycloud
Рет қаралды 24 М.
AI can't cross this line and we don't know why.
24:07
Welch Labs
Рет қаралды 1,5 МЛН
The Dark Matter of AI [Mechanistic Interpretability]
24:09
Welch Labs
Рет қаралды 145 М.
1% vs 100% #beatbox #tiktok
01:10
BeatboxJCOP
Рет қаралды 67 МЛН