9:09 Dude, really good point about human priors affecting our understanding of quantum mechanics. Also, I'm listening to this on studio monitors and the intro music woke up the astronauts on the ISS.
@alexeypolevoybass3 жыл бұрын
So I'm not the only one with studio monitors here. Nice.
@Vedranation8 ай бұрын
I was wearing headphones and holy shit my ears weer blasted
@henrikf87774 жыл бұрын
My two cents is that the "core algorithms" in our brain isn't necessarily better it's just that we're born with a lot of really well-optimized algorithms already. Human babies can already detect faces and look into faces of people and as you said, imitate people. Recognizing objects in 3D like this and mapping it to the agents own body would on a blank slate require lots of training. I don't think we should look to mimic these well-optimized modules (like making an algorithm that learns by imitation) but instead get to the core of what allows them to be created in the first place.
@4AlexeyR Жыл бұрын
I agree with that. It is much more complicated process. Yes, one point of view that a human (and other creatures) have some pre-defined well tuned algorithms. And this fact is part of miracle of a life. The human brain a more complicated system. Thus it has own schedule of body evolution in time. And in this case we can see different stages of learning and understanding in quality and quantity. You can find some literature about it. For example, when a child starts to identify similarity of objects in a different scales (as I remember it starts after 3 years old age cca). And a lot of other body growing properties there are. It looks there is a very wise program or plan how to learn a mind step by step according to the state in a time.
@EdViaja5 жыл бұрын
This is the nicest channel about AI (personal account) I saw in KZbin. Continue with this work, it is excellent. RL has a big potential in real-world problems.
@faizanahemad5 жыл бұрын
Hi Xander. Great Video. An orthogonal question: How do you find such interesting papers? Like what do you use as source and how do you filter what to read and what to ignore? Possibly a vlog on how to actually find good recent papers and how to decide what to read??
@ArxivInsights5 жыл бұрын
Great question, since many people have already asked I wrote a blogpost on this a while ago: blog.ml6.eu/catching-the-ai-train-c0c496959999 But it kinda takes a while to create your own, healthy filter bubble in our digital media mess. Give it soms time, and remember: all those recommendation engines don't train themselves, so give feedback whenever you have that option! :p
@faizanahemad5 жыл бұрын
@@ArxivInsights Thanks Xander! Went through your blog, great ideas. and keep making these vids, they are awesome.
@ericfeuilleaubois407 жыл бұрын
Damn great video! Carry on ! Makes it very easy to get into these advanced subjects :)
@alexanderkurz2409 Жыл бұрын
5:03 "to test the presence and influence of different kinds of human priors" ... this is pretty cool ...
@davidm.johnston89946 жыл бұрын
Nice video man, just subscribed :-) If I may give you just a bit of criticism, watch out the audio levels because the ending music and some "pop" sound fx in the beginning were much louder than the speech.
@zmajoslavomirdedamrazizovi42907 жыл бұрын
i liked and subbed, but the intro and outro scenes blew out my eardrums and gave me a mini heart attack so i had to panically lower the volume, as well as adjust it during and after the video with the babies...
@ArxivInsights7 жыл бұрын
Good feedback, I'll definitely check this in my next edits before exporting!
@zmajoslavomirdedamrazizovi42907 жыл бұрын
i noticed this with a few other youtubers i follow in their early videos... i have high hopes for your channel :D
@ArxivInsights7 жыл бұрын
Fixed the audio levels in the new video :p Thanks for the good feedback! ;-)
@zmajoslavomirdedamrazizovi42907 жыл бұрын
yea, I saw man, it's awesome :D I think i was the second viewer, but i didn't want me to be the first comment again :) I hope you realize that when you reach a 100k subs, I'm gonna be telling people that i fixed your channel and that I'm the reason :D (kidding, to my friends only)
@maheshpatel20052 жыл бұрын
Very nicely explained..thanks a lot
@HeduAI6 жыл бұрын
What an awesome video!!!! Thank you! The edited games examples blew my mind...
@swapnilmeshram99916 жыл бұрын
great video. You are really doing something different from others who present knowledge on ai. Depth of knowledge is amazing.
@JKarioun6 жыл бұрын
This is great content, thank you for doing those videos, looking forward to the next ones!
@maskedman68903 жыл бұрын
Well there is a saying that you have to have done a task for 10,000 hours to be actually great at it. And adult humans when they turn 18 years old they have already spent 155,520 hours. In that time a person could only be best at 15.552 things. But another variable that comes is the involuntary actions we performs like walking, running, balancing. If we have done those actions then they just become muscle memory. So then you could read while walking, or sing when running or whatever. Also it doesn't affect the performance of the other task we are doing even though multitasking. So probably we could be great at more than 15.552 tasks, or even a 50 or 100 tasks if you are being productive 100% of the time. The point is that humans take years to perform these many tasks efficiently, so robots/simulation would also need years of training to be a conscious AI or something. I mean if it's possible for a single a program/robot to accumulate all the tasks it is learning throughout the years .
@ativjoshi10496 жыл бұрын
Elegant and clear explanation. Great speaker and nice content.
@inkwhir6 жыл бұрын
Wonderful video!!! Thank you for sharing this paper!
@iggoncharenko6 жыл бұрын
inkwhir мултики😉😑
@bernardfinucane20617 жыл бұрын
The concept of "priors" is what deep learning is about. The shallow layers are the priors of the deep layers. In image classification, edges are the priors of shapes, shapes are the priors of patterns, patterns are the priors of objects and so on. It would be interesting to do transfer learning to see how well a network trained with supervised learning on internet images could use its priors to do on a RL task like this.
@alexanderkurz2409 Жыл бұрын
"The shallow layers are the priors of the deep layers." Yes, and I will remember this quote. But humans like have priors that are not learned (as Chomsky famously has been arguing since the 1950s).
@NikosKatsikanis6 жыл бұрын
Great videos on making complex stuff approachable
@ankaandrews10932 жыл бұрын
This video is excellent!!
@lorenzoblz7996 жыл бұрын
I always wonder how many other "abstract" priors are actually learned by the model. For example, playing breakout I think that the model does learn what a trajectory is, what a bounce is, even what time, space and movement and causation are. And even "identifying" itself with the white paddle. Not in a sophisticated introspective/self-conscious way but in a very raw/"instinctive" way. If you place the paddle in a certain location and wait for the ball to arrive there are a lot of "concepts" you are relying on. Looking at this from the opposite perspective there are many more priors that humans bring to the game: space, time, the idea of a "puzzle", winning/loosing, dependencies (you need this to open that). Liked the paper and the video, thanks.
@ArxivInsights6 жыл бұрын
Hmm I'd beg to differ on that point actually. It's quite proven that current Deep Learning systems are unfortunately not doing much more than simple curve fitting (although these 'curves' might be very complex, like recognizing images for example). Check out this nice blogpost: www.vicarious.com/2017/08/07/general-game-playing-with-schema-networks/ They clearly point out that standard RL policies completely miss the underlying point off the game, they simply (over)fit actions to pixels through learned feature extractors. This is one of the biggest open problems right now in Deep Learning: how do we go from the powerful, gradient based curve fitting models to more 'general intelligence' type systems...
@lorenzoblz7996 жыл бұрын
The article discusses the possibility for the model to have a "human like conceptualization" of the world. And of course this is not true: the "world" the model "lives" in is a flat grid, not our physical world, there is no reason, and no possibility, for it to discover concepts similar to ours. Bricks are not "human bricks" to the model, just "those thing at the top that works in that way". Its "umwelt" (to use a big word) is completely different. Like the time and space in the "mind" of a wasp are different from the ones we use but, in my opionion, are nonetheless there with a similar role. I also disagree with the author concept of "small changes": these changes are small from the point of view of a human, who has a very specific conceptualization of the game. But these may be huge for the model, like changing fundamental physical constant for us humans (gravity, time "rate", etc.). In our mind we have the concept of bullets and their speed because we have seen different examples of them, otherwise we probably would not even have two categories to express these two concept. Like: let's make the time 10% taller. Dismissing intelligent-like behaviour as "instinct" has been done a lot with animals, erroneously more often then not. Even AlphaGoZero is no more then "curve fitting" but I would not dismiss it as a dumb "pattern matching" thing. And we may also say that human mind is nothing more than "electric signals" but this does not make us more or less dumb.
@pfever6 жыл бұрын
I just found your channel, pretty cool! Looking forward to see more videos of yours =)
@bernardfinucane20617 жыл бұрын
One reason why reinforcement learning works so well on games is that there is a more or less infinite amount of data to work with, and reinforcement learning needs that. But the technique will have to be improved to deal with real world problems.
@rooster4436 жыл бұрын
Same capabilities as our prior knowledge and processing speeds
@sedi43615 жыл бұрын
so what i conclude out of this. its crucial to combine computer vision (what do we see and the knowledge of what we see) with deep reinforcement learning to generate something close to primitive intelligence.
@alexanderkurz2409 Жыл бұрын
3:12 This reminds me of Chomsky's critique of AI and LLMs. Any comments?
@MrJorgeceja1236 жыл бұрын
Awesome video and explanation mate! Can you do one on inverse reinforcement learning or GANs in general? That will be great for the community! Thanks!
@garrett77546 жыл бұрын
Loved the video and explanation ! But I wished you offered a solution to teaching AIs priors. That might delve to far into the current problem of transfer learning, where there might not be an answer yet.
@ArxivInsights6 жыл бұрын
Exactly, we currently don't really have a good solution to the prior / transfer learning problem although many many people are working on this since it is a fundamental problem in all application areas of AI. I might do a follow up video in a year or so when there are exciting new developments!
@omarlopezrincon6 жыл бұрын
greeeat !!! you only need to normalize the volume of the whole video hehehe
@basti78486 жыл бұрын
LOL the Website says "You have already participated in our experiment before, hence you cannot take part in the experiment again. Thanks again." for the first time I visited it.
@ArxivInsights6 жыл бұрын
Bastian Schwickert I noticed it too :p The updated links are here: rach0012.github.io/humanRL_website/
@basti78486 жыл бұрын
All I had to do was clear the cookies to access the site, but still thanks for the updated links :)
@bjbodner30976 жыл бұрын
Thanks for another great video! Keep doing what you're doing:) Loved it!
@miladbanan24455 жыл бұрын
Thank you
@albertomartel65086 жыл бұрын
Hey awesome video, this channel is underrated (by no. Of subscriptions) Keep up the good work
@vanderkarl39272 жыл бұрын
Man that ending music was loud lol
@williamkyburz6 жыл бұрын
You forgot to put (pop up) some great universities in Holland. Like Ghent. Wish you the best in your Ph.D. studies. Looking forward to reading some of your research. Peace
@virgenalosveinte59156 ай бұрын
awesome vid
@snippletrap6 жыл бұрын
Killer channel, watched all your vids. But try to treat the acoustics of your recording space! Around 6:24 - :25 you can really hear the flutter echo.
@ArxivInsights6 жыл бұрын
Yeah, I know! I was using a very bad mic at that time, switched to a clip-on in the later videos so it's a lot better there :)
@Arsonade6 жыл бұрын
Your videos are awesome but please normalize your volume!
@ArxivInsights6 жыл бұрын
Adam Chess Haha I know, small mistakes were made in the first vids, the other ones have it normalized :)
@Arsonade6 жыл бұрын
Arxiv Insights, cool. Keep up the good videos. I think you've got a great format here
@rakhilsoman92996 жыл бұрын
So in an RTS game which has action based supervised feedback. Can I claim it will emerge with new tactics over time or do I need RL to claim the same.
@asddassl94536 жыл бұрын
For the future videos you might want to tune down the intro and outro music a a bit. It's much louder than your talking. Otherwise, a great video!
@ArxivInsights6 жыл бұрын
Many people made this comment :p I screwed up there xD! This is fixed in all later videos :)
@verdiergun6 жыл бұрын
Great great great video!
@muhammadhelmy55758 ай бұрын
4:00
@stephenkamenar6 жыл бұрын
0:24 1 minute? you mean 1 second?
@Fire64 жыл бұрын
Yeah xDD
@jackalstrategy96756 жыл бұрын
Nice video! Do one on google AI deepmind and openAI's race to conquer dota and starcraft2 please!
@itsjacobhere3 жыл бұрын
Is it just me or is the intro and closing music super loud?
@damnit2585 жыл бұрын
so is it good to let "fully" our prior understanding of the world influences our future decisions?
@juleswombat53093 жыл бұрын
Yes very rather interesting. Deep RL learning is much too slow learning and frustrating to apply in many practical uses and generalisation to real problems. So boot strapping with some priors, would be attractive to shorten training times, even if this is considered a return to feature management.
@JoshuaAugustusBacigalupi6 жыл бұрын
Ironically, the same analogy he used to explain our struggle in understanding quantum mechanics is very likely applicable to our struggle to understand the cognition itself.
@ArxivInsights6 жыл бұрын
Great point. I guess it would make sense if it turns out that a brain cannot by itself fully understand how it works... We might need some Machine Learning algorithms to show us :p
@MadScientist5126 жыл бұрын
I think he's got that first part backwards with the statement "Why are humans so good at this?" He says that the AI took 36 hours to solve a simple level, but that's actually the length it took the AI to learn from scratch how to play games, with no prior experience, and then solve the level with no knowledge of the objective; he even mentions that people have an advantage due to prior experience. This puts that figure in a completely different light, making one wonder how long it'd take for a kid who's never played games to learn the same thing, without assistance or knowledge of the objective. I know some older people who'd probably never get there :) Chess and Go AIs routinely learn the game from scratch just by playing themselves and they surpass top human level after only a couple of hours, they have to test them against other AIs because humans are too slow to catch up, which is kind of uncomfortable. It'd interesting to see Elon Musk apply some OpenAI to platformers, hopefully it's just a matter of time.
@ArxivInsights6 жыл бұрын
It's true that I should have placed the "36 hours" statement into richer perspective. The RL agent starts from scratch so you can't compare this with a human playing, very true! On the other hand, humans are able to transfer those prior forms of knowledge to new games the've never played before, and this is where AI currently fails quite miserably. No matter how many games you train the AI on, once you apply it to a new game, you basically have to start from scratch... And with AlphaGo, the "couple of hours" it took to train their latest system is what we call "wall-time", which is the total time from starting the python script to it having finished training. But you can throw as much compute in there as you have available, so you can't compute 'wall-time' with human learning time since there could be thousands of worker threads training simultaneously!
@MadScientist5126 жыл бұрын
Arxiv Insights That's even worse, you can't just discount 'wall-time' or turn the time it took to go through 4 million frames into 36 hours simply because it would take a human that long as that's how computers work, and you can't multiply the thread-count by the time taken or some such for a human-equivalent figure, humans aren't single core CPUs executing code. It's real-time human vs real-time AI, that's how it's been done since before Deep Blue defeated Garry Kasparov and the measure of super-human AI performance.
@ArxivInsights6 жыл бұрын
Well your argument is valid as long as you are talking about very scalable simulation environments where it's easy to spin up 1000 threads that do the same thing in parallel. As soon as you arrive at the real world (say eg training a robotic arm to pick up objects) real-world sample efficiency becomes very important. It's true that Google can use hundreds of robotic arms in parallel, but they are Google. For industrial applications of RL, the real-world sample efficiency is a rather important metric!
@WildAnimalChannel6 жыл бұрын
Good video. I learned something. Which is a rarity.
@TheOneMaddin6 жыл бұрын
Overall a good video, but I feel the note about QM was unnecessary. Physicist understand QM pretty well and its not strange when you have the right mind set. It is just presented as a very strange and unintuitive subject to the public, simply because this is how you get the attention for it. All these old quotes about "QM cannot be understood" are very questionable today.
@ArxivInsights6 жыл бұрын
Good point! Perhaps I should not have specifically mentioned Quantum Mechanics.. Nonetheless I still feel AI is great when it comes to finding new solutions in domains where humans have strong priors that are not necessarily optimal (like AlphaGo for example)..
@wunkewldewd6 жыл бұрын
I think it was a good analogy. If you notice, he didn't actually say that scientists don't understand QM, he just showed the Feynman quote saying that. But his point is totally correct -- even though we obviously can use QM effectively, it's still very "unnatural" to humans because it goes against a lot of our intuitions/priors/experiences/etc. So you might have a different definition of "understanding", but I think it's fair to say that between concepts like entanglement, tunneling, and a probability density rather than a definite position, it's much weirder than more classic fields like CM, EM, SM, etc. It's not that you can't understand it, but it definitely is weirder.
@CosmiaNebula6 жыл бұрын
The right mindset takes a lot of training to get, and it is strange not as a marketing stunt, but because that's how it is for humans. I mean, what kind of mind set would find both of these situations natural? 1. double slit experiment for electrons. Inference pattern appears. 2. double slit experiment for footballs. No inference pattern. And many more, such as the delayed choice experiments, which violate intuitive causality.
@palfers16 жыл бұрын
It's almost a century since the birth of QM. "New ways of thinking" indeed!
@TheAcujlGamer4 жыл бұрын
Great fucking video!
@loopuleasa6 жыл бұрын
cool channel
@triularity2 жыл бұрын
Just curious.. how much money was spent on this research to validate the obvious? =)
@MaxLohMusic6 жыл бұрын
Monkeys figure out object permanence faster than humans... More powerful models sometimes take longer to train... maybe those AI's aren't as stupid as we thought
@MarcoMugnatto6 ай бұрын
Not true. AlphaZero took 4 hours to learn how to play Chess better than any human, beating even IBM's Deep Blue
@pawncube20506 ай бұрын
You missed the point, obviously computer can train faster, but still need more training, case in point, alphazero played 44 million games. Its about amount of samples, not the time it takes to go over them
@codyheiner36366 жыл бұрын
Maybe the monkey experienced significantly different environment during its upbringing to learn object permanence more quickly? How well controlled was that experiment?
@judgeomega6 жыл бұрын
but an adult human has many years of training built in. when comparing AI to human performance, you need to compare a newborn human to the AI. things look a bit different in that frame
@ArxivInsights6 жыл бұрын
Exactly, hence the whole idea of prior knowledge. The difference is that our brains somehow succeed in rapidly transfering this inbuilt knowledge to new tasks and problems we encounter in the world. So far, neural nets can't quite do this...
@m07hcn623 жыл бұрын
I just got smarter after watching this.
@georgplaz6 жыл бұрын
In other words: human evolution and societies nurture were overfitting us..
@otonanoC5 жыл бұрын
How could you make a video like this and not mention MONTEZUMA'S REVENGE? No existing AI agent can play that game. And it's not just a matter of "human priors".
@boqsc03 жыл бұрын
The example with games is a nonsense, it's cultural preset that allows to perform well. Would take me hours or days to figure out what the game is actually about and what to do, if I had no previous preset to make a search and initiate the actions you expect and propose in that example
@SurferDudex99 Жыл бұрын
Lmao this must be a joke. Anyone who supports this theory has no understanding of the exponentially nature of how AI learns.
@versag37766 жыл бұрын
Good point, teach AI quantum physics with a reward for creating teleportation devices
@АлександрБагмутов6 жыл бұрын
Result was absolutely obvious and this is not f hindsight. Waste.
@JohnSmith-lf5xm7 жыл бұрын
Very nice video until mentioning quantum physics... WTF...
@ArxivInsights7 жыл бұрын
How do you mean? I think the analogy is very striking: natural selection has provided us with perceptual systems that are good at discerning predators and modeling the linear trajectories of a moving object etc.. Not so much to reason transparently about quantum entanglement or wave-particle duality. We are inherently limited by our in-built intuitions about how the world works. In a field very far removed from ordinary perception (as in quantum mechanics, where mere measurement affects a systems state) this becomes a significant hurdle if we want to objectively explore the laws of nature. I truly believe Machine Learning will play an increasingly important role in man's scientific endeavor because it is mostly unconstrained by our prior assumptions and evolutionary baggage.
@JohnSmith-lf5xm7 жыл бұрын
Your channel tells the development of a science that works at the scales of nature at which any Quantum BS is not applicable. Brains operate by electro-chemical interaction between cells. Cells get their metabolic processes going by the use of the ATP molecule. that's it!.. If you do not eat, your mitochondria does not make ATP, then cells can not perform its duty and die. Game over... sorry ... that is all.. is not that your souls stays playing video games in another dimension because of the wrap of the space-time. no! or that "go" is being played in all the possible outcomes in multiple parallel universes... no. none of that non sense. Those THEORIES of Quantum Entanglement and Wave-particle duality are just scientific misunderstandings... think on that like when the most advances minds though that the earth was flat, or that the earth was the center of the universe. The plain true is that still no ones knows how an atom looks like neither an electron or a photon... for some years all the science went down the path of quantum mechanics until it reach this point now at which it does not even make sense... hopefully other theories could come and save the day. look at this video and see how crazy the thing can go if you believe that those nonsense theories are possible. kzbin.info/www/bejne/e6WtfHWlo72mqpY And then look at for example this video which shows that other mathematical models can explain real phenomenon with out the need of bending the rationality. kzbin.info/www/bejne/jXrchY16fdqUqLM Your channel is amazing. I only watch two videos and they already help me with my work. Thanks!
@ArxivInsights7 жыл бұрын
Well I completely agree with you that the brain is just a physical system (although there are some very interesting ideas out there that disagree with this. Check out this video if you really wanna dive down the rabbit hole: kzbin.info/www/bejne/pZLHmHuemcyFoas ). And I also agree with you that nobody really knows what atoms, electrons or photons really look like, they are mathematical models that try to explain experimental measurements, nothing more. My only point is that we all have strong evolutionary biases that might sometimes cause us to miss good solutions/explanations and I think Machine Learning can provide a totally new approach to doing modern science!
@JohnSmith-lf5xm7 жыл бұрын
Thanks for the reply I just want to post this last video that kind of sets my view also of being careful with extrapolating subatomic particles mathematics to macro-scale reality (like ourselves). kzbin.info/www/bejne/bnWqmKl7a81_d7c
@alexfloyd57307 жыл бұрын
I think he was just making a point that humans use their common sense understanding of the world to help make decisions, and sometimes that common sense understanding is wrong. We can use our rationality to overcome our biases but that doesn't make it easier. A system without these built in biases may find unintuitive solutions to problems or it might even find it faster than a human would if those biases affect us too greatly.
@sausage4mash7 жыл бұрын
very interesting, thank you . I arrived here due to Alpher go zero beating the stockfish chess engine, I was blown away by its games, to me it showed signs of a deep understanding, playing moves whose benefit would be beyond a brute force calculation, but maybe I'm jumping to conclusions , guess I'm very curious.
@ArxivInsights7 жыл бұрын
Well, I think deep learning is radically shifting what we actually mean with "deep understanding". Chatbots (despite being incredibly 'stupid' and narrow) are pretty close to passing the Turing test (depending on the skill of the judges of course), a benchmark devised by one of the godfather's of AI almost 70 years ago. It's very fascinating to see the 'requirements' for true intelligence shift as Machine Learning progresses to display ever more fancy tricks and skills ;)