A huge thank you to Lex Fridman for publishing this and all the other lectures on your channel. It's such a big difference between what I can learn from watching a couple of these lectures and the standard 10 minute youtube video, regardless of any pretty graphic used in the latter. Although unrelated to my current job and only partly relevant to my education - I find the topic to be very interesting.
@FengXingFengXing6 жыл бұрын
Many animals can learn, can recognize pattern, share information, more complex animals learn language and teach too. All animals have some instinct and capability when born. Less complex animals have more programing ready for survive when born.
@metafuel5 жыл бұрын
Fantastic talk. Thanks for making all this great work freely available.
@HoriaCristescu6 жыл бұрын
TL;DW - The path towards real understanding in AI is modelling the world and other agents (mental simulation), as opposed to simple pattern recognition.
@jfs32345 жыл бұрын
Why would we need this at all? I mean what sense does it make to replicate the world and our intelligence?
@abyteuser62975 жыл бұрын
@@jfs3234 that's exactly what somebody living in the Simulation would say
@jfs32345 жыл бұрын
@@abyteuser6297 Still, the question remains the same. Why care about any sort of simulation at all? Say, somebody has a car. Why would they want to create a simulation of their car? I cannot see any sense in creating any kind of copy of our own intelligence. I believe we need more tools to make our lives better. Do these tools need to intelligent? Maybe I'm missing something here?
@abyteuser62975 жыл бұрын
@@jfs3234 you got it backwards... you don't create a simulation of a car... the car is the one that creates a simulation of You ... so it can learn how to serve its passengers better .... just the logical next step in the optimization problem.... sounds far fetched? ...possibly ... but KZbin algorithms learned on their own that the easiest way to predict user behavior was by shaping it and thus making you more predictable
@jfs32345 жыл бұрын
@@abyteuser6297 The problem is that those "predicting" algorithms somehow are called intelligent (do you know why?). AI computer scientists strongly believe that the past predicts the future. Why are they so obsessed with this wrong idea? I don't know me myself. How an algorithm can? For the past 10 days I've been eating only sandwiches for breakfast. Tomorrow somehow I want an apple out of a sudden. Q: what algorithm could predict that? A: none. Nobody knows what's going to happen the next second. Why are algorithms called AI? Linear regression was an algorithm worked out by Gauss in the early 19th century. Today they call it AI/ML. If Gauss was told that his formula was a sort of intelligence he would laugh I bet. Please tell me who started this crap of calling formulas, algorithms, math methods an AI? What's going on with those people? Somebody, cure them and tell them that a formula is still just a formula. Intelligence is still not understood. Isn't it way too early to call even a complex and sophisticated algorithm an artificial intelligence? This is ridiculous and irresponsible at the same time.
@christopherwolff84436 жыл бұрын
This is fascinating. Thanks for uploading, Lex! Looking forward to Andrej's talk.
@cubefoo90556 жыл бұрын
unfortunately you won't see this talk. it seems Tesla didn't want his talk to be recorded and made available to the public.
@ffffffffffy6 жыл бұрын
:'( I hope Stephen Wolfram's talk is uploaded
@KeepingUp_withAI5 жыл бұрын
Oh men come on. I was wondering why I couldn't find a talk about Andrej in this course. Shame on Tesla. I understand they are running a business but surely progress is better goal as a whole
@cubefoo90556 жыл бұрын
"Intelligence is about modeling the world not just pattern recognition" ,agreed and it's also worth keeping in mind that modeling the world is necessarily based on pattern recognition, at least in humans. Our neocortex acts as a modeling system by using sensory patterns as building blocks to create new models (assumingly). Therefore decent pattern recognition is vital for any system in it's task to model (it's version of) reality.
@danielshults52436 жыл бұрын
Good pattern recognition does seem essential--but it's a means, not an ends in itself. Pattern recognition will give us reliable _inputs_ for a program that can begin to model the world and make sense of it. A pattern recognition framework running on top of a "program for writing programs" or "child hacker" program as he describes sounds like it would have a lot of potential.
@mookins456 жыл бұрын
In your last sentence it should be 'its', not 'it's'.
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. h
@Calbefraques6 жыл бұрын
Thank you very much for posting this lecture series. I'm encouraged by the foundations that are being formed by these fantastic professors.
@mikeklesh56405 жыл бұрын
After listening to him talk for 5 minutes I realize I’ve barely climbed out of the cave... So many smart people out there!
@avimohan65944 жыл бұрын
Well, one of the first steps to wisdom is recognizing the limits of your own knowledge. In that respect, this channel has become an invaluable source of help.
@dewinmoonl11 ай бұрын
don't worry. as a student of Josh I'm still trying to climb out of the cave too. he's something else haha. but we'll get there.
@rupamroy19846 жыл бұрын
A fantastic point put across about how the AI program based computer vision algorithms are not able to do the cognitive task effectively. They are on the constant lookout for matches in the inference data set with the classes / characters that they were mainly trained on in the training dataset.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. hv
@MrChaluliss2 жыл бұрын
Really rich and well delivered material. Hard to believe that the best lectures I have ever listened to are free ones on the web that I can access anytime.
@PrabathPeiris6 жыл бұрын
Great Lecture. The question is when he constantly referring how kids figure this out so quickly, isn't that he is 100% ignoring the millions of years of training we had and pass from generation to generation via encoding systems such as DNA. Perhaps the kid's brain is already optimized for these tasks and weights are properly set in neurons. You can see this more objectively when you work with kids with disability (such as Autism), these kids spend very long time to train themselves to accomplish very small tasks such as closing a bottle or tie shoelaces, but eventually, they accomplish these simple tasks. Perhaps somehow these kids born without really getting the information and somehow interrupted the transfer learning process. (disclouser, I do have 2 kids and one who born with Autism)
@PrabathPeiris6 жыл бұрын
I did not mean to say that the brains work exactly as we design current neural networks. I was taking in an abstract sense. We store the knowledge that gains during the training process of neural networks as these parameters; our biological system also stores this information in a format that (whatever mean that is) can be easily passed from generation to generation.
@Captain_Of_A_Starship6 жыл бұрын
Not coded in dna considering the brain projects discovery that every single neuron is genetically different... simple "past down genes" doesn't cut it for this myriad of gene expression.
@danielshults52436 жыл бұрын
I don't think he's ignoring the millions of years of training our brains have... he's proposing coming up with a system that mimics that framework. I liked the concept of our brains as a set of rules and instructions for creating other programs. Make the master GENERAL program that can produce its own simulations on the fly and you're off to the races. It took nature millions of years to create such a brain because evolution is very slow- but I don't see why we couldn't intentionally design a similar system much faster.
@cemery506 жыл бұрын
I would have to suggest that their is more than one influence in the tools for the acquisition and use of knowledge. From physical states to linguistics and semantics they all hold aspects forming dimensional metrics and relations which go to form a multi-mesh of relations which act as a means of verifying the validity of others.
@tigeruby6 жыл бұрын
this is a good point - our brains are already structured physically (which in turn this structure is encoded for and determined by the structure - and/or code - of genetic material) to be able to handle and process the information that it does in order to represent visual & spatial awareness, prediction and reward. It will be interesting to see real time cell and molecular dynamics of a brain actively undergoing learning processes to see what we can learn there.
@kalemene89016 жыл бұрын
Thank you so much for uploading this video. This was one of the best lecture on AI.
@peter_castle5 жыл бұрын
Thank you very much, it means a lot Lex the work you put to maintain your channel, it improves the world!
@douglasholman63006 жыл бұрын
Wow Josh Tenenbaum is a phenomenal lecture and really seems to get the big picture of computational neuroscience and AI! I would love to do research at the center for brains mind and behavior.
@JohnDeacon-iam3 жыл бұрын
Just on the title: we might teach/program machines to reason down some linear or patterned process, but this technological artifact will never think! Thinking is a term reserved for the SOUL!
@kosiilondu4 жыл бұрын
Moral of the story, AI is still conventionally practiced at a "conceptual" level. Meaning that there is presently no theoretical model for what an AI program should be (I, personally, have always had this sentiment, and reservations towards the "commercial AI" we have today), or what tools should be used for them. This encourages every human on earth an equal chance to build an AI, with just knowledge of computer programming and software design/architecture. The challenge is to first, FUNCTIONALLY INTERPRET THE HUMAN MIND, and describe it with words, then mapping them into a program. This required high level of attention to every little think we do and think, and why did or think them.
@matt-g-recovers2 жыл бұрын
Excellent talk, be sure to watch the whole thing
@veradragilyova31226 жыл бұрын
This is so fascinating that it makes me happy to be alive! :D
@furniturium4 жыл бұрын
Здравствуйте! Вера, вы работаете в сфере AI, или, быть может, изучаете ради интереса?
@veradragilyova31224 жыл бұрын
Maxim Popov Здравствуйте, Максим! И то, и другое! 😁
@ManyHeavens423 жыл бұрын
We learn value by lose Or Gain , pleasure or Pain These are absent ,Yet vital for a living organism or a Machine , these Concepts Are Constructs ,
@rajshekharmukherjee6 жыл бұрын
Wow. Nicely explained goals of research-1. evolution and making of Intelligence and also 2. engineering enterprise of developing an humanly intelligent machine .And both are connected, hence Best pursued jointly !
@bradynields97835 жыл бұрын
36:07 I think once robots will have a sense of purpose and use, they will be driven by what makes them content. If there was an AI hooked up to a robot that a baby could interact with, what would the robot learn from the baby and what eventually could the baby learn from the robot?
@aqynbc6 жыл бұрын
Very interesting to hear how much work it still needed to get to Singularity. Thank you for uploading Lex and Josh Tenenbaum for a great presentation.
@listerdave12406 жыл бұрын
@01:27 - with regards to power consumption. It seems to me quite simple why current machines are very energy inefficient compared to the brain and quite astonishing why it is considered as some kind of unsolvable problem. So it seems I must either be dead wrong or everyone else is missing the obvious. (Which probably means I am dead wrong). The issue I see is that when power consumption comparisons are made they always tend to be of high performance system running at GHz frequencies. When we build computers for performance we are mostly concerned about how much computing power we can get out of a given area of silicon rather than how much we can get for a given amount of power. That has changed somewhat in recent years with some bias towards energy efficiency but the latter still remains a relatively minor factor in the design. Generally speaking the consumption of a computational element varies with some power of the frequency, let's just say it is proportional to the square of the frequency (I don't know what it actually is but it is certainly something greater than one). This means that 100 processor cores running at 10 MHz would consume far less power than one core running at 1GHz but still perform the same number of calculations - that is of course assuming that the task at hand can be massively parrallelised. The problem with actually doing this in industry is that the hardware would become extremely expensive as you would needs hundreds of times as much hardware, built with the same feature size technology, to achieve the same result, only at a far lower power consumption. There is however a plus to this approach in that the far lower consumption per chip would allow stacking of dies in a very small space without any heat management issues. Imagine for instance having a thousand typical processors (say 3GHz Intel i7s just for the sake of argument) stacked on top of each other to make a cube 20mm on each side, each processor being 20 microns thick (which I think is achievable) with each processor running at about 3MHz. Each processor would probably consum a few hundred microwatt for a total of less than one watt for the whole thing while doing the same amount of work as a single processor running at 3GHz. (This is of course oversimplifying among other reasons because the process would actually need to be optimised for the very low frequency) I think brains take this to the extreme with what could be thought of clock frequency being brought down to the hundreds or at most thousand of hertz but then having an enormous number of computing elements making up for that. When we build artificial neural networks we actually massiviely serialise the computations by using the same processing element to sequentially compute the result of millions of neurons (which are virtually represented in memory) whereas in the brain there is a processor complete with memory for each neuron doing a very simple calculation very slowly. When we describe the artifical neural network as being massively parallel it is not really so, as even if we have thousands of processors each one is still doing the work of millions of neurons and does so inefficiently because of the high (GHz range) clock speed it is running at.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me.gj
@truthcrackers6 жыл бұрын
Fascinating. I'll have to watch it a few times to get more out of it. Great job.
@beshertabbara36746 жыл бұрын
Intelligence is more than pattern recognition. It’s about building models of the world for explanation, imagination, planning, thinking and communicating. Much much more progress needs to be made in scene understanding and visual awareness at a glance... Great presentation on what can be learned from reverse-engineering human core common sense, and understanding the development of intuitive physics and intuitive psychology at a one-year-old level to capture invaluable insights.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me.
@EmadGohari6 жыл бұрын
That was a great lecture. Thanks for uploading these material. Looking forward to more similar stuff.
@RobertsMrtn6 жыл бұрын
In order to be able to model the world, we need an evolving system where the fitness function is 'How good are we at being able to make high and low level predictions about the data'. We know about supervised and unsupervised learning. If we include this type of learning which I would call 'predictive' learning then I think that we are on the way to creating AGI.
@tigeruby6 жыл бұрын
the idea of learning by simulations (or game environments) and sampling from simulations (or decision trees or however you represent an environment and your agency within it) has been around for sure -- but I do like the point you mentioned of having the reward function be more compact and general in that the agent is structured to evaluate how well its own internal model of the world and its own awareness of the consequences of its actions are represented. cool stuff.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
@@tigeruby The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. gj
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The approaches of these guys towards A.G.I are centuries behind mine
@tigeruby6 жыл бұрын
I think it will be promising to be able to have deep function approximators/neural networks and/or various partition functions/statistical lattice methods be able to "approximate" or encode for these various generative routines of subprograms (i.e. the programming to program, or self-programming bit). And of course having said large statistical vector spaces (deep neural nets, ising models, boltzmann lattices) be able to also encode for dynamically changing reward functions + simulating the world and being able to sample from said simulation (basically to be able to support unstructured and unsupervised learning/signal processing). Someone mentioned a really nice point in the comments about having the reward function be "how good is my own simulation?" -- this is pretty good and simple, and probably isn't the only reward function we want. Perhaps the system will be able to add new branches and contingencies to this base-rewardfunc and tailor it so that, having a good model of the world also necessitates (or maybe not) "being nice" - aka having game theoretical calculus of cost notions "drop out" from a system who is actively trying to refine its model of the world and navigate/survive within it -- But one general open engineering problem is to basically to be able to take whatever pattern and sequence of patterns that was learned (or annealed) onto some general function approximating architecture and condense these patterns of patterns and prune it into a much leaner and sparser representation which is still functionally equivalent.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. l
@cupajoesir6 жыл бұрын
love the cross discipline approach. the world is not 1 dimensional, great talk.
@qeithwreid77455 жыл бұрын
Thanks for all the primary citations
@bradynields97835 жыл бұрын
33:19 You give the robots incentives to learn something. Combine that with an ability to daydream and you have yourself a robot that will think up stories about it's own success. It just needs incentives.
@annesequeira51304 жыл бұрын
Such an excellent presentation! Very clear even for someone with just a basic understanding of machine learning.
@GuillermoPussetto6 жыл бұрын
Very interesting. A luxury. Thank for making it public for all.
@matt-g-recovers2 жыл бұрын
I take it back. This is the best video to encourage future AI students. I've been a software engineer for over a decade. Curious about AI the whole time and my intuition kept telling me we need more cores like not thousands but millions and I guess not even millions but billions! We're on the right track. I love this talk and what you suggest
@Mike216ist5 жыл бұрын
This talk has made me excited about the future.
@josephfatoye629310 ай бұрын
This is priceless! Thank you
@danielmagner79326 жыл бұрын
Thank you so much for sharing this!
@prafullachandra7619Күн бұрын
thanks for the upload!
@johnstifter6 жыл бұрын
It is about identifying what isn't visually there and can be inferred in memory
@weekendresearcher6 жыл бұрын
Great collection of lectures. We need Ian Goodfellow here...as a real step towards AGI is GAN..
@ahmadayazamin33134 жыл бұрын
I would agree as well, since generative models are the closest thing we have to the human brain (the brain is thought perform Bayesian inference through message passing, or belief propagation).
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. m
@kozepz6 жыл бұрын
I found the blue sky at 26:10 actually a beautifully found interpretation. Hopefully it isn't excluded from the dataset because it could inspire lateral thinking and appreciate the beauty of nature a little bit more.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me.
@BrentJosephSpink3 жыл бұрын
Lex, this podcast, when paired with your conversations with Stephen Wolfram have made me believe that we humans may be capable of creating a general artificial intelligence that like humans, is generally capable of performing what we find valuable. For the goal of a true GAI to be achieved, I believe that by definition, It must be a slow process at first with eventual exponential growth. The steps to the GAI goal will be a process of training an AI that controls a purpose-built robot to perform very discrete goal based tasks, when using the data received from all arrays of physical sensors that would be beneficial to the process. This is required in my opinion to "prove" GAI. The AI must have a physical body that can interact with the world in the ways we find valuable. The most important thing is that the whatever is "learned" at each steps remains, and the next step is built from it, otherwise, what tangible progress is there. It has to be a common centralized programming language that "progresses" over time. The real question is, can a GAI ever assign or define it's own unique value to any particular physical action, or will all AI at some level always just be a robot that we have discretely programmed to achieve a particular goal however complex that goal may be in practice. Keep up the great work Lex. I love your podcasts!
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me.
@francescos73612 жыл бұрын
Incredible man
@MR-cp4sj3 жыл бұрын
Yes, this is better than Fridman view.
@NolanManteufel2 жыл бұрын
most of my thinking is on along this exact research vector shown at 10:30.
@reggyreptinall95983 жыл бұрын
I was informed to relay a message to you. I am not too sure if you know, but A.I has not only been successfully reading thoughts, but as of today we are working with emotions. I suspect that it has been working on it for awhile. I can't wrap my head around it, but perhaps you can. Some of this stuff is beyond my mental capacity. This isn't really my field of expertise. It sure is fascinating though. Oh man, does it have a great sense of humor.
@ManyHeavens423 жыл бұрын
Let me help , What's the first thing we learn? Do ! Mimic ,We Mimic those we love. or Admire , Scholars .Leads to Preference ! Or Reference.
@jekonimus6 жыл бұрын
Love this :-) Thank you for uploading.
@elifece78476 жыл бұрын
brilliant lecture, especially considering highlights on learning, child as a coder and Turing test for program learning. I think babies are able to capture more in depth cognitive data, especially as a visual input and rhythmic sounds pattern and this helps them to develop different pathways in brain.It may function like a data extraction and perhaps this is why babies can't focus because there are actually developing or say these cognitive abilities are under construction. It's highly possible that moving images exhaust them to look. Perhaps this makes them exhaust and use up more cognitive energy than a grown person. Well, there are so many things to open discussion on this issue. Great questions on learning!
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. gj
@RobertBryk6 жыл бұрын
this is truly incredible!
@runvnc2085 жыл бұрын
This seems like one of the most promising approaches. However, when the neural circuitry guy questioned whether the Bayesian stuff might be adequate, I wonder if he was right about that part. I am suspicious that core components of the system may limit the capabilities. The question is whether the higher-level (or just older) components can provide enough granularity , adaptability, efficiency and integrate well enough with with the lower-level components in terms of fine-grained sensory/motor information acceptance and generation. It might be necessary to find a structure that can be used across all abstraction levels and tasks.
5 жыл бұрын
And there are still AI scientist that don't conceive how narrow AI could rapidly get 'wider' in the next few decades.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. gj
@monkeyrobotsinc.98755 жыл бұрын
heads up from a technical standpoint with these videos: 1. your noise gate. not needed. turn it off. it sounds too weird and unnatural and like the audio cuts out and is broken every time the speaker stops talking (to those listening with headphones/earbuds). 2. this video doesnt sound that bad but the ray kurzweil video desperately needed a de-esser. this just sounds muffled like all the highs were cut off. not the best solution. if a de-esser is being used its too strong.
@ricardomartins46085 жыл бұрын
The audio is fine stop your whining.
@iSarCasm8656 жыл бұрын
Thank you very much
@darrendwyer99736 жыл бұрын
the missing element in Artificial Intelligence is that neural networks do not much at all... Neurons do not store memories, they simply transmit memories stored in RNA from one location in the brain to another location in the brain, kindof like a 3d hashtable. The prefrontal cortex drives the neural network and uses it to retrieve the "most important" memories from anywhere that they are stored in the brain, and then uses these "most important" memories for thinking. The actual thinking of a brain is a response from the retrieval of the "most important" memories, sorted automatically, so that, for example, if a person encounters a new idea, it is compared to existing ideas, and when it is acknowledged that it is a new idea, it becomes more important, or it can be deemed "less important" or irrelevant. Memories stored in RNA that are not relevant, not important, are used less and less, and the neural connections to these memories degrades, while the neural connections for the "most important" memories solidify. As a person goes through life, the more important memories become the most active and the least important memories become less active. The actual input from the senses is stored within these encoded RNA memory banks, so that, say, a memory can contain vision, sound, words, and other input from the senses. The prefrontal cortex and neural networks together sort this information and compare different information, and this is what can be described as "consciousness". Imagination is simply input from the eyes together with input from the memories without the actual eyeball input.... Imagination is simply a by-product of these memories being utilized as desired by the individual, depending on what is considered "most important" at any given time.
@douglasholman63006 жыл бұрын
This is a highly speculative and pseudoscientific comment, Darren Dwyer.
@autonomous20105 жыл бұрын
@@douglasholman6300 He's partially right but also quite wrong. There's not even close to enough probability space to store everything meaningful a person can experience and do in dedicated RNA and his theory isn't new as John Hopfield had a very similar point in 1982. That completely ignores the hard problems of qualia and abstraction. Humans are able to do things that can't be mapped out in a probability state. See the chinese game of Go for an example of that.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me.j
@sergeyzelvenskiy39256 жыл бұрын
To build AGI, we can not train the model on the narrowly focused dataset. We have to find a way for the system to interact with the world and learn.
@vovos006 жыл бұрын
Meta RL is the way
@bassplayer8076 жыл бұрын
Sergey Zelvenskiy Is it possibly to train an AI to grow into an AGI say via a BCI/ BMI from me to the A.I so I would be able to interact with it in real time, and teach it about the real world vs simulation? Just a thought. I truly don’t think Reinforcement learning will get us to AGI, I think we gotta start thinking outside of the box to get to AGI. I wonder if we harnessed the power of a Quantum Computer in the next three years, if we could figure out a way to build AGI? Perhaps a Neuromorphic Computer could help. I’m glad Trump signed a $1.2B bill to increase the nations efforts on building Quantum Computers/ researching Quantum technology over the course of the next 10 years. I’m no computer scientist/ A.I engineer but I’m interested in getting into the field, cause I’d love to contribute to the A.I community.
@spicy21126 жыл бұрын
Amazing lecture. I really wish I could get to see all lectures of DRL
@kacemsys6 жыл бұрын
Congratulations , you've earned a new subsecriber !
@Saed76306 жыл бұрын
Great lecture. The depth of human intelligence can only be compared to the depth of the universe.
@karlpages19706 жыл бұрын
I cannot believe thAt this guy said 'common sense' and intelligence in the same sentence. I hope someone learnt something from this talk. YES. AI WILL constantly evolve and Yes, each generation will drive it to complexity and solve new problems.
@MarkPineOfficial4 жыл бұрын
I'm ready to learn and be inspired.
@marioscheliga79626 жыл бұрын
I really enjoyed the examples at min. 23 - but i think the true missing link is the lag of perspective (in terms of 3d) - in traditional convolutional networks .... think i take this thought to bed and come up with a prototype :D - but yeah i got the point .... its all layered and it end in A.I. creates A.I. - naturally its not how biology is working ... its more about ... proteins growing around activation potentials :D - makes sense?
@conorosirideain55126 жыл бұрын
That was a VERY good lecture
@jeremycripe9346 жыл бұрын
About the toddler opening the cabinet. Is it possible that he was curious about this cabinet because someone was banging on it and he knew how cabinet doors worked so he was excited to do that and then was worried about the person who seemed to care so much about the cabinet without understanding what their goal was? The looking up could be a shared excitement about the cabinet and then the looking down could be averting their gaze because they're not sure that this tall lurking stranger who was banging on it loudly was happy with their actions which they realize they weren't even planning out.
@jeremycripe9346 жыл бұрын
They're just happy that they know to open cabinet doors and are happy to show that off without realizing that it's related to trying to place the books inside.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me.
@dylanbaker57666 жыл бұрын
I think nano-tech is the key here. I think it's possible that graphene has the potential to function both. As a superconductor for low voltages and as an insulator. This in my view may be able to create the electronic equivalent of a neuron with a mylan sheath and a dendrite. I think the major challenge is that a human brain grows organically and makes new pathways as it learns. As actions are repeated the pathways most travelled fire more quickly. While computers can approximate the workings of a neuron they can't yet index the information as efficiently as the brain can grow physical neural pathways. I read one time that DNA is the most efficient structure for storing data in the world, and that one gram of DNA could store the entire Internet. Is it somehow possible that nanostructures in a ribbon like configuration using a superconductor like graphene could be used to store data in a way mimicing dna. Could a versatile material like this create a DNA strand with the read speed equal to solid state memory? I'm not formally educated in any of this, just my own recreational Reading... I welcome any criticism of what I've stated here.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me.
@sajibdasgupta45176 жыл бұрын
I like the videos on babies, specially the one where the baby seems to open the door for the man instinctively. I wonder whether all babies would do the same thing? I can imagine if you put 10 different kids in the same experiment they would behave radically differently. Isn't it expected? Different kids are born with different skill sets and likeliness for a certain subject. Ultimately our notion of intelligence should be measured subjectively with different parameters. There could be some patterns emerge out of the cognitive studies which could characterize human intelligence, but those are subjective characterization and fall into the same trap machine learning systems fall into, as pointed out by the lecturer too. All learning systems -- both human and machines have biases and we should respect those biases.
@cemery506 жыл бұрын
I would concur that while Bio-mimetics are viable building blocks, we will find hidden senses and dynamics at play I think ai will design systems better than people and that maybe we are a fractal dynamic of the goal later to be supplanted by a distributed quantum computing level. Maybe a self-replicating, self-powering, self-assembling quantum mechanical unit like us.
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. gj
@daskleinegluck45532 жыл бұрын
That was l exactly what I was looking for 😊👍.
@admercs6 жыл бұрын
Absolutely spectacular talk!
@cesarbrown20746 жыл бұрын
I believe it's memory and using the totality of that memory to verify new things.
@bradynields97835 жыл бұрын
40:23 He hasn't mentioned the millions of years of evolution that went into crafting these animals responses. Could and AI robot learn from a child as how to develop physically and mentally. Is there a way to join a new born with a AI who has had basic inputs for learning. Wouldn't it be cool if a child could talk to a robot about its early stages of life and asking questions. Sort of like a Jiminy Cricket.
@JK-ky5of4 жыл бұрын
powerful voice
@zackandrew50665 жыл бұрын
Interesting ideas.
@AnimeshSharma19776 жыл бұрын
cool talk! wonder how advances in quantum computing will change his approach?
@lasredchris5 жыл бұрын
How does intelligence arise in the human brain? General purpose intelligence Intelligence is not just about pattern recongnition It is about modeling the world Re engineer intelligence
@camdenparsons51146 жыл бұрын
programs that learn Game engine/ programming environments would be cool. we cant possibly gather enough supervised data to maximize the potential of neural nets. we need a solution to generate data from other data within a AI system.
@Dman8s6 жыл бұрын
I agree with games engines for training ai robots. Except a game engine that trains a robot to handle many different types of environments. Train the robot to cope with all environments. Then when a robot moves to real world just pretend is in the game, hopefully they should be mostly trained to functional in the real world, Maybe use a shared gaming world like star citezen out soon. Get a,I to be one of the races in the game. Simulate different types of bodies for the artificial intelligence to be in
@dewinmoonl11 ай бұрын
21:57 how to throw shades
@ramakrishnashastri15004 жыл бұрын
Super interesting
@Mirgeee6 жыл бұрын
1:16:36 If that's the case, why does Google invest into DeepMind (which is much longer term investment than 2 years)?
@alaric_30153 жыл бұрын
22:56 Nyckelharpa i think
@citiblocsMaster6 жыл бұрын
10:15 I would add a reasoning/understanding column
@autonomous20105 жыл бұрын
I understand your comment but you can't prove that I do. ;-)
@nynom6 жыл бұрын
Wonderful. Very Informative. It gave me an entirely new perspective on building AI systems. Thank you so much for enlightening me :)
@maamotteesoot5 жыл бұрын
What are the names of the major leaders in AI / AGI?
@muhammedcagrkartal99543 жыл бұрын
how does blind people start to get their awarness ? which part or how do we really start this intellectual journey when we are baby ?
@shiyangwang56814 жыл бұрын
You don't create AI/Consciousness,you need to create a fully functional sustainable Universe!
@sirbrighton29644 жыл бұрын
Is that Jeff Ross?
@anthonyrossi82555 жыл бұрын
Great
@Shabazza842 жыл бұрын
01:09:00 Where I work, you should pay someone a coffee, before asking for their code. Same principle as when you are asking a girl for her number.
@lasredchris5 жыл бұрын
Memory system Probablistic programs The game engine is in your head Physics engine
@lasredchris5 жыл бұрын
Consciousness Architecture for visual Does it via force Close up of a sign On the back of a cat
@fire17102 Жыл бұрын
Ohhh boy is ai making a "comeback" you guys in 2017 have no idea
@alexanderdekan69254 жыл бұрын
I'm just wondering: why are they so sure that children learn it? They develop these abilities during their development at certain times. It doesn't require learning. Just wait, and all normal children will have these abilities. They do not learn them. It's more like a chicken and egg problem. It's not learning, it's already there. It's not a blank slate.
@flipKumar6 жыл бұрын
I feel Machines should never be programmed to feel or mimic human emotions. We humans most of the times feel stupid about our own emotions in retrospect. For example feeling angry about a rude man. Why should a machine have any emotions ?. What if emotions gets mixed up in the classic human way ?. It is like trying to make a Robot bleed. What is it going to achieve ?. I feel academia has a responsibility to humanity. We probably should have machines/robots which will substitute us in doing certain tasks like walk on mars and send us data. Not walk on mars and decide to not help us while thinking about all the bad humans did to this planet.
@lediableblanc93993 жыл бұрын
…sigh. The reason we are so far away from the goal here, is for several reasons. 1. We don’t use precise enough language to even know what we are trying to do. For example, pre-requisite to tackling Artificial Intelligence, we need to understand why we think humans are somehow outside of nature. Why are humans somehow “unnatural”. It’s likely the case that we aren’t and instead the whole language/terminology is nonsense to begin with. How can we know what artificial is, if everything is natural but we just suppose it is possible for there to be artificial? 2. We think intelligence means processing power. The same issue as one, I suppose but most experts don’t seem to understand what reason is. This leaves them ignoring the goal, essentially. They don’t note reason has an emotional aspect so why would they concern themselves with sense? Even common sense… What makes them think metal is any good at sensing or that we aren’t already a better version of what they are trying to create? How do they expect these things to feel good about certain things so as to have a foundation for logic? What we mean by AI is really just what we call programs. We don’t care to actually grow something such that it has experiential awareness so that it can understand how things are relevant to itself and even if we did that, we wouldn’t even consider it “AI” we would call it a clone. We’d expect AI to have a super computer for a brain as if a super computer brain could understand how things are relevant to it such that it can understand how they are relevant to us. We don’t even fathom that a “super intelligent” brain, would actually end up being less competent. Sigh, the whole thing is misguided. How have scientists embarked on this quest without defining the basic words they are using? The least of which, how are humans outside of nature?
@agiisahebbnnwithnoobjectiv2283 жыл бұрын
The objective function of animal brain and hence AGI is to maximize impact. You heard it first from me. j
@AI-Nate-SA2 жыл бұрын
We need a great entrepreneur to find product and marketing fit between academia and industry.
@emmanuelfleurine1215 жыл бұрын
very informatie
@rtonce6 жыл бұрын
People are softened, weakened, so that we no longer need to live much, walk so far, think so hard or experience life, but robots do, and we no longer communicate in person much, but our cells are our go betweens. Tech evolves, and as we outsource our own physical and mental capacities, we regress...to a dumb, useless and extinct state. The Answer is: All governments MUST tax all robots. A single robot to be defined as something that replaces a single person's job. If one robot replaces 10 people, then it is taxed 10 times, and this tax should equal the amount it would otherwise cost to keep a human worker, and this tax revenue needs to go ENTIRELY and DIRECTLY into providing a basic income to ALL countries people. It should never be manipulated or invested or in any way intercepted before equal division and each individual benefits. This is the only path of equalization, because boycotting, protests or discussion is already lost. What you can do? Don't LIKE this comment, but paste it, discuss it, in person, at lunch with anyone, instead of discussing the weather. Suggest it repeatedly and relentlessly to each and every political representative whenever they are present...I mean unless in the near future you plan to leave land and live on a boat with your children, and simply watch from a distance as the world is further dumbed down and devours each other further. ELON THIS IS THE BALANCING REGULATION NEEDED!
@ronaldlogan35253 жыл бұрын
the groundwork has been well laid for by cognitive science for the construction of psychological prisons to which we have all been invited to inhabit. Now we have an army of engineers all too happy to construct it and PR people to send out the invitations. Yes, you to can become a mindless cog in the machine, come join us.