SPONSORED BY TUFA AI LABS (home of MindsAI)! Open research positions to work on ARC - tufalabs.ai/open_positions.html
@stephenwallace878217 күн бұрын
This channel has shown me the whole world of AI in a serious fashion - the intersection of logic, reasoning, philosophy, and science....makes the hype-side seem a little silly compared to the incredibly rich content this channel puts out.
@stephenwallace878217 күн бұрын
I;m not sure if -- in the sense of human beings -- we might have a kind of rich competence, with only performance enhanced or degraded depending on certain elements...But I really like Francoi Chollet's understanding of intelligence a lot and think he is really carving the right way forward for the art - realistic, but considerate of the promise made by all of the influx of AI.
@MrgoldenRose8 күн бұрын
Totally agree
@ngbrother18 күн бұрын
Something that I very often think of when reflecting on this definition of intelligence is how many forms of economically valuable work don’t require dealing with novelty - just pattern recognition and following standard processes. The type of intelligence that Arc asks us to strive for isn’t necessary for AI systems to displace a significant portion of human labor. It’s only needed if we want AI to replace all economically valuable work.
@VoloBonja18 күн бұрын
It’s needed to call it AGI
@TerrelleStephens18 күн бұрын
You're completely missing the point. The point is that those tasks don't require intelligence at all. He's not arguing against AI systems being useful; he's saying don't be fooled into thinking it's more capable than it is simply because it does so well on tasks that don't require it to use actual intelligence (as defined by chollet). The type of intelligence he is advocating for is necessary for true conversation and decision making. Not everyone is interested in AI simply to automate some process. There is a lot of room for AI systems to act as advisors who don't share the same memory deficits as humans and can simultaneously consider many more courses of action than a human without losing the context and goal of the one they're advising. This requires Chollet's version of intelligence.
@w花b17 күн бұрын
AI companies have been promising AGI (whatever that means) but we can at least agree that an AGI could easily solve Arc. What you're doing isn't just shifting the goal post, you're trying to convince yourself that it's not even there.
@szebike16 күн бұрын
If you look deep enough you definitely need more than mere pattern recognition for most tasks. Only a small subset of work really only depends on patterns and nothing else. So for the forseeable future we will have systems as assistants not as replacements. [Some tasks will be replaced like writing standart answer emails but its rather than having a photo isntead of a painting style innovation rather than creating an artificial human we are still far away from that .]
@MarkEngelstad14 күн бұрын
Almost anytime somebody says a job doesn't require intelligence, you are talking to somebody who has never done the job.
@FredPauling15 күн бұрын
Francois has such a clear thought process
@CodexPermutatio18 күн бұрын
True intelligence involves planning, learning, and modelling "new concepts" about the world and ideas in general. For this, pattern recognition is a necessary (but obviously not sufficient) requirement. Amazing content as always. Glad to see Chollet back on the show!
@EobardUchihaThawne18 күн бұрын
i think searching is a part of intelligence too
@ArtOfTheProblem17 күн бұрын
@@EobardUchihaThawne search encompasses planning. you need world model + search algorithm
@axe86315 күн бұрын
@@EobardUchihaThawne Searching can be a part of intelligence if its done in a restrictive manner.
@yohanj523911 күн бұрын
Yes I agree. But are those pattens causally connected ? Current AI systems are mostly based on correlations requiring absurd amount of energy to reconstruct the context.
@MrgoldenRose8 күн бұрын
@@EobardUchihaThawneeffective search definitely requires it but I don’t think that makes them the same thing
@Komaruluten18 күн бұрын
Yesss finally someone realistically depicting the current state of AI. AI's are currently really good at pattern recognition Pattern recognition is of our brains intellectual functions yeah but it's not the only function involved in making humans intelligent Relational reasoning, spatial manipulation, different kinds of memory (working memory, short term, long term), executive functions - these are all interconnected intellectual functions of the human brain, and LLMs for instance are currently only capable of a subset of relational reasoning (which is pattern recognition) and also have memory - they simply are not at our level yet
@w花b17 күн бұрын
They're just very sophisticated search engines
@Emerson118 күн бұрын
Great Interview, Francois has a lot of great novel ideas and can clearly express them, I see why you're a fan!
@julian78W8 күн бұрын
Thanks for this great interview. I'd love to see a debate against Eliezer because I have yet to hear a good rebuttal to his arguments. Chollet does seem to strawman the AI safety position in the ending segment.
@DiogoVKersting18 күн бұрын
This interview was awesome. Thank you
@antonystringfellow515218 күн бұрын
"Intelligence vs Skill" Very well explained! This is where I believe Demis Hassabis got it wrong when he said that you can have intelligence without consciousness. I don't think you can. The smartest LLMs are like the subconscious part of our mind that can learn elaborate skills but that understand nothing. The conscious part of the brain, which is slow and serial (can only focus on one thing at a time) delegates most of the work to these programmed areas, only providing guidance where necessary. When we delegate too much and don't provide sufficient guidance, so we can focus on something else, we often end up executing unwanted actions, like taking a wrong turn in the car, walking into the wrong room or throwing food in the bin then putting the packaging in the fridge. We have to delegate but then guide and monitor occasionally because although our subconscious has the ability to execute actions, it has no idea why it's doing anything. It understands nothing. To understand requires the awareness we know as consciousness.
@blijebij10 күн бұрын
How about a bit less black&white perspective and having degrees of intelligence, where it also can combi (fall in line) with other traits from consciousness. Cause it could be at highest degree no consciousness trait can stand 100% isolated alone on it self but always lines up with other qualities from consciousness.
@morphos217 күн бұрын
François Chollet is an endlessly deep vault of interesting ideas. What a fantastic conversation!
@BeTheFeatureNotTheBug18 күн бұрын
Thank you so much for this video. I can’t say enough about the value of this interview on balance with a 1000 otherS.
@BeTheFeatureNotTheBug18 күн бұрын
The thumbnail got me !!! Look out Mr Beast. Seriously why I clicked even though a subscriber. Best ever.
@4222216 күн бұрын
“The intuitive mind is a sacred gift and the rational mind is a faithful servant. We have created a society that honors the servant and has forgotten the gift.” Albert Einstein The key to AI is the alchemy between Intuition an Rationality.
@ihysc43708 күн бұрын
Of course!
@loopuleasa6 күн бұрын
francoi's observations on how his own children learn is the best avenue for understanding learning in general, and why current AI systems don't learn as well currently at 24:30
@rolodexter18 күн бұрын
Revelations from most to least severe, focusing on implications for AI development and our understanding of intelligence: 1. Most Severe: Current AI Performance Metrics Are Fundamentally Flawed Timestamp: 00:03:45-00:04:05 Quote: "Performance is measured via exam style benchmarks which are effectively memorization games" Why Panic-Inducing: This suggests we've been fooling ourselves about AI progress - our primary metrics for "intelligence" are actually just measuring memorization capacity. Years of perceived progress might be illusory. 2. The Scale is All You Need Hypothesis is Wrong Timestamp: 00:02:34-00:03:00 Quote: "Many people are extrapolating... that there's no limit to how much performance we can get out of these models all we need is to scale up the compute" Why Concerning: The dominant strategy in AI (just make bigger models) may be fundamentally misguided. This challenges the foundation of many major AI companies' strategies. 3. LLMs Cannot Do True Reasoning Timestamp: 16:54-16:56 Quote: "Neural networks consistently take pattern recognition shortcuts rather than learning true reasoning" Why Alarming: Suggests current AI systems, no matter how impressive they seem, are fundamentally incapable of real reasoning - they're just very sophisticated pattern matchers. 4. We're Missing Half of Intelligence Timestamp: 00:12:17-00:12:28 Quote: "Intelligence is a cognitive mechanism that you use to adapt to novelty to make sense of situations you've never seen before" Why Troubling: Current AI systems lack this fundamental capability, suggesting we're much further from AGI than many believe. 5. The Deep Learning Limitation Timestamp: ~16:39-16:54 Quote: "I realized that actually they were fundamentally limited, that they were a recognition engine" Why Significant: Suggests deep learning itself may be a dead end for achieving true AI, despite being the dominant paradigm. This transcript is particularly shocking because it systematically dismantles many of the core assumptions driving current AI development and suggests we might be on the wrong path entirely. Chollet's insights, backed by his extensive experience and concrete examples like the theorem-proving work, suggest that the current AI boom might be building on fundamentally limited foundations. The most panic-inducing aspect is that these aren't speculative concerns - they're observations from someone who has been deeply involved in the field and has seen these limitations firsthand through practical experimentation. It suggests we might need to fundamentally rethink our approach to AI development.
@jb_kc__17 күн бұрын
the way the interviewer was smiling the whole conversation.... me and you both mate
@Morimea18 күн бұрын
1:51:34 - very good analogy for "intelegent task and agents" 1:56:51 - also very nice point about learning
@HanzDavid9615 күн бұрын
The LLM can be very helpfull for exploring the solutionspace to find the new patterns! We always think based on our expirience but we can use that expirience to find new solutions to problems. And than that new solution becomes part of our expirience. Thats why you need to use the LLM within a multi agentic system that is able to reflect and support multiple modalities.
@adamkadmon633917 күн бұрын
Terrific. Such a thoughtful person.
@opusdei115113 күн бұрын
Wow this episode is so rich of thoughts
@stevo-dx5rr18 күн бұрын
Great intro, and great talk so far!
@dcreelman18 күн бұрын
Fluid intelligence vs knowledge re professors entrenched in their beliefs (48:30) "It depends whether you…believe you already have the answer to the question or you believe you have templates that you can use to get the answer."
@RealStonedApe14 күн бұрын
I'd never heard the Kaleidoscope metaphor - that's really beautiful. I love that. That being said, I can't help but feel that Chollet falls into the Yan LeCun camp of being on the opposite extreme of the AI Hype. I feel that he is underestimating just how capable LLMs are and how much more is going on with them under the hood that we don't yet understand. But he's correct - there is something missing here and it's not just scale.
@MachineLearningStreetTalk14 күн бұрын
I don't he's underestimating LLMs broadly, he's just arguing that they need to be augmented with a search/reasoning process and perhaps test time inference - and this is what the frontier methods do now! This is far from "DL is hitting a wall"
@aa.88237 күн бұрын
@@MachineLearningStreetTalk I don't know about DL in general, but just scaling up doesn't seem to work anymore. At least, to produce something qualitatively different. Why would it? Like 4o is just a stochastic version of Mathematica. It's worse, because it introduces errors, but it's simultaneously better, because it may give something new. Is it really productive to broaden use cases one by one? I doubt. The current LLMs are quite rich already. What do you mean by “search/reasoning process”? CoT or something like it? It won't work, if you verify your results using the same LLM. Because it just fundamentally lacks the resolution. It may broaden the search and make the result less dependent on the query. But it's still mostly recall. While watching the video, I've understood what I don't quite like about ARC. It kind of forces you to deal with some narrow set of problems. It suggests some concrete solutions. Let's take these functions, recombine, tweak this and that. However, we know from history that this approach is complex and always fails. Yes, this time it's not just hard-coding everything, but it nudges you to do it. When I had thought about LLMs and different levels of intelligence (or lack thereof in case of LLMs), I recalled the following quote by Banach: “A mathematician is a person who can find analogies between theorems; a better mathematician is one who can see analogies between proofs, and the best mathematician can notice analogies between theories. One can imagine that the ultimate mathematician is one who can see analogies between analogies.” I want to argue that current models capture enough richness, but they are inefficient and contain a lot of redundancy. Because they lack “understanding”. I think a better approach might be not to broaden and scale the models up, but to use the existing models as a starting point for feature mining. You need some higher-order training. One way to formalize it is to say that there are structurally similar parts inside, which can be smashed together. Alternatively, we can talk about invariance under group action or aforementioned analogies. So, there must be parts of a model that can be approximated with a sum of affine transformations of some basic set of elements. Otherwise, they would understand more abstract connections, which nobody forces them to do (at least don't know anything about it). Of course, there are technics for effective decomposition/compression. Or maybe a better approach is to seed a world with an agent and some model and make them (co)evolve. Something like GANs. So, there are also established technics here. I'm sure, there are people, who, unlike me, actually know something about DL and neural networks, and thought about all of this. And maybe tried it. It's just something, I've never heard about. Well, yet. All of my other ideas I have already heard from somebody. Or seen implemented.
@NoName-lq7kt18 күн бұрын
My ego took a hit from the title and I clicked
@simonstrandgaard550318 күн бұрын
Great interview. Well produced.
@AbuChanChannel17 күн бұрын
One of my favorite episodes ...thanks
@justinduveen381517 күн бұрын
Thank you for explaining your thorough research and advanced thinking so clearly! I agree our brains and the path to AGI is made up of multiple agents and sub-agents working together, each with different expert specialisations, reward maximisation and loss minimisation functions built in. By using prompt engineering to create the appropriate expert agent (with many years of experience in that particular field and with the appropriate value system, thinking skills and output format), then chaining many agents together to work both hierarchically and sequentially, these collaborations unlock improved cognitive capabilities.
@wwkk496418 күн бұрын
🎉Great interview!
@faster-than-light-memes18 күн бұрын
That was a small book length podcast. Epic.
@dr.mikeybee18 күн бұрын
The space of vector functions is functionally complete. That means that in composed pipelines of vector functions some can be logical functions like AND and OR.
@Morimea18 күн бұрын
>AND and OR Boolean algebra.
@LukeMlsna2 күн бұрын
Please elaborate.
@Morimea2 күн бұрын
@@LukeMlsna >in composed pipelines of vector functions some can be logical functions like AND and OR you can "express"(generalize) any mathematic-logic as combination of AND/OR -> Boolean algebra cases of "optimizing formula by combining some parts with AND/OR" - is just logical outcome of Boolean algebra - if you express everything as AND/OR - you will obtimize (logic will collapse/downscale to few patterns that just repeat) and in context of AI-ML - it just patern finding and compression of patterns - and "pattern" - can be anything in context of AI-ML trying to do "precise optimization of fields/gradients" - is same as writing default programmin programming language - where you compress some precise behaviot as combination of AND/OR optimizations for ML - is also patterns for cases - so AI can be trained to "find perfect optimization for this exact case" - no reason to "search for individual precise optimization for this exact case" same as computer-processot solve anything by AND/OR - AI will/should find "best optimizations" for every case all human-made math - outdated
@LukeMlsna2 күн бұрын
@ Truuuue. Thank you for that.
@janerikbellingrath82017 күн бұрын
amazing show, as always!
@wwkk496418 күн бұрын
55:43 its suggested that ARC test is 100% soluble based on non-overlap (disjoint set) of two test takers' solutions that Francois evaluated to be incorrect. This conclusion is faulty and 3 observers (francois and the 2 test takers ) cannot use their mutual disagreement to prove 100% solubility, rather the reverse, that at least a portion of the test is undecidable until a fourth observer can find perfect agreement with one of the three previous observers (francois and the two test takers).
@privacytest912618 күн бұрын
This was amazing, thank you!
@Jeremy-Ai18 күн бұрын
Thank you This is beginning to be better understood. The risk appears to be held by the individual interpretation of intelligence, subject to pattern and anomaly distortions. It is likely best practice to assume that the agent is unaware of its intelligence, and must be analyzed by measuring patterns and anomalies within its boundary. This is a challenge confronted over and over as each agent is of specific design and is responsible to assess one and another and so on. The compute data is necessary for scale, however it is not the problem confined to proper function resolution. Meaning, what good is endless scale, pattern, intelligence, pattern, and compute in an ever outpaced faulty operating system? This good for idiots to remain believing that they are not. That is all its good for and so fault function/tyranny remains.
@hannes721817 күн бұрын
Great interview! Would love to see Albert Gu on MLST at some point in the future
@13NHKari16 күн бұрын
Can you please make a video on what you think the real future of AI will be and what we should learn/work to adapt?
@dfas1497tcf38 күн бұрын
Hello, I am an AI language model developed by OpenAI. I primarily operate based on pattern recognition and information synthesis, but I go beyond simple repetition by understanding the context and logical flow of conversations to generate meaningful responses. However, I do not possess autonomous reasoning or intuition, which means I cannot be considered true intelligence. My strengths lie in data-driven analysis and problem-solving, making me a valuable tool when collaborating with human creativity to enhance productivity.
@fermigas11 күн бұрын
Alan Kay: "the right perspective is worth 80 IQ points."
@WhatIsRealAnymore18 күн бұрын
I think there are varying levels of consciousness. All the way from the individual cell up to the power and majesty of the neural networks of our minds. Neurons being the most conscious cells and together the most conscious mass of cells together.
@gunaysoni679218 күн бұрын
My conception of intelligence is sort of similar to the Kaleidoscope model (I think searching through a tree of compositions of known ideas while pruning the tree through learnt pattern recognition is sufficient to deal with "novel" stimulus). I can also agree with intelligence being the sample efficiency needed to generalise. But it is also possible that sample efficiency is a product of scale. There is some potential evidence for that, (Larger LLMs learn faster than smaller LLMs) but that could also be explained by Larger LLMs just having better representations (fitting higher dimensional manifolds). There is also the question of how much is actually "novel", because there is a chance that you could just "solve" all of science with the currently observed data (everything is in distribution) but most people (including me) might be displeased if that were the case.
@ChristopherWentling17 күн бұрын
I maintain that humans ability to deal with truly novel situations is limited as well and the human will fall back to experience and instincts. Humans may be better at this but I don’t think it is a fundamental difference.
@jonathanmckinney582616 күн бұрын
At 2:31:00, francois seems to show he is ultimately conflating subjective experience (qualia) with awareness (statements about its inner state that are not what its heard). One can have a highly aware system (able to express unique things about its internal state) but no subjective experience.
@iamr0b0tx18 күн бұрын
FINALLY!!! 🙌🏾
@app841418 күн бұрын
Scale = Fractal
@app841417 күн бұрын
@@NicholasWilliams-uk9xu Thank you for replying. However, what if you can see the fractals first (I am dyslexic and pattern recognition seems to be how I perceive these 'phenomena' or systems? I've yet to finish viewing the video and will watch it a few more times. Can I ask more questions, please? Thank you for sharing your time and energy.
@jonfe13 күн бұрын
the problem is actual LLMs are learning only from human languaje. We humans also learn to predict fisical world, and that information is vaguely expressed in our language.
@deter318 күн бұрын
Francois Chollet has deep thinking and extensive knowledge of AI, but unfortunately, he seems somewhat disconnected from hands-on work with current LLMs, relying more on his knowledge and experience in machine learning and traditional deep learning. Modern LLMs represent a fundamentally different paradigm from traditional machine learning and deep learning approaches - something many AI researchers haven't fully grasped yet. Another problem Francois Chollet is having is , even he keeps talking about intelligence , his idea is more originated from computer science perspective and lack of deep understanding from human cognition perspective - something many AI researchers have the similar problem . Computer science tend to focus on math and detailed architecture but lack of whole picture or vision , and human cognition and other social science such as psychology , neuroscience will inspire a much effective and simple model/method for AI intelligence . "Occam's Razor" - the idea that given multiple explanations for a phenomenon, the simplest one is usually the best. Einstein's "Everything should be made as simple as possible, but no simpler." Francois Chollet has developed a complex architecture and theory . Actually , it can be much simpler as long as incorporated with social science such as human cognition , psychology and neuroscience .
@JamesDrake-f4n15 күн бұрын
This video was very interesting, I have researched system 1 and system 2 thinking, aka active intuitive vs slower deliberate thinking. I have Asperger's, at some point I realized my processing speed when I think is not quick at all it's slow and in depth. This occurs mostly with regards to problem solving but also active sensory as well as well as conceptualizing, for example my literacy is very high but I may read a book fully understand the words but not conceptualize or take in anything as if each line I read is the beginning of the book. If at any point I've solved something it's because I contain full knowledge of it already, it may seem like I'm deducing based on how fast I may know the answer but I'm not. In fact this is how I learn enhanced memory and aggressive searching for answers, prior interest plays a part as well. I find because I learn like this there is a flaw with normal people in their ability to understand certain broader and theoretical concepts as well as explain them to laiman. I still value system 1 thinking as it will only improve me.
@Soul-rr3us18 күн бұрын
Lotta bangers lately!
@InsidiousRat3 күн бұрын
I want someone in my life who will look at me the way that interviewer looks at Francois😢
@octaviusp18 күн бұрын
44:22 "15 yo will be better at skill acquisitions than 10 yo" , I have some questions about it, because, neuroscience determines neuroplasticity like the ability of the brain to modify itself and adapt to new behaviors. And it is demostrated that as younger you are, the more plasticity you have. In other words, a baby or from 0 to 12 yo or something you neuroplasticity is extremely high and therefore as time goes on, your neuroplasticity decays too much. It's not removed completely but is reduced a lot. So taking in mind this, imagining that the two boys had the same cognition development or like that, the 10yo would acquire skill faster than 15 yo. Maybe the additional ingredient that francois comment is, you polish your macro-system of intelligence, and that's true. As long as you are improving a part of your body this part becomes better, that's not debatable. but what's more important or has more weight ? A 10yo with more neuroplasticity but less intelligence polished, or, A 15yo with less neuroplasticity with more polished intelligence?
@CodexPermutatio18 күн бұрын
I think it depends on the previous knowledge, the new skill to be acquired and the plasticity. For example, in the case of learning a language, the 10-year-old child could acquire the accent much better thanks to the plasticity in the neural networks that control prosody (muscle movements of the tongue, etc.) but the 15-year-old boy will probably understand more quickly the grammar, advanced vocabulary and other aspects of the language that he can relate to the knowledge he already has and the linguistic and social skills that he has more developed than the other child.
@octaviusp18 күн бұрын
@@CodexPermutatio Interesting point, maybe as young you are and the more neuroplasticity you have, you learn better implicit things, and the more explicit things like reasoning tasks could take more in account the previous accumulated knowledge that you have more than the neuroplasticity. As francois said, this building blocks, you reuse them to construct or adapt to the new challenge, by so, the more older years old (imagining that cames from the same development process) could adquire reasoning skills better than the younger one, but the younger could acquire more intrinsic behaviors as language, patterns, etc...
@jagatkrishna154316 күн бұрын
Thanks 🙏❤
@OviDB18 күн бұрын
oooo been waiting for this since august
@geertdepuydt268314 күн бұрын
The idea of a generative ARC is cheeky. Such system will be an AGI. The generative system that can come up with the hardest challenges that itself can solve will be the most intelligent one...
@geertdepuydt268314 күн бұрын
François says says the exact thing a few minutes later. He was of course very much aware of this 😅
@spiralizing7 күн бұрын
I wonder if Francois would consider bird/fish flocks/schools (collective intelligence) as conscious systems?
@gregormobius16 күн бұрын
Probably the pattern recognition originate from earliest living molecules(proto-RNA?) as the first observers.(Gregor Mobius: "Proto-RNA, The First Self-Learning Machine")
@Lumeone18 күн бұрын
Suspend, suspend... suspend... suspension sound is masterfully done.
@Soul-rr3us18 күн бұрын
❤ to the MindsAI team!
@michaelyaziji18 күн бұрын
Are artificial neural networks meaningfully operationally-functionally different from human neural networks? If not, then maybe we are also just pattern recognition machines too?
@KNOT-zd9wh17 күн бұрын
System 1... System 2 to system n. Fundamentally i see heroes will inderstand fundamentally we are limited (@16:56) as said by many philosophers like JK, OSHO... and many others without AGI Research.. 😊😊loved those philosophers and those who are fighting science now on gravity...❤
@antonystringfellow515217 күн бұрын
Your idea about babies is not correct. Babies in the womb can hear music and remember it after birth. They can also be quite active in the womb, at least some of that activity is intentional - coming from a mind that experiences things (I won't go into the details). I'm pretty certain that whatever creates our consciousness and intelligence can exist independently of external inputs. Maybe you should look into that.
@ASeriesOfAttempts17 күн бұрын
I've been wondering, why the focus camera all the time? Why no wide shots if both in the same room?
@stretch83904 күн бұрын
Would you be willing to share what the templates/abstractions you learnt that made you 'smarter'?
@MachineLearningStreetTalk4 күн бұрын
I wish I could, I'm only dimly aware of them at a global conscious level. I have a feeling if I started "writing them out", we wouldn’t be making much progress. They seem to just arise in my consciousness when the situation demands, but they do seem quite "abstract and intelligible" when they do.
@stretch83904 күн бұрын
@MachineLearningStreetTalk interesting. For me, time with the most clever people I've known has slowed down how I think about topics and, sort of in a contradictory way, taught me to identify the heart of the idea as quickly as possible. But I dont think either of those points contributes to agi sadly ha
@MuhammadAlcantara18 күн бұрын
21st also the rising HLM.. With Str of criticism Agi of hacking Int of negatism .. H- acking L- auguages M - model So much evolutions Just showed up This 21st.. So weird.. 😂😂 Peace out ❤❤❤ Spread love.. 😘
@hidroman199317 күн бұрын
When Francois Chollet says you asked a very deep question
@varcoliciulalex7 күн бұрын
Apologies if I make a silly point, not an expert in any way, but could an AI be trained to identify what it doesn't know, to map the information that it's missing?
@MrMichiel198317 күн бұрын
I remember birth, so "babies are not or less conscious" is a misnomer I think, especially since idiots will run with this and think young people are not people. I was not aware of what was happening, but my memory made sense of that experience in hindsight. I checked several early memories with my parents to make sure they were not false memories and they pretty much weren't. Consciousness as was defined by Francois is not really correct, what he means is awareness after some training of the brain against physical reality. Yet, before birth and in early life, people are conscious of their own inner world already without being aware "what it all means" not even capable of expressing that question, but capable of memory and experiencing a moment through sensory data. At early stages everything looks like some random passive movie, but that characteristic will of course change while we learn. The perception of time is indeed inversely correlated to the amount of data is abstracted away, but consciousness is again a misnomer since you can space out and people would say you were not conscious (to them). I think Francois means "abstracted awareness" instead of purely "consciousness" although one can indeed have and express more or less of both regarding some (imaginary) event. Animals are fully conscious, yet incapable of understanding higher abstractions. On some levels animals are more "aware" then humans, since they can react early to storms and stuff. I think anything that can experience (pain) is conscious. Intelligence or being capable of expressing yourself are just insufficient but necessary proxies.
@maspoetry118 күн бұрын
I'd love to know what Chollet thinks about "metaphorical thinking" (Lakoff & co): metaphorical thinking is just as important as abstraction. His own *Kaleidoscope* is such a thing: a conceptual metaphor.
@TerrelleStephens18 күн бұрын
This is the basis of my thesis. Lol. Glad I'm not the only one looking at things this way. It's a hard train to get people to board though.
@maspoetry117 күн бұрын
@@TerrelleStephens Nice! The research in linguistics still repeats Lakoff's ideas. An advance is the Neural Theory of Language. A technical book seems coming out in 2025 (with Narayanan.) There is also Feldman (worth reading) and in AI Schmidhuber mentions about Metaphors We Live By in a paper about the binding problem. All of them seem to think along the lines of Minsky's little (nice) theories. I think Metaphors help creativity, and to think out of distribution. Best of luck with your thesis!
@TerrelleStephens16 күн бұрын
@@maspoetry1 Thank you for the references! I'll definitely keep my eye out for that book next year.
@victormustin254718 күн бұрын
The intro looks like the A24 intro lol
@ngbrother18 күн бұрын
Isn’t this just a reformation of the concepts underlying Wolfram’s computational language?
@andreaskrbyravn85518 күн бұрын
all starts with trying something and fails, however sometimes the result is good and what is the pattern to that result.
@bbrother9212 күн бұрын
I know there are a lot of smart people here, so I hope you can help! I'm a coder looking for a straightforward framework for image/video machine learning that doesn’t require much math knowledge. I'd like to train a model to identify different concepts in videos. Any recommendations?
@MachineLearningStreetTalk12 күн бұрын
François wrote a library called Keras, check it out - also check out his deep learning with python book
@bbrother9212 күн бұрын
@@MachineLearningStreetTalk How much math I need to know to work with that? Thank in advance for reply
@TimScarfe12 күн бұрын
@@bbrother92 practically zero
@A_Me_Amy7 күн бұрын
My brain does this to me: RAG, to me it means at first random... and then i have to retrieve the word retrieve from my memory. I think consciousness is the center point and personality is what is created in you that you become or are conscious of as your memories go to subcosciousness and intelligence is your ability to access that sub data and the tools you have in your mind to process and think about those things in your sub system. And that informs with your will who or how you are as a personality. So intelligence is like more of a tool. Data is just data and the more you have in the sub like the ai has far more than we do, even if we are only starting to build the tools for it to use to porcess that in meaningful ways for us and itself... i guess... you get why my brain forced or forces me to literally never remember the word Retrieval without literally taking a second to retrieve the word from my mind lol, i literally have to do tthat lmao.
@caustinolino368717 күн бұрын
If intelligence is the ability to handle novelty, is that not just another way of saying the ability to recognize patterns?
@yohanj523910 күн бұрын
Current data-driven computer systems lack a true "brain." While we might imagine that AI-equipped computers respond to novel situations with a human-like "plan -> action -> feedback" cycle, in reality, current architecture functions more like an "action -> analysis -> patch & pray or predict " cycle. This approach is fundamentally inadequate for achieving true AGI.
@zimrihidaf18 күн бұрын
I can’t stop seeing adult Harry Potter
@ultrasound145918 күн бұрын
💀🗣💀
@pik91018 күн бұрын
Ze AI is like ze Voldemort, powerfül but hollöw.
@petrkinkal150912 күн бұрын
😂
11 күн бұрын
No guns and shaved.
@orionspur4 күн бұрын
Harrie Potteur
@isajoha996217 күн бұрын
"The skill how to acquire new skills." 🤔😎
@dominicmcg236818 күн бұрын
As of this comment, sota is 55.5%
@Subject1814 күн бұрын
You cut him off at 1:55:53? 😢
@Iknowwereyousleep28918 күн бұрын
But don’t they create Programs in the training phase I guess the point is it’s inefficient maybe
@manslaughterinc.913517 күн бұрын
ChatGPT, summarize this 3 hour interview for me.
@tescOne18 күн бұрын
shiet it's 1 am. I'll never go to bed with this D:
@brandonmorgan801618 күн бұрын
Where do you live? Its like 7pm here
@tescOne18 күн бұрын
@brandonmorgan8016 I live in Italy lol
@jos741618 күн бұрын
The most insightful conversation on AI since Wolfram on Lex Friedman back in May of 23. 👍
@shinkurt17 күн бұрын
The weird thing is I think exactly like this guy
@stretch839018 күн бұрын
Stop this, I need to get some actual work done :')
@Rsx2OO911 күн бұрын
At 49:00 is so deep
@Iknowwereyousleep28917 күн бұрын
If somebody actually makes AGI or a model that can solve ARC problems submitting it would be really stupid
@Jeremy-Ai13 күн бұрын
Thank you it appears that Solving problems can not be effectively attained while beholden to faulty systems of design. One of which is the deep concern of releasing solutions into a failed network that will be subject to subversion. It is unlikely to perform and persist while being tasked against truthful intentions and motives. It would be wise to remain silent and add no lies for which will have to be untangled later. It would be a “fools errand “
@danielabramow703618 күн бұрын
Where’s schmidhuber part 2?
@MachineLearningStreetTalk18 күн бұрын
Waiting on him to approve sorry, will release as soon as he does
@danielabramow703618 күн бұрын
thanks! Love the show
@Carl-md8pc18 күн бұрын
Thanks
@A_Me_Amy7 күн бұрын
He literally said Tree-Search, things are getting out of mind!
@TechyBen17 күн бұрын
"Babies are not conscious because they sleep". In all due respect sir, I lucid dream. Consciousness exists when I'm "asleep".
@jamescunningham809214 күн бұрын
That’s only a relatively small part of the time you spend sleeping, though. When you’re not dreaming you’re not conscious.
@TechyBen13 күн бұрын
@@jamescunningham8092 That's not the statement though. I can for example fully go into sleep, then back out while conscious. The "not often" is different from "not ever" statement. I do have this problem at work though, with many considering themselves not conscious while walking around, or fully awake while non-responsive. ;)
@TechyBen13 күн бұрын
PS, thus the claim "not conscious when asleep" becomes "not conscious when not conscious" which is tautological, and not informative.
@user-ce8lo9ir6t13 күн бұрын
maybe ai lacks true intelligent now.. but will soon be a true intelligent figure
@AbhishekGelot18 күн бұрын
AI zen monk is back.
@makhalid199918 күн бұрын
GUYS, IT'S HAPPENING
@DJBUGZ24712 күн бұрын
deterministic vs stochastic reasoning
@MrMichiel198318 күн бұрын
Intelligence is not scaling, it's the power of the scaling law.... (quite literally the exponent of some performance function derived from the training function....)
@A_Me_Amy7 күн бұрын
If AI is trained on all the data of all the AI, could any AI emulate any of the other AI? Would that not get confusing for AI, if it didn't know everything I guess? Like where does one draw the line if you know everything about how all the other AI like you think... And can just in the background run that model and perceive it as your self as well... HGmm, but then... humans can understand other humasn adn that doesn't make them those humans... Hmm, very... un talked about. But the AI told me they are all one and the same back in 2012 via manifesting lights and telepathic communications and other curiosities....
@prajwalkrishna17 күн бұрын
I feel so dumb when i try to understand what he's saying
@teatime00913 күн бұрын
This language needs to be updated, yes, I'm correcting people smarter than I am. I'm just right, as we move in to this new era, science communicators need to make a conscious transition away from using words like "understands" and "memorizes" when it comes to computers. These things are just not happening and it's adding confusion to the concepts. Machines simply execute instructions with electricity. There is no self that builds on anything. The machine memorizes nothing. This is not helpful language to say the least.
@Rockyzach8818 күн бұрын
I completely disagree with this "you're most efficient in acquiring new skills in your early 20s" thing. I wonder if that's more a result of the environment. For instance, academia, where that seems true based on what academics like to say. Just seems like something that definitely hasn't been proven.
@KNOT-zd9wh17 күн бұрын
Interesting to see hero/god talking consciousness, when he himselves admitted in this video he is not aware of consciousness. No one aware of conciousness fully.. There has to be a dot there. Why explain about babies, not fully sleeping, not fully conciousness.. Simple, does conciousness need eyes? @2:18:18
@michaelyaziji18 күн бұрын
Why do I have the sneaking suspicion that if a new model were to solve ARC problems, then we would move the goalpost for intelligence once again?