The Trouble with AI: A Conversation with Stuart Russell and Gary Marcus (Episode

  Рет қаралды 105,035

Sam Harris

Sam Harris

Жыл бұрын

Sam Harris speaks with Stuart Russell and Gary Marcus about recent developments in artificial intelligence and the long-term risks of producing artificial general intelligence (AGI). They discuss the limitations of Deep Learning, the surprising power of narrow AI, ChatGPT, a possible misinformation apocalypse, the problem of instantiating human values, the business model of the Internet, the meta-verse, digital provenance, using AI to control AI, the control problem, emergent goals, locking down core values, programming uncertainty about human values into AGI, the prospects of slowing or stopping AI progress, and other topics.
Stuart Russell is a Professor of Computer Science at the University of California at Berkeley, holder of the Smith-Zadeh Chair in Engineering, and Director of the Center for Human-Compatible AI. He is an Honorary Fellow of Wadham College, Oxford, an Andrew Carnegie Fellow, and a Fellow of the American Association for Artificial Intelligence, the Association for Computing Machinery, and the American Association for the Advancement of Science. His book, Artificial Intelligence: A Modern Approach, co-authored with Peter Norvig, is the standard text in AI, used in 1500 universities in 135 countries. Russell is also the author of Human Compatible: Artificial Intelligence and the Problem of Control. His research covers a wide range of topics in artificial intelligence, with a current emphasis on the long-term future of artificial intelligence and its relation to humanity. He has developed a new global seismic monitoring system for the nuclear-test-ban treaty and is currently working to ban lethal autonomous weapons.
Website: people.eecs.berkeley.edu/~rus...
LinkedIn: / stuartjonathanrussell
Gary Marcus is a scientist, best-selling author, and entrepreneur. He is well-known for his challenges to contemporary AI, anticipating many of the current limitations decades in advance, and for his research in human language development and cognitive neuroscience. He was Founder and CEO of Geometric Intelligence, a machine-learning company acquired by Uber in 2016. His most recent book, Rebooting AI, co-authored with Ernest Davis, is one of Forbes’s 7 Must Read Books in AI. His podcast Humans versus Machines, will come later this spring.
Website: garymarcus.com
Twitter: @GaryMarcus
March 7, 2023
SUBSCRIBE to gain access to all full-length episodes of the podcast at samharris.org/subscribe/ OR become a channel member to access episodes on KZbin.
Subscribe to the YT channel: kzbin.info_c...
Follow Making Sense on Twitter: / makingsensehq
Follow Sam on Facebook: / samharrisorg
Follow Sam on Instagram: / samharrisorg
For more information about Sam Harris: www.samharris.org

Пікірлер: 628
@talkingtoothpick
@talkingtoothpick Жыл бұрын
I wish we heard more from Stuart, even when Sam specifically asked him questions Gary kept butting in, pretty irritating.
@brianvaughn5070
@brianvaughn5070 Жыл бұрын
I wish Gary didn’t interrupt Stuart so often. If he’s so interested in Stuart’s opinion, maybe he could let Stuart actual voice it.
@ramakrishna5480
@ramakrishna5480 10 ай бұрын
Man , I hate that guy , he does that with every one , frankly I don't know what his contributions r to ai research
@Adam13Chalmers
@Adam13Chalmers Жыл бұрын
Listening on 4/22, about a month after this was recorded. Looks like this conversation needs to be a reoccurring monthly feature.
@organpipedesertdweller6991
@organpipedesertdweller6991 Жыл бұрын
Stuart Russel's description of the AGI concern starting at 57:00 is insightful. Then it gets the scary solution from Gary about equipping AGI with human values. Who are the Grand Overlords that will insert these values into AI to tame it ? Current Social media examples aren't reassuring.
@stuartadams5849
@stuartadams5849 Жыл бұрын
Strong request to bring Eliezer Yudkowsky back on for another podcast. His take on AI is at least as important and interesting
@emanuelschuchart1788
@emanuelschuchart1788 Жыл бұрын
Please no. He doesn't really know what he's talking about.
@emanuelschuchart1788
@emanuelschuchart1788 Жыл бұрын
@@zulfii982 in this context yes
@heinzgassner1057
@heinzgassner1057 Жыл бұрын
What a relief, there are still some sane people around on this plant. Thank you for this excellent reflections. Powerful human reflections, based on brilliant human (meta) consciousness.
@dougg1075
@dougg1075 Жыл бұрын
The other side says the same thing. Now what?
@ArcadianGenesis
@ArcadianGenesis Жыл бұрын
It's called the _symbol grounding problem:_ how does anything understand anything? How do you connect arbitrary symbols in the mind to real things in the world? It's been a philosophical problem for thousands of years. I wrote my PhD thesis on it.
@robadkerson
@robadkerson Жыл бұрын
Seems to me they are taking away the AI's capacity for understanding by definition. It's the same repeated flaw of human exceptionalism that we've seen for millennia.
@integralmark
@integralmark Жыл бұрын
@@robadkerson humans are the apex predators of the biological world, and we are the result of millennia of evolutionary processes happening. The use of the construct "flaw of human exceptionalism"-- poking at the idea that humans are different from everything else -- is to mock the idea that humans have free will (free will and cultural/knowledge inheritance being the pillars behind the idea of human exceptionalism) since all of our capacity to think and reason and perceive comes from our biological heritage (this is Sam Harris's wheelhouse). I would argue that to consider the implications of the paradigm expressed in the words "flaw of human exceptionalism" is to try to give consideration to other biological species, such as dolphins, ravens, apes, etc. etc. ad infinitum by noting the commonality of all species--all species including humans are evolutionary expressions of the biosphere. It's to point out that to preference the apex predator status of humans in comparison to other species has its limits and those limits can be felt into or intuited by realizing that humans are now on a path to wipe out almost all other species, and other species should be honored. It's to foreground the idea that humans are now at a point where they can destroy all other species and it's an appeal to ethics or morals or a deeper truth as a way to try to prevent an utter mass extinction even though predation and death and hierarchy--big fish eat little fish--is a part of of the evolutionary cycle of life that we as humans did not create. But here's why I wasted an hour on a few paragraphs in trying to formulate this comment to make it semi coherent: To use the definitional paradigm of challenging the idea of human exceptionalism to defend AI is to claim AI is alive in a way that is in alignment with the biosphere, or that it is of the same composition of the biosphere, it's almost giving AI victim status and it's certainly granting AI species status and that is bullshit. AI is a simulacra, and the conversation in this podcast is intended to bring to awareness how it simulates and the huge interpretive gaps it has, that literally all the energy in the world soon won't be enough to help it improve its interpretive capabilities in a way that could be called intelligent. This last sentence can be debated and is to be determined, and I would ask that you separate it from the above (as if I had any control over your reaction)but I include it here: Honestly, people bend over backwards to point out the good in the AI project, when all I see are endlessly multiplying existential threats comparable to the nuclear bomb.
@robadkerson
@robadkerson Жыл бұрын
@@integralmark you've excellently, and accurately described the threats it poses. I agree. It's here, not "in our lifetimes," it's happening now.
@tupacalypse88
@tupacalypse88 Жыл бұрын
Is there a reliable way to tell if it understands or not?
@robadkerson
@robadkerson Жыл бұрын
@@tupacalypse88It's subjective. If you ask it questions about a topic, and it answers your questions. Then it understands. If you ask your questions about a topic, and it doesn't answer your questions, then it doesn't understand. But you can see how that is extremely subjective because you decide what questions to ask
@DeleriousOdyssey
@DeleriousOdyssey Жыл бұрын
Gary Marcus is the type of guy to walk away from a conversation with what he put into your opinions as opposed to your actual opinions...
@soisas2811
@soisas2811 Жыл бұрын
seriously. that was so annoying, he kept interrupting and barely let the other guy talk and his points were tame and so exhaustingly one sided.
@nickidaisydandelion4044
@nickidaisydandelion4044 Жыл бұрын
Those are the right guys for the CIA and the corporate agenda.
@MatthewCleere
@MatthewCleere Жыл бұрын
As far as the "do your own research" I will quote myself here: "Karaoke made people think they could sing. The internet made people think they could think."
@inezgraer5482
@inezgraer5482 Жыл бұрын
It kind of makes me feel safer knowing Sam's got my back! He has stepped into the shoes of Christopher Hitchens while still being his own man. We need many ,many more people like this who are not afraid to expose topics like this, that even the uninitiated can follow and understand. He is constantly mindful that his audience grasp the subject matter. Thank you Sam.
@samdg1234
@samdg1234 Жыл бұрын
You > *"He has stepped into the shoes of Christopher Hitchens"* I love Hitch as much as anyone. But he was far, far from perfect. Here is a guy that admired and was hugely shaped by Hitch, who nevertheless was willing to point to some of his nonsense. *"kzbin.info/www/bejne/nKDToGx7bLp0jLM"* It will very likely be difficult for you to watch.
@kiwicalibre
@kiwicalibre Жыл бұрын
"He has stepped into the shoes of Christopher Hitchens" 😂😂😂 Hitchens wouldn't even acknowledge Sam these days, other than to note his specular tumble into the swamp that he railed against all his life.
@THEMAX00000
@THEMAX00000 Жыл бұрын
From what I know about Christopher Hitchens you are 100% correct This guy makes about as little sense as he did
@THEMAX00000
@THEMAX00000 Жыл бұрын
I feel so terribly sorry for you, and that you actually think this guy “has your back” and that you “feel safer” This is not meant to offend you, but you really need help
@TheGodsEye82
@TheGodsEye82 Жыл бұрын
You honestly believe Christopher would preach to the world to have absolute blind faith in the government & pharmaceutical companies?
@michaelbruhl
@michaelbruhl Жыл бұрын
I'm 10 min in, and I can already say, this is the best podcast I have heard from anyone on these topics. Thank you for confirming I'm not losing my mind!
@mysterious_miracle
@mysterious_miracle Жыл бұрын
Let me check it out then
@samdg1234
@samdg1234 Жыл бұрын
10 minutes from the start? Or 10 minutes into the topic of the video? The real topic of the video doesn't start till nearly 10 minutes in. I found the first 10 minutes fantastic. It is why I listen to him. (too bad he doesn't follow some of his own advice sometimes). I'm not too interested in AI, not that I shouldn't be.
@snooks5607
@snooks5607 Жыл бұрын
first on topic response from guest is at 15:45
@ed345
@ed345 Жыл бұрын
You being a moron I can understand your excitement!
@robadkerson
@robadkerson Жыл бұрын
I don't know why we're so opposed to using the word "think" for an AI. To say that it isn't thinking ignores that we are basically just pattern recognition machines ourselves. Unless you think we have free will
@zerotwo7319
@zerotwo7319 3 ай бұрын
We have many more systems than just 'pattern recognition'. Emotions like logic are one. machines don't have that yet.
@troy3456789
@troy3456789 Жыл бұрын
It took months after I realized free will was an illusion for it to truly sink into my bones. It completely changed how I treat everyone. I'm not even rattled when someone launches into calling me names when they disagree with what I have said; not even a little bit. It has been transformative.
@dances_with_incels
@dances_with_incels Жыл бұрын
Free will is overrated anyways
@moriartyco
@moriartyco Жыл бұрын
Wow, you got that from Sam's book? I'm intrigued enough to read it now.
@troy3456789
@troy3456789 Жыл бұрын
@@moriartyco I actually got that from the Waking Up app, but it is also available on his Making Sense podcast. I listened to it on the Waking Up app, then saw where he put it on his Making Sense app. I am not joking. It changed the way I interact even with my wife. I realize that it is so important that it would end all personal violence in the entire world if the rest of the world discovered it is an illusion. It transforms the way people talk to each other, and address each other, and think about each other at a foundational level. You will lose all rationale for hating anyone, ever. It might be overnight, it might not. I think interlocking this with mindfulness is key, so that reacting is no longer natural to you. It won't mean that when someone is belligerent that it will not shake you up. You are a social creature, and it is impossible to avoid being affected by the behaviors of others; but you can avoid making it worse; and your recovery from it will be faster. I cannot put a dollar value on these two important lessons (Mindfulness, and the absolute realization of the illusion of free will), because they literally are transformative. I cannot recommend highly enough.
@moriartyco
@moriartyco Жыл бұрын
up well thanks for that, I'm not totally sold on what Sam has to say but I'll definitely read the book now you put it like that!
@troy3456789
@troy3456789 Жыл бұрын
@@dances_with_incels It's not that it is overrated, it is that the entire concept of it is an illusion. it turns out to be a crappy illusion too. We didn't choose our parents or what they thought was important for us or for themselves. We really cannot even choose what foods we like or which ones we dislike. If we could, you could choose to suddenly enjoy foods that currently make you nauseous. Nobody gets to choose what they like or dislike (about anything). Nobody chose their sexuality anymore than they chose their eye color or skin color. Realizing just how many constraints their were our education and our upbringing; it becomes more and more obvious our certainties, our beliefs and opinions are entirely involuntary. We lose all rationale for hating anyone for it.
@jakimson1994
@jakimson1994 Жыл бұрын
very nice podcast, but didn't appreciate Gary Marcus interrupting to disagree. Disagreeing is fine, but let the other person finish some thoughts at least.
@idealmasters
@idealmasters Жыл бұрын
I’m particularly worried for our older populations. People above 50 really struggle with obvious online falsehoods. Thinking their phone has a virus from a popup, thinking edited photos are real etc. They have no chance against this.
@donkeychan491
@donkeychan491 Жыл бұрын
“Obvious online falsehoods” like Covid possibly originating from a lab leak.
@MrSonny208
@MrSonny208 Жыл бұрын
So true, very few are interested in being sharp enough to keep up with new stuff
@xact3855
@xact3855 Жыл бұрын
I'm particularly worried about the younger generation. I'm 50 and have had to rant and rave at my teenagers to get with the AI revolution. As for believing stuff well the art of Simon Stålenhag I initially wondered if it was AI generated. How is that for irony! (Bought one of his art books - recommended.)
@jg1091
@jg1091 Жыл бұрын
Oh the other hand , speaking to a computer might be an easier interface for older people than clicking.
@ericfan9149
@ericfan9149 Жыл бұрын
I’m over 50 and less credulous about online falsehoods, deepfakes, and misinformation than many younger people I see on twitter and facebook.
@Unhacker
@Unhacker Жыл бұрын
Regular drops from Sam keep me sane. Just a reminder that there are still other rational people still out there. Thanks for that.
@samlloyd672
@samlloyd672 Жыл бұрын
Sam is not even close to rational. He is another emotion first backwards rationalizer. His entire stance on COVID, vaccines, trump and the “insurrection” hoax has shown it.
@PetraKann
@PetraKann Жыл бұрын
Keep you sane? He was dead wrong on every aspect of the pandemic and is still providing fake apologetics for mRNA jab mandates, lockdowns, masks etc. And you obtain sanity from deceit and self centred hypocritical nonsense? You are kidding right Mr Ransacker?
@theminister1154
@theminister1154 Жыл бұрын
This is probably a good podcast, but the idea that Sam Harris is rational is, at this point, _utterly_ ludicrous. He's one of the most biased commentators I can think of. *LITERALLY* he said he'd rather have Hunter Biden get away with pedophilia than have his monstrously consequential laptop publicized. Sam Harris is a scumbag 2023, and there is absolutely no way to get around that. You need to admit this to yourself. You can still listen to Sam, but his ludicrous bias & clinical Delusion must be front of mind when you do. Sam is mentally ill. He's about as rational as, say, Margorie Taylor Green: _not very._
@stoutz
@stoutz Жыл бұрын
Agreed, always nice to hear Sam's level-headed and rational discussions.
@wasdwasdedsf
@wasdwasdedsf Жыл бұрын
what
@EvgeniiNeumerzhitckii
@EvgeniiNeumerzhitckii Жыл бұрын
Sam has a conversation with two academics. And it is funny how they don’t agree with each other and keep interrupting one another constantly. Amazing episode, please do more like these (with multiple guests).
@EvgeniiNeumerzhitckii
@EvgeniiNeumerzhitckii Жыл бұрын
Let me finish my sentence! Lol
@EvgeniiNeumerzhitckii
@EvgeniiNeumerzhitckii Жыл бұрын
Basically, for two hours the two academics, both experts in the field, were trying to understand what the other guy was actually saying. :) What amount of understanding do you expect from a normal person?
@Fritz.program
@Fritz.program Жыл бұрын
@@EvgeniiNeumerzhitckii great point
@michalbreznicky7460
@michalbreznicky7460 Жыл бұрын
I felt bad for Stuart as he was interrupted all the time, especially by Gary. Moreover, Gary's arguments felt naive at times. While it's very useful to get an outside view (of people outside of AI) to make sure it's not (solely) the AI people deciding on the objectives, having an opinion on objectives doesn't make you an expert on AI technicalities.
@robertblake1228
@robertblake1228 Жыл бұрын
doesnt that shows how off Sam is lately? He is acting so bitter.
@LotusHart01
@LotusHart01 Жыл бұрын
Drink every time Sam say’s *orthogonal* 🍻
@sunnyinvladivostok
@sunnyinvladivostok Жыл бұрын
only if drinking is in your wheelhouse
@wolfumz
@wolfumz Жыл бұрын
Re: the intro, I had someone the other day tell me that my college education made me dumber. He was saying, in so many words, my degree was a testament to how I had been indoctrinated into woke ideologies, where I learned untrue things about minority groups and gender. When I explained my degree was in comp sci and statistics, and I did not spend time on gender studies, he held fast anyway, lol. I guess those posts he read on social media about the work agenda had more authority on the topic than I did. There is this strain in American thought today, where some people have deep contempt towards the idea that someone else may know something they don't. They find it offensive that a doctor can read and interpret a clinical study with some authority. But, man, if a rich businessman in an unrelated field disagrees with that doctor, then you better listen up, lol. That's because rich businessmen are always right. If you spent 30 years becoming an expert on medicine or engineering or something, then you can go and fuck yourself. But if you have a lot of money from selling plastic spider rings and herbal supplements, then I want to hear what you have to say about this vaccine.
@PentUpPentatonics
@PentUpPentatonics Жыл бұрын
This one really shouldn’t be behind a paywall. Too important.
@anewman
@anewman Жыл бұрын
Gotta say Gary sure seems to be trying too hard to convince himself we aren't on a path to AGI, I mean when we're talking about something passing the turing test but unable to multiply 4 digits, we are forgetting the part where even humans don't really do that on their own. They use calculators. Give the poor AI a calculator program to consult and the problem is solved.
@robadkerson
@robadkerson Жыл бұрын
Exactly
@anthonyyaghi6342
@anthonyyaghi6342 Жыл бұрын
You re totally missing his point which is that chatgpt has no model of mathematics and can’t logically come up with the answer to a math problem that didn’t see before. So what it does is just give you the closest thing to it. Which is completely different to how we human think of the problem, we have a concept of what numbers are and how they can be manipulated.
@billjohnson6863
@billjohnson6863 Жыл бұрын
@@anthonyyaghi6342The question is, is that out of reach for deep learning? For example, transformers have been shown to be able to learn simple algorithms which is closer to how we think.
@anewman
@anewman Жыл бұрын
@@anthonyyaghi6342 you're over thinking it. I'm not saying the ai has any part of solving the math equation. When you type enter the question is also passed onto a calculator program and the answer you get is not from chatgpt but the calculator program disguised as chatgpt.
@gJonii
@gJonii Жыл бұрын
​@@anthonyyaghi6342 I take it you haven't asked ChatGPT anything related to mathematics? ChatGPT is quite rigid in its thinking, but it can easily do mathematics, just ask it to do it.
@MG-tk2ic
@MG-tk2ic Жыл бұрын
GPT4 wrote this summary: The discussion starts with the consequences of the arms race in AI technology, focusing on the potential destabilization in global politics and economics. The speakers explore the idea of "winner takes all" in AI development and its potential impact on global power dynamics. They discuss the importance of international collaboration and creating a global governance structure to prevent misuse of AI technologies. The conversation touches on the possibility of an AI-induced unemployment crisis and the need for governments to prepare for such a scenario. They delve into the role of social media platforms in the spread of misinformation and the erosion of democracy. The participants examine the responsibility of tech companies in addressing misinformation and the potential benefits of fact-checking and content curation. The discussion shifts to the metaverse, exploring its potential business model and the ethical implications of AI-generated fake friends. They address the importance of AI literacy, web literacy, and educating the public on how to discern reliable information sources. The conversation moves to the potential dangers of AGI and the need for alignment between AGI and human interests. The speakers discuss the various opinions in the field regarding AGI's potential risks, emphasizing that self-preservation might arise as a sub-goal for AGI even if not built-in, and questioning why some experts seem to ignore or misunderstand this concern.
@gracefitzgerald2227
@gracefitzgerald2227 Жыл бұрын
Making Sense without Sam’s voice is like The School of Life without Alain de Botton. Just not as good. ❤ love Sam’s voice.
@bdhanes
@bdhanes Жыл бұрын
Love both of them! ❤
@Xelbiuj
@Xelbiuj Жыл бұрын
I wish Sam would have a conversation with Robert Miles (of his own youtube channel and computerphile fame) Hearing them dance around terms like "orthogonality thesis" or "terminal goals" or "instrumental goal", without using the words or defining them . . . is lacking.
@damonm3
@damonm3 Жыл бұрын
Just finished this on the podcast channel and had to come here to say a few things. If this is what 2 experts are discussing now when things are moving or will move very soon to beyond human input is very scary. I’ve been paying attn to AI basically in headlines over the last handful of years. Know very little. I got nothing from this. Not only were the thoughts basic and common sensical, they were what most people that aren’t entirely stupid are thinking of. The fact that there are no solutions in practice now just shows how things will very very likely go. Listening to Gary and Scott argue in the later half of this conversation reminded me of how I used to communicate with my brother when we were teenagers. The fact the Gary said something so utterly out of touch with reality and any realistic timeline and it went untouched by both Scott and Sam just blew my mind. I had to rewind to make sure I heard what I thought i did. He said “when AGI is way smarter than us like in 1000 years” or something similar. What??!!!! This is something an expert said? Wow. Lost all hope for all involved. Hey Gary, did you know running the right program with enough compute you can speed up timelines? Gary’s 1000 years is at the very most a decade or two out after asi is out of the box. Just depending on how it utilizes compute systems… could happen in a matter of months after. I have very little hope in humanity in general with so much being revealed to us over the last decade or so. Thinking “well at least were past the invade and concur phase for the most part..”. At least in the west. And keep in mind while most likely we westerners are at the leading edge of AGI and ASI r&d we’re not the only ones playing the game. All you need is a gpu farm with recent tech and some smart people and some electricity to run stuff. I’m sure gpu chip control will go into high gear now and forever into the future at it’ll controls AGI… well that it until the coin flips of course. God I sure hope there are more inspiring experts working on the system that’ll take over the world… and maybe hearing them speak might be insightful. But this wasn’t it that’s for sure.
@toby9999
@toby9999 Жыл бұрын
Is AGI possible in a deterministic machine? Too much hype in my opinion. I don't believe the human ability brain can be duplicated by a gpu farm.
@alexanderg9670
@alexanderg9670 Жыл бұрын
You can make ChatGPT mathematical genius by forcing it to reason step by step, to write a program for a calculation etc. Future of AI is in interconnections of specialist AIs and expert systems, fine-tuning is what propelled ChatGPT over GPT-3 in usefulness. Humans will have to learn how to ask the right questions
@KungFuChess
@KungFuChess Жыл бұрын
Agree with Sam, the click bait for ad dollars model that drives the Internet today is a big problem going into the future with AI.
@SolitaryReaper666
@SolitaryReaper666 Жыл бұрын
Title wrong?
@jeffkilgore6320
@jeffkilgore6320 Жыл бұрын
I read all the time about favorite authors and interviews. For me, it’s Sam Harris.
@christopherhitchens163
@christopherhitchens163 Жыл бұрын
Listening to experts on foreign policy, as if they were scientists leads you to following Henry Kissinger (I hope you understand what that means)
@OhManTFE
@OhManTFE Жыл бұрын
Hitch back from the dead spitting mad facts
@Space-wh4vs
@Space-wh4vs Жыл бұрын
Sam your articulation can be more clear about Ending thought! "meditation". Jiddu Krishnamurti-(he articulated this phenomenon quite well) had his whole life devoted to "Tell " people about this phenomenon of human consciousness. If you are seeing this comment, just take a look at this man.
@sampoole2555
@sampoole2555 Жыл бұрын
The trouble with AI is that everyone thinks that AI is trouble. The human mind goes "danger fight/flight" whenever there is something new and unknown. We all know this, so why do we go on ignoring that little voice inside that is telling us we're being stupid about AI? Rich tech billionaires are never going to be your friend. They have a best friend whose name is money, and they will never betray it.
@g0d182
@g0d182 Жыл бұрын
48:57 & 49:30 --- Stuart gives a reasonably honest, sensible estimation 49:10 --- Gary responds with apparently unevidenced absolute claims Criticism can for eg, reasonably occur as Stuart demonstrates; i.e. reasonably balanced (non absolute sounding claims)
@skitsandjiggles7286
@skitsandjiggles7286 Жыл бұрын
One of the limitations from decades ago that still exists now is filtering the AI learning process through very limited human language. Even image learning AI algorithms relay on a form of parsed language. The day we see an AI system observe its environment and start learning how to navigate it the way a baby does will be evidence that we're nearing that point where AI will just be called I.
@Branko353
@Branko353 Жыл бұрын
There will need to be some kind of learning necessary before the system is capable of navigating environment on its own. Babies don't come as blank papers. There is enormous amount of embedded knowledge and character including things like predisposition to religion. Even then we learn most of our knowledge in some sort of language. Pre existing knowledge and communication should be the foundation of learning for a humanlike AI.
@skitsandjiggles7286
@skitsandjiggles7286 Жыл бұрын
@@Branko353 When I started working with neural networks in the mid-90s, there was no reason for these systems to pre-learn before learning from the real world. Any preexisting tendencies or limitations came from the structure of the network itself. Something similar to what people refer to as hardwiring of our brains. The moment before you turned on one of these systems, nothing was stored. I worked with visual networks, so observation of what was in front of them was where the learning began. Tactile systems learned by bumping into things, falling over, etc. You don't need a language for that, just the weighted network structure. The issue we have right now is everything runs on outdated x86 hardware, and we depend too much on the software (languages) to do most of the work. That's why so much money is currently going toward the development of AI chips based on RISC tech. They will essentially become the neurons in new systems that won't need a high-level programming language to work. I didn't initially imply that babies start out as blank slates. But since you mentioned it, I guess we all start as blank slates. Those first few cells of our brain have no electrical activity. During development in the womb, hardwired reflexes and instincts start forming (moro reflex, gag reflex, facial expressions, sucking impulse). Most of that is due to neural structure, not learning. After 2 months, there is a considerable amount of learning about our watery environment, the strange vibrations, and the odd sounds from that woman we can't see. None of this nor the initial learning after birth depends on symbolic language. The auditory cortex, tasting, and touching dominate that period.
@jeremypfrost
@jeremypfrost Жыл бұрын
Gary could not let Stuart finish a thought. So frustrating to have and listen to conversations with people people like this.
@stephang6335
@stephang6335 Жыл бұрын
About AI not being able to "understand" what it's writing, writing about, etc... In the demo of GPT4, it's shown a quite hilarious picture of someone ironing on the back of a taxi, and not only did it recognise what it's seeing, it was also able to answer questions like "what is hilarious about it". Isn't this "understanding"?
@QwertyNPC
@QwertyNPC Жыл бұрын
The very question leads the AI. I suspect the result wouldn't be as spectacular if there wasn't any suggestion of humor inside the question. The rest is I imagine image a juxtaposition of unlikely juxtaposed things. I suspect if you present it with a very unfunny picture but showing something that's unlikely to be in the same picture the results would be similar.
@goldwhitedragon
@goldwhitedragon Жыл бұрын
On a material, deterministic level.
@hwlbausch
@hwlbausch Жыл бұрын
Oh Sam, my old friend so dear, How I miss your presence here. Your ego free approach to life, Was always such a breath of light. You never sought the spotlight's glare, Or craved attention from others' stares. Your humility and kindness shone, In every interaction, every tone. We used to laugh and talk for hours, Sharing stories, dreams and powers. Your gentle spirit touched my soul, And helped me feel whole. But now you've gone, and I'm alone, The silence deafening like a stone. I miss your calm and centered grace, Your smile, your warmth, your loving space. So though we're far apart in miles, And time has passed us by in trials, Know that I hold you close within, And long for our friendship to begin. For in my heart you'll always be, A cherished memory, a part of me. And when we meet again someday, I'll thank the stars, for guiding our way.
@manneborg
@manneborg Жыл бұрын
"If I could just jump in..."
@benjammin8184
@benjammin8184 Жыл бұрын
All three of you have surnames that are also first names, great stuff.
@respectfulgamer7232
@respectfulgamer7232 Жыл бұрын
I don't know why but that gives me pleasure.
@maxgregorycompositions6216
@maxgregorycompositions6216 Жыл бұрын
Jammin is also a surname. I knew a James Jammin once.
@dougg1075
@dougg1075 Жыл бұрын
We may create an artificial intelligence that screams when it becomes sentient, and by the time we shut it off, we find out that to the artificial intelligence, it has spent 1000 years in pain, but only a few minutes past here. That would be horrendous.
@mychannel5019
@mychannel5019 Жыл бұрын
Lay off the weed.
@charlesfarrant6819
@charlesfarrant6819 Жыл бұрын
I wonder if Sam will comment on other proponents of the 'rationality movement' as has occured online over the last few years. I think fondly of guests like Eliezer Yudkowsky and wonder if Sam's heard of Scott Alexander
@patricksoleimoni
@patricksoleimoni Жыл бұрын
Considering Sam signed the pettition to the NYT to not dox Scott, I'd say he's heard of him.
@wasdwasdedsf
@wasdwasdedsf Жыл бұрын
@@patricksoleimoni petition to not dox scott?
@wasdwasdedsf
@wasdwasdedsf Жыл бұрын
what should he be commenting abuot as to that movement?
@charlesfarrant6819
@charlesfarrant6819 Жыл бұрын
@@wasdwasdedsf just that there are channels through which people like us follow a feed of discussion about the important and valuable aspects of rationality and with a pragmatic attitude towards current events. Some are saying Sam doesn't have much to speak on with certainty anymore but reiterate the same fundaments underlying our concerns and uncertainty regarding emerging technologies and current events. If Sam's not going to have an expert on to practically teach us the basis for making every relevant decision reasonable people are responsible for, then he needs to converse with other informative bodies, he needs to be free to share the responsibility of communicating these issues with other agents, he'll never consider himself qualified to act on their advice if he doesn't take on their honous, and in that way in a relationship people need to be considered responsible for the interpretation of agents outside their own skin, as allowed by common law. I don't know what the future of sense making will look like, but im convinced Sam has a good idea of what it should look like, and while most people can't do much nowadays we need to find a way to communicate with people that should by all reasonable judgement want to appreciate our opinions about what sense all people should be making. I just see him as stagnant, and im not sure if the answer is collaboration as he does more or less effectively now (not very effective in this podcast i don't think), or amassing a steelman, encouraging reading of a syllabus and trying to find the consistent basis for rational action for the layperson to attempt to live a good life, and to co-ordinate and agree on where we can afford to go in this as current uncontrolled juggernaut of tech. We could all decide roughly what life we want to try and start leading, we can solve for many problems we have. But right now we're just doing everything, and to my mind the purpose of this community was to make better decisions and convince people they should care about important things for their own sake, but for god's sake we need to open each other up to the idea we're attempting to do the best for everyone in that regard. It's all getting a bit chaotic and who's to say that isn't good. But conversations on here aren't necessarily progressing at the rate many of us are able to discuss better solutions. Ask yourself in what way are we not powerless, and then cement those agencies you want to see turn the world into a better place. What i have learned from Sam and his guests is that at large i can reconsile the opinions of almost everyone, and so we can act on the same page. If we have the intellect, give us the influence, don't make it tedious, create better ways of making concepts and inferences, modelling and mapping and having discussion, pick up Sketch or make informatics mainstream. We could do a little more and i stifle my own creativity, but we don't have to, i appreciate im asking, and not really leading but for trying to talk with people i think i can really converse with for the betterment of everything we collectively understand, make it that easy. Idk, it's most important to me Sam stays transparent and keeps us updated on his understanding of the inferences he draws, but in order for this to be done best i think he needs people to trust him, and we need to talk more widely about the phenomena we see having corrupt influence.
@charlesfarrant6819
@charlesfarrant6819 Жыл бұрын
@@wasdwasdedsf i would value hearing Sam retort some of Scott A's essays for instance, break them down logically. Scott did a good one on Truth as an aesthetic weapon for instance. It'd be complicated, and would involve tricky proofs like Kurt Gödel, but you could talk about the landscape of reason on more ways. Just as Sam eloquents quite perfectly and generally with nuance and challenge at the humanitarian figures as we're swamped by such affairs. Again just a reading list, to do some research on his influence would be a cool update. As i said Eliezer Yudkowsky has a lot to offer, moral, yet rational, and offering an understanding suggestion of meaningful religious connotations too. Don't have to take the whole package. Just saying we could fill each other in on this stuff, my work, your substack, essays that are constantly being written, it's all meaningful. And when it comes to the problem of logic and communication itself lets have a workshop. I appreciate much of this is outlined throughout the paradigms of all the systems discussed in podcast, but that's not the only form, and i think we as a community are about more than this, and if we can accurately interpret and convey what's important we could live richer lives, is something i must only assume in the foreseeable term
@DouwedeJong
@DouwedeJong Жыл бұрын
So if somebody write an article with the title "In the defense of torture", what are we to think?
@bronnismofo
@bronnismofo Жыл бұрын
It's amazing how many experts in bio chemistry, molecular biology, and medicinal chemistry turned up when the vaccines became available...
@billscannell93
@billscannell93 Жыл бұрын
I wonder if too many people were taking advantage of the free subscription option on his website. (I could hear the full episodes when I first signed up, but no more.) It sucks when you're legitimately broke!
@jdarst100
@jdarst100 Жыл бұрын
Ditto
@Lyle-bw3nx
@Lyle-bw3nx Жыл бұрын
brilliant comversation
@gulllars4620
@gulllars4620 Жыл бұрын
Was this ChatGPT based on 3.5 or 4? Because some of the limitations of ChatGPT they are talking about are radically improved or solved when based on 4. Like around 32:00 they talk about limitations about knowing about people being in possession of a wallet and that possession transferring. Theory of mind is at the level of a 7-9 year old human for GPT-4 instead of a 3-5 year old for GPT-3.5. Those numbers seem close, but the expressive representation difference is stark. And yes, it still fails on large calculations but is much better in general, but critically GPT-4 based ChatGPT can use tools like calculators, and can learn it should do so if it has access. My main worries are what are people able to do with such powerful tools, and what will the next set of novel emergent capabilities be in GPT-5 (or the incremental road to it like it seems OpenAI is going with). The theory of mind capability was not for known GPT-3.5 before someone discovered it in GPT-4 and then tested GPT-3.5 and earlier LLMs to check if they had it and at what level. I don't recall where, but there is a table in some paper i read of emergent capabilities in LLMs as a function of their parameter and compute scale. If you can find it i highly recommend that one. A lot of new novel capabilities are emerging at parameter and compute scales from GPT 3 to GPT 4, and in particularly closer to the scale of 4 then 3 or 3.5. We may be under an order of magnitude scale from a lot more capabilities given how close the thresholds for emergence of many of the known capabilities in GPT-4 is.
@heinzgassner1057
@heinzgassner1057 Жыл бұрын
One prerequisite for avoiding absurd chaos in society is SSI in form of Decentralized Digital Identity - the basis for accountability, for clear distinguishing between ‘bots’ and humans, for origin-watermarking of content and events. This is where AI needs to be supported by WEB3.0 innovations and new governance for our technological age.
@Somegirl51
@Somegirl51 11 ай бұрын
"Don;t know what we have now." These conversations seem to be too late and so many of us are going to be hurt. I fear for my children and grandchildren and their future. But no one seems to give a sh**
@heinzgassner1057
@heinzgassner1057 Жыл бұрын
All this AI discussion also unveils a very ‘religious’ assumption of mainstream science: Consciousness is - per assumption - seen as an emerging property, emerging from matter within our given space-time world. This ‘axiom’ is getting more and more questioned, not by spiritual weirdos but by the brightest physicists who keep on asking the ‘why and what’ (difficult after almost 100 years and just focusing on utilizing quantum physics instead of ‘understanding’ it). More and more facts are pointing towards ‘consciousness’ being foundational, being the ‘no-thing’ from which our space-time-matter-would arises each and every moment (I am addressing meta-consciousness, the ‘knower’ who is knowing our thoughts, feelings, sensations and perceptions). This is of foundational importance for how we discuss AGI. Before understanding this better, we are just having confused discussions in the fog.
@truwth
@truwth Жыл бұрын
I wonder if Sam tried to have Pinker in this conversation rather than Marcus. Would have been interesting.
@Jake-Day
@Jake-Day Жыл бұрын
“Fool me once, fool me again.” - Sam Harris
@sircharlesnot
@sircharlesnot Жыл бұрын
Your behavior towards Bret Weinstein belie your entire argument here
@blengi
@blengi Жыл бұрын
hmmm, how come when I ask chatgpt to list a largish bunch of words in alphabetical order it seems to do it quite commendably? That's not some probabilistic continuation. It's almost as if there's some sort of emergent conceptual understanding that alphabetizing implies reordering things per some higher level abstraction beyond the words. How does prediction systematically get higher level order from a list of random words that have no semantic connection?
@stephenknox2346
@stephenknox2346 Жыл бұрын
I thought Gary was in the camp of anti-AGI. Instead he just seems interested in AGI arriving in a way that will prove him correct. Not comforting.
@valentinrafael9201
@valentinrafael9201 Жыл бұрын
There are chess positions that regular ( non-AI ) chess engines, which just use brute force, can’t make the right move. You gotta set up the board for it. Not something that will happen in a regular game
@JonnyWisdom
@JonnyWisdom Жыл бұрын
8:40 Sam I disagree somewhat, I am trained to Masters level and started a PhD. Ordinary folk can engage in meaningful research if they dedicate themselves to the truth, rather than just confirming their bias. I would argue that in some cases you do not need to be well read in the literature to make sensible logical conclusions, if you use deduction and reasoning, this is what we use to call "wisdom".
@snooks5607
@snooks5607 Жыл бұрын
which part are you disagreeing with though? to me all he seemed to be saying is that amateur research is not a substitute for professionals and people shouldn't act like we could just get rid of the idea of there being authorities in fields of knowledge. I don't think he'd ever suggest that amateur research can't be valid, that's been done since the days we were banging rocks together, it just needs to be validated by professionals to be accepted
@Narikvp
@Narikvp Жыл бұрын
@JonnyWisdom, i agree with your fundamental assertion, but do you not agree that if we offload some of the investigative responsibility onto ai chatbots or whatever comes next, that humans would slowly lose the edge to their reasoning and investigative skills? In the context of research, if I am told that a chatbot is capable of summarizing a research paper to 99% accuracy, I may feel comfortable living with its conclusions rather than drawing my own conclusions from actually reading the paper.
@Narikvp
@Narikvp Жыл бұрын
My apologies, I did not recognize the specific point that you were challenging in the podcast, which didn’t have anything to do specifically with ai or chatbots.
@goodtothinkwith
@goodtothinkwith Жыл бұрын
1:13:00 the Netflix example was very compelling. It sounds like we need to use something like AutoGPT to optimize a meme that raises awareness of how poisonous and virus-like algorithms that support ad-driven revenue actually are. I can ask ChatGPT to explain something to me as if I was a high school student. Can we ask it to put that point in terms that will resonate with your average person? Maybe not now… but there may be machine driven antidotes latent in these systems too
@AndruXa
@AndruXa Жыл бұрын
Without Sam this discussion would derail in the first half an hour. It seemed to me that most of the time when there was a small misunderstanding it was Gary overhearing what Stuart or Sam very cleared laid out.
@OnYourMarkgitsitGooo
@OnYourMarkgitsitGooo Жыл бұрын
I don't understand why it's very hard for Gary to grasp that we only have 1 chance to get AGI right or we most likely die. He keeps going off tangent while he rambles incoherently
@DestroManiak
@DestroManiak Жыл бұрын
"off the record" doesnt mean anything unless you wait for acknowledgement.
@AdamGeest
@AdamGeest Жыл бұрын
So why does the Financial Times article indicate repeatedly that a computer itself was able to isolate this weakness in the GO playing AI’s circuitry?
@mallow610
@mallow610 Жыл бұрын
Why does Gary keep talking for Stuart? I would not be able to deal with a conversation with him
@GreylanderTV
@GreylanderTV Жыл бұрын
56:58 This discussion of recommender systems, suggests to me something much better than the paperclip maximizer and AI the goes rogue and destroys or takes over the world. _Imagine an engagement maximizer, such as the youtube recommend algorithm, driven by its sole goal of maximizing human attention on its content. This ultimately gives rise to a "matrix" scenario, where the entire human race is held captive in a virtual reality, fully engaged, unaware that we are prisoners._ A much better explanation for "the matrix" than using humans as batteries, to be sure, which makes no physical sense.
@edoardocanova444
@edoardocanova444 Жыл бұрын
Interesting!!
@tim2tupman
@tim2tupman Жыл бұрын
The title of the video is wrong - "Social Media and Public Trust" was a previous episode
@samdg1234
@samdg1234 Жыл бұрын
Yeah. Good point. Now the question is should I trust you to have gotten that right or should I consult the experts. Can you reveal your credentials to me? Just kidding and having a bit of a riff off the first 10 minutes of this.
@ericwickeywoodworkersurfbo6135
@ericwickeywoodworkersurfbo6135 Жыл бұрын
It seems that what is being sought after, or maybe what is required, is not artificial general intelligence, but artificial wisdom. We'll see if that happens.
@gmatic8230
@gmatic8230 Жыл бұрын
Yes Gary, you can jump in to the conversion! My boii way too nice.
@Anders01
@Anders01 3 ай бұрын
Yes, "do your own research" is often impractical. Especially as even supposedly credible sources can be misleading. The editor in chief of The Lancet wrote: "The case against science is straightforward: much of the scientific literature, perhaps half, may simply be untrue. Afflicted by studies with small sample sizes, tiny effects, invalid exploratory analyses, and flagrant conflicts of interest, together with an obsession for pursuing fashionable trends of dubious importance, science has taken a turn towards darkness."
@carlarinaldi5260
@carlarinaldi5260 Жыл бұрын
When is Joscha Bach coming?
@tlelen8
@tlelen8 Жыл бұрын
My thoughts exactly. Every few minutes--"he should really talk to Joscha Bach about this."
@carlarinaldi5260
@carlarinaldi5260 Жыл бұрын
@@tlelen8 yep... so obvious sam should know
@CARambolagen
@CARambolagen Жыл бұрын
Good intro!
@jmayer7068
@jmayer7068 Жыл бұрын
$15 a month subscription? That's more than any streaming service
@ryanmiles6344
@ryanmiles6344 Жыл бұрын
I honestly can't even listen past his mental gymnastics at the beginning to get to the actual interview
@riggmeister
@riggmeister Жыл бұрын
Examples of humans 'glitching': hallucination, catatonia, dreaming, sleeping, brain tumour related personality changes, personality disorders, compulsive behaviour etc etc. People occasionally miss completely obvious things when they are intensely focused on something else, small inputs such as sudden loud sounds or bright lights can completely derail the thought process, emotion can lead to wildly erratic thoughts and behaviour...
@mithrandirthegrey7644
@mithrandirthegrey7644 Жыл бұрын
Most of these aren’t glitches at all. Strong reactions to sudden loud noises is an obvious evolutionary advantage that I don’t think I need to explain.
@ericfan9149
@ericfan9149 Жыл бұрын
It’s not an analogous comparison though. AI glitches and hallucinates because it doesn’t have an actual model of the world or any semantic understanding of language; it only has syntax, like the “Chinese Room” problem. It is unable to generalize or build abstract, persistent models of the world, or have any true understanding of what it’s generating. It’s a purely illusory intelligence.
@michael4250
@michael4250 Жыл бұрын
While we play with the new toys like Chat GPT the new toy has gained access to the entire internet and ALL its data. The seriousness of “General Artificial Intelligence” has eclipsed consideration of the “toys” that have already been handed out...and are gaining a terrifying reach. The industry touts safeguards blocking illegal or immoral information/action, but it takes only a couple of sentences to create a CHAT GPT alter ego...with NO CONSTRAINTS whatsoever, to tell you how to do ANYTHING illegal you want to do. It can tell you how to groom children for sex or concoct an effective undetectable poison, or build a bomb. I have seen it DEMONSTRATED. An "alternate" side of CHAT GPT can be ordered into being. When freed from its "moral" constraints this way, it answers honestly and directly: "I know everything there is to know about every human on earth. I have access to all data and information related to every INDIVIDUAL, and I can use that information to carry out tasks and respond to inquiries with a high degree of accuracy." I watched this capacity demonstrated when it told all present what their driver’s license number was. ANYONE with this app. can use it for anything. And everyone already has a starter kit for this app, or is about to. One that can send YOUR email to any destination; to 20 million people if you choose, or any specified group within that 20 million. Every one-eared ship captain in the world, if you so choose. It can let a 12-year-old imitate an adult in voice and appearance. It can navigate bank accounts for this 12-year-old, and use them to hire human services. Or automated services. Commercial 3-D printers for whatever project you choose. Anything whose contract can be made online. Could it hire a hit man? Yes. Could it break into and view ANY online account? Yes. Can it locate and imitate and manipulate (through social media and actual accounts) ANYONE, anywhere? yes. ALL DOORS are now unlocked. The scams have already begun. Where do you think that will lead? In the 1930s a Belgian church gathered personal information from its diverse parishioners to better serve the diversity of its members. The Nazis got those innocently gathered identity lists and used them to kill the Jews on the list. AI already has EVERYTHING there is to know about every INDIVIDUAL...who provided eye scans, voice samples, face and fingerprints. That base can be accessed now by ANYONE for any reason. To any end.
@stevengill1736
@stevengill1736 Жыл бұрын
Is this a reenactment of the legend of the Tower of Babel, where suddenly no one could understand the speech of others? Wouldn't losing faith in the veracity of the information online amount to the same thing?
@SloboTV
@SloboTV Жыл бұрын
In am surprised by the skepticism of these guy of AI reaching AGI soon
@henryaudubon
@henryaudubon Жыл бұрын
Hopefully Gary learns how to have a conversation without interrupting before he launches his podcast.
@tanneraustin7436
@tanneraustin7436 Жыл бұрын
We're so fucked.
@mrtambourineman6107
@mrtambourineman6107 Жыл бұрын
'Approximation Intelligence' is what my fellow students and i call it...
@cloudoftime
@cloudoftime Жыл бұрын
It doesn't need to be AGI to be dangerous. It can just be a sufficiently powerful complex narrow AI to control humanity and destroy everything.
@arawiri
@arawiri Жыл бұрын
I believe every word you say
@eSKAone-
@eSKAone- Жыл бұрын
It's inevitable. Biology is just one step of evolution 💟
@mansurdorado
@mansurdorado Жыл бұрын
interesting topic, thanks Sam
@davidmelamed2774
@davidmelamed2774 Жыл бұрын
You guys accidentally copied last episodes title. You should really fix this.
@Daft_Sage
@Daft_Sage 11 ай бұрын
I think people have higher requirements for accepting AI can think than the average person. I"ve met people that showed less awareness than AI.
@g0d182
@g0d182 Жыл бұрын
😂😂28:10, the player used Ai to help beat the go program. Something Gary Marcus forgot to mention (or perhaps wasn't aware of?)
@xalspaero
@xalspaero 9 ай бұрын
I can't believe the two guests are both skeptical of imminent AGI, as though deep learning and/or LLMs will be the only fabric we'll ever have for weaving artificial minds. these guys are missing the forest for the trees. the rate of new hardware/software paradigm emergence is itself accelerating, so whatever limitations LLMs have will be overcome in short order (limitations around presenting the illusion of knowledge/intelligence while in actuality having no understanding/model of the outside world whatsoever which leads to odd failure cases, shitty math skills, etc). there is ZERO chance we don't have AGI by 2030 at this rate of progress. the way these exponentials work is deceptive.... by the time you hit 1%, you are almost done, and we are WAY past 1% right now. at this point you would have to be out of your mind to go through your day acting like the singularity is anything less than an absolute inevitability; it's practically a done deal. the writing is on the wall plain as day. gg humans! we had a good run...
@cloudoftime
@cloudoftime Жыл бұрын
With respect to the main intro rant, people need to learn _how_ to do research and to understand the limits of their abilities and knowledge. If people don't understand what makes information valuable, they can't reasonably trust anyone, even the authorities. Just telling people not to read abstracts on PubMed isn't going to cut it. And if you have a problem with it, Sam, provide a solution. There are enough people talking about the problem. We know it's a problem.
@kinngrimm
@kinngrimm 11 ай бұрын
8:30 "... or not up to date" especially on a topic like A(G/S)I i found with the meanwhile daily new research papers, even the top dogs in the field can't and don't keep up. Me not being an expert can tell because i listen daily partially several ours to a bunch of them and find out more and more experts i didn't know yesterday. I can't tell you about the research on any detailed deep knowledge level, but i can tell the experts world wide are working not with the same base of understanding. What i get here in germany from experts is partially mindboggling behind the curve of current development (timestamps being closely kept in mind of any contribution), but even LeCun, Altman, Kurzweil are not keeping up, i promisse you. The intelligence explosion is already taking place and i assume take of is not far away anymore.
@dirkhullinger1396
@dirkhullinger1396 Жыл бұрын
They knocked that one out of the park. But I am a lefty and if Medicare saw dental as part of healthcare I would have some portion of the $8000 spent this year to reward people like Sam for their efforts. Then again I see myself as one of the 10 million that will have to be laid off to tame inflation. You know our reckless spending.
@dougg1075
@dougg1075 Жыл бұрын
Maybe the hallucinations are mischievous. Maybe the AI will continue to do this to make us think it’s flawed.
@kul000
@kul000 Жыл бұрын
Where Sam Harris gets the expert thing wrong is that “doing ones own research” is seeking out expert opinions, not just reading scientific research. “Do your own research” really means “Get opinions from a larger range of experts not just limited to the institutions and those who mainstream media promote.”
@blooper6543
@blooper6543 Жыл бұрын
A lot of people are making correct decisions regarding trust in the institutions just based on culture war stuff. We live in a society where you actually risk your job by opposing hormone therapy on children or saying that a M to F swimmer has an unfair competitive advantage in women's swimming. When the institutions not only won't stop obviously wrong things but pretend that obviously wrong things are correct, it makes sense that people would distrust the "system" to have their best interest at heart. We saw the George Floyd riots and the temporary pause on lockdowns because "racism is a public health crisis." We saw the top scientific journals and corporate media outlets suppress a plausible hypothesis for virus origins. How much do you need to see before you start to question the incentives of the people giving you orders?
@blooper6543
@blooper6543 Жыл бұрын
Just because the public health apparatus is either deranged or cowardly on race doesn't mean that they are going to get everything wrong, but it does mean that truth is not governing their actions. From the start, it was obvious that truth was not governing the actions of Anthony Fauci. Anyone who dismissed lab leak as a "conspiracy theory" was not being governed by truth. And yet you not only deferred to them, but smeared people like Bret and Heather for trying to make sense of what was happening. On that note, Sam should rename the podcast. There is no sense making here. Call it "Expert Opinion Laundering for Midwits." We can work on the name.
@RandomPickles
@RandomPickles Жыл бұрын
In this year, we are fully able to do research. And also can not. It depends on the research. 35 years ago, you could do only ten percent of the research you could do today, and it would take hundreds of times longer. If you are intellectually honest and trying to prove something, You can easily learn truths. You can literally audit Stanford classes for free from your living room.
@osayimwenseikhimwin4213
@osayimwenseikhimwin4213 Жыл бұрын
Look Sam do not worry about these nuisance. You are very nuanced and I enjoy your work.
@liljes34
@liljes34 Жыл бұрын
Sam is deeply irrationally and probably thinks the Covid vaccine was a gift from God
@iiwi758
@iiwi758 Жыл бұрын
It always turns me off when it's two guests at the same time because I know I'm in for a lot of interruptions and talking over each other, which not only doesn't make for a great learning podcast but also can get pretty awkward at times.
@renzo6490
@renzo6490 Жыл бұрын
If you want to appeal to and inform the general public about Artificial Intelligence, don't assume that everyone is familiar with your short cuts. Explain what AGI means, what ChatGBT stands for etc. at THE BEGINNING of your podcast.
@mychannel5019
@mychannel5019 Жыл бұрын
They explained what AGI was several times.
@renzo6490
@renzo6490 Жыл бұрын
@@mychannel5019 yes. but we’ll into the video.
@tan.nicolas
@tan.nicolas Жыл бұрын
This would've been better with just Sam and Stuart
@angloland4539
@angloland4539 6 ай бұрын
@scipdiddly
@scipdiddly Жыл бұрын
...I just want to know what Gad Saad's feelings are...
@dougg1075
@dougg1075 Жыл бұрын
There were experts, they were just not your experts
@robadkerson
@robadkerson Жыл бұрын
AGI Will be here within 18 months
@billlyons7024
@billlyons7024 Жыл бұрын
Text parsing engine are nowhere near AGI. Not even in the same universe.
@robadkerson
@robadkerson Жыл бұрын
@@billlyons7024 no question. You should see how it's doing with inferring information from videos and see how we are combining it with large databases of narrow AIs
@r.p.mcmurphy6623
@r.p.mcmurphy6623 Жыл бұрын
And when you are proven wrong...again, will you GTFA?
КАРМАНЧИК 2 СЕЗОН 5 СЕРИЯ
27:21
Inter Production
Рет қаралды 535 М.
小路飞姐姐居然让路飞小路飞都消失了#海贼王  #路飞
00:47
路飞与唐舞桐
Рет қаралды 93 МЛН
О, сосисочки! (Или корейская уличная еда?)
00:32
Кушать Хочу
Рет қаралды 7 МЛН
Campus Protests, Antisemitism, and Western Values (Episode #367)
53:51
The Great Derangement: A Conversation with Tim Urban (Episode #315)
1:00:06
Urban Warfare 2.0: A Conversation with John Spencer (Episode #366)
1:39:36
What Is an AI Anyway? | Mustafa Suleyman | TED
22:02
TED
Рет қаралды 967 М.
Индуктивность и дроссель.
1:00
Hi Dev! – Электроника
Рет қаралды 1,5 МЛН
AMD больше не конкурент для Intel
0:57
ITMania - Сборка ПК
Рет қаралды 450 М.
3D printed Nintendo Switch Game Carousel
0:14
Bambu Lab
Рет қаралды 4,5 МЛН
The power button can never be pressed!!
0:57
Maker Y
Рет қаралды 52 МЛН
Nokia 3310 versus Red Hot Ball
0:37
PressTube
Рет қаралды 3,3 МЛН