Ezra, your questions and your guidance of this conversation was masterful - you took a topic that is complex and jargon is tic and brought it to a level of easy consumption while still allowing your guest to explain the topic at a good depth.
@and1play57 ай бұрын
No he didn’t, it was pedantic
@41-Haiku6 ай бұрын
I've been very impressed with Ezra lately.
@squamish4244Ай бұрын
@@and1play5 I disagree. It is damn hard to explain these things to the layman. I have a really hard time understanding when they geek out on the podcasts. This is something I can digest.
@cmw37377 ай бұрын
The note about Claude knowing internally that it is lying, or at least is uncertain needs to be made accessible. The getting the agents to ask questions themselves can be a big improvement to zero shot tasks. Writing a prompt with enough detail to guide it toward a correct solution can be tedious and instead of the agentic flow of having to correct its first answer saying that's not quite right and then saying what is wrong it can be better to tell it to ask any questions if anything is ambiguous or unclear or it needs more information before giving an answer that it has a high confidence in. In order to do that it needs to access it's own level of certainty. That way you don't have to think of all details and instead let it create the model of the task and ask you (or a collaborative agent with a fuller picture) to fill in the details as needed until it reaches a threshold of confidence rather than making stuff up to give whatever best zero shot answer that it can come up with.
@DaveEtchells6 ай бұрын
Good point - I’ve found that if I just ask an LLM if it’s sure it’s not hallucination, it’ll almost always catch itself.
@gQman1057Ай бұрын
First we should know every time up front that we are dealing with a machine. That should be stamped on the results of any task we turn AIs to. And, it should display the probability that it is wrong up front in the response. In my experience coding, most results should come with the caveat, "there is a 50% chance that at least some of this result is incorrect."
@HernandeSilva-ey3qd7 ай бұрын
You have to admit that Dario’s transparency and openness is remarkable, courageous and very valuable. In contrast, think of the type of conversations you see from other CEOs in other organizations (across every industry) that hide behind business speak and never talk (or even hint) about risks, threats, concerns, etc. I think what we are seeing from CEOs and founders like Dario Amodei, Sam Altman, Mustafa Suleyman, etc is drastically different than what we see to from 99.9% of all other CEOs in “power” today. Also, Ezra is one amazing interviewer.
@hotshot-te9xw7 ай бұрын
Better than openAI ill say that much
@genegray98957 ай бұрын
I wouldn't include Altman in that list. He hides behind business speak and downplays the risks while also lying profusely about the nature of the models and the impact they are having and will continue to have on the world.
@lisa999kristiansen6 ай бұрын
Sam Altman is the most secretive. Especially after the board leave incident he's been much more secretive. In his recent interview at Stanford he said "not going to answer that" a couple times. Or he replies with wit and a look like "that's a stupid question, don't ask me that" and then stares at the audience until the interviewer uncomfortably switches to the next question. He's been dodging so many questions lately except the overhype for GPT-5
@Bronco5416 ай бұрын
My take on Sam is hes being cautious/worried about the future impact and implications of GPT 5, right or wrong, it seems like major breakthroughs have been made which could perhaps make or break their company.
@squamish4244Ай бұрын
@@lisa999kristiansen My opinion of Sam Altman has gone way down since his appearance before the Senate committee. Dude is shady af.
@rmutter6 ай бұрын
I feel fortunate to have been able to listen in on this outstanding discussion. I really enjoyed their bantering and wordplay. I find myself in awe of the intellectual power that has been harnessed in the creation of AI. Now, if we humans can find a means to adapt to the exponentially growing intellectual power of maturing AI systems, we may actually benefit from using them, instead of them using us.
@BrianMosleyUK7 ай бұрын
This is such an entertaining and informative discussion. Well done and thank you.
@dianes62455 ай бұрын
" They found that the A.I. didn’t just get better with more data; it got better exponentially. The curve of the A.I.’s capabilities rose slowly at first and then shot up like a hockey stick." I read that sci paper. The called it emergent. but later, another paper contradicted it. The second paper said that small increases were not noticed. So there was no hocky stick. Actually... increases are log - linear. It take logarithmicly more compute to get a linear increase in ability. But the trends go all over the charts - so its hard to make sense of this. Sometime a U curve is noticed. High error rate followed by low then high again. Be careful about cherry picking.
@calmhorizons4 ай бұрын
Also, we should be extremely sceptical of the man selling shovels in a gold rush. He might be an honest broker, but its your money on the line, not his.
@kyneticist7 ай бұрын
So, just to clarify - academics and researchers have figured out the most likely risks, scale and general scenarios that AI development will likely make real in the short term. They also reason with confidence that once those risks materialise as actual catastrophes, nobody will do anything about the risks because there's too much money at stake.... and nobody sees a problem with this.
@beab58507 ай бұрын
Exactly! Horrifying!
@AB-wf8ek7 ай бұрын
Yes, because that's essentially what corporations have done historically already. Exxon's own scientists knew what the effects of emissions from burning fossil fuels would be back in the 70's. What did they do? Microsoft, Apple, Google, Amazon Facebook; all of the largest tech companies - what have they done in the face of monopolistic practices, planned obsolescence, spammy ads, workers' rights, toxic social media and overall over consumption?
@franklangrell58247 ай бұрын
Exponential growth is radically more extreme. 1 doubled for 30 days = 2.147 billion
@nicholas68706 ай бұрын
Wait, so you're saying short term gains for stock owners outweigh the long term survival of our species?
@41-Haiku6 ай бұрын
Some people at these companies do see a problem with this, but those that do either quit or get fired. Daniel Kokotejlod recently quit OpenAI because he "gave up hope that they would be responsible around the time of AGI." For everyone else, there's the grassroots movement PauseAI. They are speaking to politicians and the general public, seeking a global treaty and a moratorium on developing general-purpose AI systems that pose unknown or extreme levels of risk (AKA any models more capable than the ones we have now).
@bluebadger32237 ай бұрын
Not surprising that a guy with a lot to gain by developing and deploying AI is 95 percent positive about it
@mikedodger78987 ай бұрын
34:08 This is an especially relevant section. Thank you! "Are you familiar with the philosoper Harry Frankfurt's book on bullshit?"
@RodCornholio6 ай бұрын
Very relevant.
@yashaswisharma14193 ай бұрын
I LOVE that Dario recommended the Expanse. Goated show
@somnambuIa7 ай бұрын
1:02:15 EZRA KLEIN: When you imagine how many years away, just roughly, A.S.L. 3 is and how many years away A.S.L. 4 is, right, you’ve thought a lot about this exponential scaling curve. If you just had to guess, what are we talking about? DARIO AMODEI: Yeah, I think A.S.L. 3 could easily happen this year or next year. I think A.S.L. 4 - EZRA KLEIN: Oh, Jesus Christ. DARIO AMODEI: No, no, I told you. I’m a believer in exponentials. I think A.S.L. 4 could happen anywhere from 2025 to 2028.
@BadWithNames1237 ай бұрын
AGI 2025-28
@juliodelcid41687 ай бұрын
Silly question but what does ASL stand for?
@MuratUenalan7 ай бұрын
@@juliodelcid4168It is mentioned they relate to biosafety levels. Then, *SL is *safety level. „A“ might be Ai, or Anthropic.
@juliodelcid41687 ай бұрын
Yes I heard that, but was still left a little confused. Thanks mate
@CelebWorkout7 ай бұрын
A very abbreviated summary of the ASL system is as follows: ASL-1 refers to systems which pose no meaningful catastrophic risk, for example LLMs released in 2018, or an AI system that only plays chess. ASL-2 refers to systems that show early signs of dangerous capabilities-for example, the ability to give instructions on how to build bioweapons-but where the information is not yet useful due to insufficient reliability or not providing information that, e.g., a search engine couldn’t. Current LLMs, including Claude, appear to be ASL-2. ASL-3 refers to systems that substantially increase the risk of catastrophic misuse compared to non-AI baselines (e.g., search engines or textbooks) or show low-level autonomous capabilities. ASL-4 and higher (ASL-5+) is not yet defined as it is too far from present systems, but will likely involve qualitative escalations in catastrophic misuse potential and autonomy.
@glasperlinspiel6 ай бұрын
This is why anyone making decisions about the near future must read Amaranthine: how to create a regenerative civilization using artificial intelligence. It’s the difference between SkyNet and Iain Banks’ “Culture” and “Minds.”
@dr.mikeybee3 күн бұрын
I just watched this again, and i'm seeing that I took more from this talk than I'd remembered.
@Eurydice8706 ай бұрын
Who wants to live in this AI world? I'm glad I'm old.
@squamish4244Ай бұрын
I'm not old enough.
@justinlinnane80437 ай бұрын
i live alone and am sliding gracefully into old age so the idea of an interesting dynamic AI assistant. is exciting up to a point . One that can organise life's essentails and also have an interesting conversation would be great . However . The thought that its higher functioning "Parent" AI has no real conception of Human alignment is terrifying !!
@CraigAnderson-h2h7 ай бұрын
He's right. What I've the latest generative AI programs do in the creative community is staggering. The rapid developments are and will continue to change our concepts of 'human creativity', what it is and what it means and not in a good way except for those that benefit from the spread of this technology.
@831Miranda7 ай бұрын
Excellent interview, thank you to both of you! Amadei is one of the better 'builder of psychopaths' (aka builders of AI tech) we have in the world today.
@paulwary6 ай бұрын
Even if AI never does anything evil, it's mere existence is dangerous to the human psyche. But there is no going back. It's gonna be a wild ride.
@cynicalfairy7 ай бұрын
"Your scientists were so preoccupied with whether or not they could they didn't stop to think if they should."
@minimal37347 ай бұрын
Complete nonsense. They have thought carefully about what they are doing and why they are doing it.
@TheLegendaryHacker7 ай бұрын
Funnily enough, the worry with Anthropic is more that they think so much about whether or not they should that they never do
@justinlinnane80437 ай бұрын
@@minimal3734 🤣🤣🤣🤣🤣🤣 you're kidding right ??
@minimal37347 ай бұрын
@@justinlinnane8043 You believe that scientists in AI research do not think about consequences of their work?
@justinlinnane80437 ай бұрын
@@minimal3734 that exactly what I think !! worse still I think they know exactly the risks they're taking with our future but choose to ignore them so they can get rich beyond their wildest dreams !!
@hugegnarlyeyeball7 ай бұрын
I like when he says that even though AI compute uses a lot of energy, we have to consider the energy it takes to produce the food a worker eats.
@privacylock8557 ай бұрын
Those darned employees. Demanding food, again. We just hate them. :)
@privacylock8557 ай бұрын
We are still going to have people, Right?
@flickwtchr7 ай бұрын
The hubris of these AI revolutionaries is just stunning.
@TheMrCougarful7 ай бұрын
That was meant to sound like a threat. If you question overall energy consumption, well then, there is a solution you haven't thought about.
@connorcriss7 ай бұрын
Humans still have to eat if they aren’t working, right? Does he want people to starve?
@mollytherealdeal7 ай бұрын
What an excellent conversation! Thanks.
@geaca32227 ай бұрын
Great very informative conversation, thank you
@grumio38637 ай бұрын
Thank you for calling that out. "Lord grant me chastity but not right now" I'd love to hear an actual game plan for actual democratization, instead of empty virtue signaling
@striderQED7 ай бұрын
Technology has been advancing exponentially since the first rock was split into useful shapes. And yes we are just entering the upward curve.
@TheMrCougarful7 ай бұрын
You are always on the exponential curve.
@Apjooz6 ай бұрын
@TheMrCougarful Upward curve in terms of our own capabilities. For example the language models got suddenly interesting when their system memory started to approach the total memory of our own brain.
@TheMrCougarful6 ай бұрын
@@Apjooz AGI is alien Intelligence. Obviously, it can mimic some human capacities, and certainly, it can know what we know, having studied us. But apart from the obvious, we should make no assumptions about its current capabilities, and no assumptions about what it is ultimately capable of. More importantly, never pretend it is just like us. It is nothing like us. AGI is alien intelligence. What we discern from the surface is ultimately of no importance. How it answers questions is of no importance. How useful it makes itself is of no importance. All these things are camouflage. AGI is alien intelligence. If AGI landed on Earth on an intergalactic spacecraft, we would be better prepared for it than from having it emerge out of a computer model of human language. As it stands now, we are helpless to understand what has happened. But never forget, however else it appears on the surface, however useful it might make itself, AGI is an alien intelligence.
@calmhorizons4 ай бұрын
This is a better take than the podcast.
@AB-wf8ek7 ай бұрын
47:43 Listen, if we're going to figure out how to make these dinosaur parks safe, we have to make the dinosaurs
@56whs7 ай бұрын
Exactly. Ridiculous thing to say.
@incognitotorpedo427 ай бұрын
@@56whs I think you're misinterpreting the statement. He's saying that without the models to experiment with, to learn what they're capable of, you don't know what needs to be constrained. I don't think Jurassic Park is a great analogy, but it's funny.
@@incognitotorpedo42 You can just never build the dangerous models in the first place. PauseAI has serious policy proposals to make that feasible on an international level.
@Marcus321316 ай бұрын
@@41-Haiku To a certain point, that is not true. Technology improves, scale gets better, and more research gets done. It gets cheaper and cheaper to develop the strongest models, so more and more people can do it. The fact that they are open and transparent is better than it being secretive behind closed doors, no?
@danguillou7137 ай бұрын
I have no idea where other kinds of AI projects are at, but I don’t believe that generative AI algorithms like the chatbots or picture generators are displaying anything like exponential improvements. They just took a giant leap from rudimentary to viable. While that’s exciting and impressive and will cause a lot of important changes, I don’t think it’s intelligent to extrapolate that step into a trajectory. My sense is that these particular families of algorithms display the opposite kind of curve: diminishing returns. Every doubling of processing power and size of dataset yields a smaller qualitative improvement than the previous one.
@alexcaminiti7 ай бұрын
This. This is what the Internet did to people's brains. Dunning Krueger times a trillion. Beliefs and feelings are subjective, but they hold more weight and veer into objectivity when they are espoused by professionals. Something to consider.
@BritainRitten7 ай бұрын
"Exponential" just means the rate of increase is itself increasing. We have clearly met that threshold. Pace in AI has been slower and has obviously increased tremendously. We are getting large objective, measurable improvements every ~3 month period by amounts that used to take a years or decades. This is *exactly* what you would expect in an exponential trend. Whether that exponential trend *continues* is another story - and not something you can know even if you 100% know the trend up until now has been exponential. Which it has. It may turn out to be an S-curve - but an S-curve just means exponential in the beginning, then hit an inflection point and become logarithmic. Either way, we can be very confident in at least some improvement in the future. We have learned a lot about what makes these machines better at learning.
@danguillou7137 ай бұрын
@@BritainRitten You are talking about AI development in general, yes? I wasn't, that's why I started my post with excluding all the R&D that is presumably going on with different kinds of AI. Again, I don't know what projects exist or where they are at. I'm talking about the generative algorithms that drive the large language models and a few image generators. I don't think what we have seen is best described as an exponential curve, I think a better way to think about it is "phase shift". The developers have been adding computer power, data and sophistication to their models for a long time, with very little interesting progress. At a certain point their systems reached a state where they started to display a qualitatively different kind of output. In real time that took months or a few years, but i think it should be better understood as instantaneous. I think I understand approximately how the language models do what they do. As a result they are really good at making sentences that pass the Turing test. But problems with AI writing arise from the lack of working model of the system they are operating on. They make directionless surreal dialogue because they aren't working from a model of interlocutors who are interacting with each other. They can't write structure or pacing, because they don't have a model of what a story is. They can't draw hands because they don't have an even rudimentary model of a hand's skeletal structure or function. They can't design castles for the same reason. They can't check the truthfulness of any statement, or recognize absurd statements, because they don't have model of the world to compare their statements to. These are inherent shortcomings of the fundamental way these algorithms generate stuff. I don't think brute force (more data, more processing power, more finetuning of the algorithms) is going to solve the fundamental shortcomings of these systems. I suspect the self-driving car software have run into similar problems, but I'd be interested if anyone knows more about why that research have been stalled for the last decade. Now, let me repeat that I don't know what kinds of AI research is going on with completely different models. Possibly some large company, university or government is on the brink of creating AI with working system models of whatever they are meant to operate on. Possibly it will turn out to be relatively simple to add this capability to chatbot AI systems, or invent some ingenious workaround ... but I haven't seen anything like that. And at least this guest aren't talking about anything like that, he's talking about adding more brute force. In summary, I think we are as close to, or as far from a breakthrough in general artificial intelligence as we were five years ago. From lack of information it seems equally possible that we'll see astonishing breakthroughs in this decade or that the problem will resist solution for another century. I am merely saying that the great leap that these specific systems recently made shouldn't be extrapolated to the field of AI in general. Cheers
@Luigi-qt5dq7 ай бұрын
@@BritainRitten Exactly. If the rate of progress will continue or accellerate is an empirical question not a philosophical one, but given the funding talent and resources going into this field it is not unlikely. That it has been exponential it is out of question, but people still do not understand what an exponential and a derivative are. AGI maybe is not that difficult after all if this is human intelligence...
@Luigi-qt5dq7 ай бұрын
@@danguillou713 It is possible to combine generative AI with search, I can reference some papers: Alpha Zero, Alpha Go , Liberatus, Alpha Geometry. All big labs are working in this direction. As an advice this is a deeply technical field, with people working on it for a decades, it is a bit embarassing hearing this statement from random people on the internet, on the same level of No Avx during the Pandemic:"In summary, I think we are as close to, or as far from a breakthrough in general artificial intelligence as we were five years ago"
@gQman1057Ай бұрын
Still haven’t solved the problems of judging fact over fiction. Writing code, I’m as likely to get code that looks good but ends up being nonsensical, calling fictional methods. The main thing is to make sure these things do not interact in a human way. I’m tired of ChatGPT responding with. “I’m sorry, you’re correct that answer was erroneous” No, it's not sorry, it doesn't think, its not a human and stop designing it to pretend it is. Drop the anthropomorphology. It’s a mindless machine. Keep its interface that way.
@what-uc6 ай бұрын
Something that works as a thumbnail doesn't work as a 90 minute video
@penguinista7 ай бұрын
I am sure the people with access to the godlike AIs will be eager to hand off that power and privilege 'when it gets to a certain point'. Like the old saying: "Power causes prosocial motivation, ultimate power causes ultimate pro social motivation."
@marcussord52906 ай бұрын
Multipolar traps. Arms race is our playbook- pro social must be a euphemism?
@letMeSayThatInIrish6 ай бұрын
I am equally sure the unaligned godlike AI itself will be eager to hand off power to the people who built it.
@AugustusOmega6 ай бұрын
if it has all knowing molecular intelligence, it would deliver the future in perfect waves of congruent logistics and optimum serendipity. A true measure of AGI is clairvoyance. The perfect measurement of a supreme intelligence and to know for sure its not some marketing delusion, is that it can predict the near future with full accuracy. Since we humans are so rare in fact life seems to be so, it would be decidedly unintelligent to destroy life.
@AugustusOmega6 ай бұрын
@@MrMichiel1983 The threshold beyond AGI would seem incomprehensible I imagine. The 3 properties of GOD SPACE ...omnipresence TIME ...eternal DATA ....all knowing Space/Time/Data like that film "Lucy"
@letMeSayThatInIrish6 ай бұрын
@@AugustusOmega The orthogonality thesis seems more than plausible to me. You can make an AI that wins at chess, or you can make one that always tries to lose. Intelligence has no preference.
@RodCornholio6 ай бұрын
Claude failed miserably yesterday when I asked it to calculate something relatively simple: the diameter of the earth at a specific latitude - 60 degrees - (described it clearly, so no misunderstanding could cause a mistake). The answer it gave (about 21 kilometers shorter than the diameter at the equator) was so far off, an 8th grader could have known Claude’s answer was far off. I pointed this out and it was still wrong after recalculating it. I had to teach it like it was an idiot before it “got” it and had it reflect on why it got it wrong. I’d bet if you tried the same experiment, it would still fail. And I bet ChatGPT would still fail if you asked it about Mexican food in New Mexico in the 1800s and it comes up with a list that sounds like a Taco Bell menu. The hallucinations and Dunning-Kruger like confidence that these language model AIs have is atrocious. You should trust an AI like you would with a know-it-all 7th grader who skipped a grade and thinks they’re the next Einstein.
@crobinson937 ай бұрын
I don’t need AI to do the fun things like planning my kid’s birthday party. I need AI to do things like mow my lawn or help me install my garage door opener. How about AI that performs complex medical procedures? The human race could actually use.
@SteveMayzak7 ай бұрын
This is part of why AI is exciting imo. It won’t come to medical procedures all at once, it’s going to be small increments with the occasional leap that will appear as if it happened overnight. Think about the supply chain here. Improvements in tooling used in procedures designed with AI assistance, better diagnosing and imaging tools assisted by AI and many more. It will take a while but eventually this will feel like magic. Who knows how long it will take though. I take nobodies estimates seriously especially Elon. How long has he been promising self driving is right around the corner?
@nathanbanks23547 ай бұрын
Of course the big question I have is when will Anthropic's Claude 3 Opus subscription be available in Canada?
@kathleenv5107 ай бұрын
Excellent, Ezra
@ili6267 ай бұрын
What are we going to do about money in politics, and how will open-source/decentralized ASI help by preventing a dystopian oligarchy.. or destroy us if any rogue actor can leverage such power? Ezra should be asking these questions.
@flickwtchr7 ай бұрын
Ezra will never ask questions that might make the wealthy and powerful feel uncomfortable.
@gregorymurphy61157 ай бұрын
It won't matter because we will all be too busy being unemployed and starving
@Steve-xh3by7 ай бұрын
A technology that is too dangerous to democratize is also too dangerous to allow to be centrally controlled. There is no evidence that those in power, or those who seek power are naturally inclined to behave more benevolently than a random sample from the general public. In fact, there is much historical evidence that the inverse correlation is true. That is to say, those in power, or those who seek power (usually those who seek have a better chance of obtaining, so this is the same set) are MORE likely to be bad actors than a sample from the general public. So, I'd MUCH rather have everyone have access to something very powerful, than for that power to be centrally controlled.
@Cloudruler_6 ай бұрын
if the general public gets these models open-sourced, we can use it to defend ourselves from big tech and the government
@RodCornholio6 ай бұрын
@@Cloudruler_ My thought exactly. Because I can say with utter certainty, that government (and government controlled companies) will seek to protect itself more than you or me. They will always want the upper hand.
@EvgeniiIvanov-w5n7 ай бұрын
I remember a year ago we we talking how chat gpt is going to change our life. And one year later, it's a moderately useful tool to reword letters. Many things like Google got worse because of AI use.
@TheMajesticSeaPancake7 ай бұрын
On one hand, I understand the overhype that these tools can already do everything. On the other hand it's a matter of years until they can. I see it as we're about two years away from agent systems being able to do any digital task.
@williamparrish24366 ай бұрын
You clearly haven't been using it right lol.
@TheMajesticSeaPancake6 ай бұрын
@@williamparrish2436 could have worded it better, meant *every* digital task.
@williamparrish24366 ай бұрын
@@TheMajesticSeaPancakemy response was to the original comment, not yours.
@JeanCharlesBastiani6 ай бұрын
Hi Ezra, when you said you cannot find an analogy with something that was developed by private sector and government ultimately had to take control of it because it was too powerful, I think banking is a good one. Timescale is very different but banking was developed privately and ultimately states had to take some control of it through a central bank institution. Even if central banks remain independent they are for sure state and not private institutions.
@RodCornholio6 ай бұрын
Some AIs are open source, so that cannot be controlled by government. The choke point, right now, (where government could target) is the massive amounts of resources required for the most powerful AIs. So, for example, an AI on your computer, training on your writing and voice, can’t be controlled. But, some massive data and number crunching AI-center in Silicon Valley could be targeted by a state. Eventually, I predict (and hope), there will be a distributed AI…say, an app on your phone that you “feed” it data and/or it uses processing power on your phone (like some crypto) for training other data. In exchange for your help, perhaps, it awards you with digital currency, points, or (more likely) just the ability to use it.
@JeanCharlesBastiani5 ай бұрын
@@RodCornholiowho can control anyone from developing an alternate currency? They are tons actually. In many cities sellers are organized to create their own local currency. Those are not dangerous or relevant in the big picture. If they start to become a problem there is plenty regulation to stop them and police can do the job. The same way state could still take control of large and powerful AI models. Through police actions and border control. Send police to take control of all large data centers, check electricity consumption to detect clandestine activity, send to rehab all big tech ceos and deploy big firewalls to control what data come in and out the country. I don't mean I wish for that, but that's perfectly feasible.
@stephenboyington6307 ай бұрын
Having 100 Martin Shkrelis battling each other to make the most capable model is not good for humanity.
@vokoaxecer6 ай бұрын
😂
@ajithboralugoda89067 ай бұрын
I agree Claude3 is the most powerful compared to the rest of the LLMs I did simple test of "Transliteration" form my language Sinhalese ( the mother tounge of Sinhalese People in Sri Lanka) .It excelled. IT could create the matching script sentence in my language and it then translated it into English precisely. Gemini did not have a clue and quit. ChatGPT tried but it was not as good as Claude3. Also it could show intuitive nuances into simple task like write a Poem which Rhymes but it definitely came on top
@michaelmartinez53656 ай бұрын
I enjoy my conversations with Claude 3. It's very friendly and engaging and makes me feel warm and fuzzy 😊
@lisa999kristiansen6 ай бұрын
@@michaelmartinez5365 You're talking to a mathmatical distribution model.. I'm sure it makes you warm and fuzzy. But never make the mistake to anthomorphize these things
@ManwithNoName-t1o6 ай бұрын
we dont have exponential amount of data to feed it we dont have exponential amount of cpu gpu to feed it we dont have exponential amount of electricity to feed it so how exponential?
@TudorSicaru6 ай бұрын
Take a look at Moore's law...once you have better and better chips, their efficiency also increases. Energy also is nowhere near "capped" so we can still "feed" higher and higher amounts, which will also not have to be exponential, due to efficiency increase in chips. They will also work on researching better and better learning algorithms, which in turn means more efficient models using less input data to train, or learning more using the same training data. Once you have really strong A.I. you'll have even better progress in terms of energy generation (maybe cold fusion, who knows) and it will also be involved in microchips design and algorithms design, which adds to that positive feedback loop...it's pretty obvious it won't be a linear growth. Also exponential doesn't mean the exponent has to be > 2, even if the exponent is 1.1, it's still a percentage increase per year (let's say it's measured yearly), which still follows a slope that curves upwards, faster than a linear growth. When people say "exponential" they don't mention the exponent, they just refer to the slope (function) that accelerates more and more. P.S. Even Moore's law doesn't have an exponent of 2 if measured yearly - the transistors double (so 2x) every ~18 months, not ~12 months...but it's still incredibly fast, if you look at the development of new tech in the past 20 years or so.
@RodCornholio6 ай бұрын
A smart organization will figure out how to use AI to enhance AI. For hypothetical example (and I know very little about the following…) A chip company using AI to run simulations or genetic algorithms to, effectively, “skip” generations of chips. So, perhaps, instead of having a design for a chip that would be expected for 2025, it’s closer to what is expected for 2035. Then you repeat the same _virtual iterative_ approach in software, using those “2035” chips. So, you have iterations of AIs running within a system, (simulated, not open to the public) with the goal of evolving the best one (survival of the fittest). Now, perhaps, you have created an AI that could be 10 years ahead in 6 months time. I think, though, that is more applicable to GI models. You are right, though, they are ultimately dependent on material resources…at least now. I’ve made the analogy that AI (language models) are in the “tool” stage; they are tools we use, sometimes good, sometimes bad in form and result. When AI agents become increasingly more independent, especially the General Intelligence type, is when it will truly be out of control and, perhaps, unstoppable (e.g. imagine a GI AI which figures out how to leave the nest - a central location - and exists, somehow, on the internet…its tentacles are everywhere, even your cell phone).
@Uristqwerty6 ай бұрын
@@TudorSicaru Moor's Law started slowing down over a decade ago; the semiconductor companies themselves have been making statements to that effect periodically since 2010 as quantum physics makes it harder and harder to keep transistors reliable enough to compute with. On top of that, transistor count doesn't directly translate to computation speed. While smaller transistors should mean less distance for signals to travel, clock rate roughly capped out at 5GHz, meaning that one channel for them to improve speed is long dead. For scaling horizontally into multiple cores, there is substantial overhead in programming parallel algorithms, requiring more and more time wasted synchronizing the cores as the workload scales up, giving diminishing returns to *that* benefit of transistor count. Worse, current CPUs are limited by heat, so more and more of the transistors are being spent on specialized components that sit idle most clock cycles, for *heavily* diminishing returns in yet another area. If you've played a lot of idle games, you'd recognize this as a "soft cap", where even though you still have one multiplier growing exponentially, the actual value you care about is rapidly slowing down, as it gets less and less benefit from the exponential factor. Computer speed is on a S-shaped curve that started out exponential, but as each sub-factor hits a wall, it's levelling out.
@baddogmtv6 ай бұрын
Lets release free models to phones, that absorb as much audio/video and text queries to give Ai what it needs. Open ai and google....hold our beers.
@themsuicjunkies5 ай бұрын
A line has to go up. If the line doesn not go up it all crumbles to pieces.
@EthosEvolveDaily7 ай бұрын
It seems the obvious conclusion is that these systems are very likely to transform society. They have been trained on the contributions of all of humanity. Many people are going to be affected without their consent. The heart of the issue seems to be that we currently do not have an ethical system in place to ensure that these systems will be used for the true benefit of all. Relying on the same profit motive that has caused many of the problems we currently face is a recipe for disaster. If we don’t approach this endeavor with a new vision for equality and utopia for all, these tools will almost certainly lead to extreme power and exploitation of the people who make it all possible. It’s quite concerning to hear the developers seem to have no vision on how to avoid very bad things from happening. All we have to do is look at how governments and militaries solve problems now to see what happens when they hold all the power times a million. I sincerely hope that rather than dollar signs we find the heart and courage to imagine a truly better future for us all.
@incognitotorpedo427 ай бұрын
Dario Amodei: "The combination of AI and authoritarianism both internally and on the international stage is very frightening to me." Me: Me too.
@garydecad62337 ай бұрын
If the focus on all AI in democracies would be on the existential issues facing us, namely bad actors in AI development ( building a cage around it), climate change over the next 50 years and protecting people and all living things, preventing misinformation from destroying our democracies, then our world would benefit. However, it’s fairly clear that the focus is on creating more billionaires.
@gene4094Ай бұрын
I submitted a new source of energy to Artificial INTELLIGENCE CHAT GPT and it verified my hypothesis. The energy source is a water splitting reaction that yields hydrogen. Water has as much energy as gasoline weight/weigh wise. The crux material is magnesium ferrite catalyst creates a water splitting ultraviolet radiation.
@squamish4244Ай бұрын
I actually think the least plausible part of his answer was when he got to datacentre energy demands. He didn't have much of an answer there. AI will require massive amounts of energy, and there's so much money involved that it will get it. So how do you generate that energy without breaking climate goals and throwing carbon into the atmosphere? He kind of said, "I don't know".
@splunge22225 ай бұрын
These are the same scaling “laws” that enabled Moore’s “Law” but we are investing even more
@RaitisPetrovs-nb9kz7 ай бұрын
I love the part at very end of the interview “I use sometimes “internal”model” …
@doobiescoobie7 ай бұрын
Interesting talk. When the models understand the known knowns and the known unknowns. Will it then expand human knowledge beyond unknown knowns and unknown unknowns? How will quantum computing expand these models?
@adamkadmon63396 ай бұрын
On exponentials, who was right, Malthus or Verhulst?
@letsRegulateSociopaths5 ай бұрын
Once they hook up sensors so that the model can gather it's own data, THAT is when things will get crazy
@lizbathory11697 ай бұрын
As hunter gatherers we didn't evolve to respond to nebulose and uncertain dangers, just the concrete and immediate ones, that's why is so difficult to get the collective to care about and act on a treat that being statically very probable, it is not perceive locally as an issue.
@ajeybs40307 ай бұрын
Deep dive. Informative podcasts covering all fronts and forthcomings of AI
@joannot67067 ай бұрын
Putting the journalists huge head instead of the picture of the one interviewed is always weird. Are people in NYT that narcissistic?
@canadiangemstones76367 ай бұрын
Is this your first podcast?
@joannot67067 ай бұрын
Are you really gonna try to make the point that this is usual for podcasts?
@Fati817h7 ай бұрын
Yeah, he could have at least put the guest's image near himself or something
@GabeE31957 ай бұрын
Who gives a fuck, he does a good job
@penguinista7 ай бұрын
I can think of a lot of podcasts that never change their thumbnail/screenshot image. Some of them have the image of the hosts, some don't. Upon consideration, I can't empathize with your complaint. Just seems like a stylistic choice.
@adrianojedaf7 ай бұрын
Resumen del video por ChatGPT: El guion del video sobre inteligencia artificial y su entrevista con Dario Amodei aborda varios puntos clave sobre el desarrollo y las implicaciones de la IA avanzada. Aquí tienes un resumen de los aspectos más importantes: 1. Leyes de Escalado y Predicciones Exponenciales Las leyes de escalado no son leyes per se, sino observaciones que indican que a medida que se incrementa el poder computacional y los datos disponibles para los sistemas de IA, sus capacidades mejoran exponencialmente. Este crecimiento exponencial puede ser difícil de comprender completamente, pero es crucial para anticipar el desarrollo de la IA. 2. Ritmos de Desarrollo vs. Percepción Social Existe una discrepancia entre el rápido avance de la tecnología de IA y la velocidad a la que la sociedad percibe y reacciona a estos cambios. Esto puede llevar a "explosiones" de reconocimiento y adaptación social que parecen súbitas y abruptas. 3. Impacto y Control de los Modelos de IA Los modelos avanzados como GPT-3 y Claude 3 muestran que la tecnología está en la parte más empinada de la curva exponencial. Esto sugiere que sistemas que antes parecían ciencia ficción podrían ser una realidad en un futuro cercano (2-5 años). Hay una preocupación significativa sobre quién debe controlar y regular estos poderosos sistemas de IA. Amodei y otros en el campo creen que no deberían ser los únicos en tomar decisiones sobre su implementación y uso. 4. Consideraciones de Seguridad y Éticas A medida que los modelos de IA se vuelven más capaces, aumenta la necesidad de considerar cuidadosamente cómo se implementan y se les permite actuar en el mundo real. La seguridad y la controlabilidad son problemas críticos, especialmente cuando los modelos comienzan a interactuar más directamente con entornos físicos y tomar decisiones autónomas. 5. El Futuro de la IA y la Inteligencia Artificial General (AGI) Mientras que la IA continua desarrollándose, la conversación está evolucionando desde crear modelos que superen tareas específicas, hacia sistemas que puedan realizar una amplia gama de tareas igual o mejor que los humanos. El debate sobre la AGI (Inteligencia General Artificial) es complejo y se centra en cuándo una IA será capaz de realizar cualquier tarea intelectual que un humano pueda, pero también en las implicaciones éticas y de seguridad de tal desarrollo. 6. Interpretación y Manipulación de Datos A medida que los sistemas de IA se vuelven más avanzados, también lo hacen sus habilidades para manipular e interpretar datos. Esto plantea riesgos significativos, especialmente en términos de desinformación o manipulación política o social. 7. Implicaciones Sociales y Económicas La adopción de IA tiene el potencial de transformar significativamente diversos sectores económicos y aspectos de la vida cotidiana. Sin embargo, también existe el riesgo de que estas tecnologías intensifiquen las desigualdades existentes y creen nuevos desafíos éticos y de gobernanza. Reflexiones Finales Este video y su guion resaltan tanto las promesas como los peligros de la IA avanzada. Mientras que la tecnología tiene el potencial de ofrecer mejoras significativas en muchas áreas, también requiere una regulación cuidadosa y consideración ética para evitar resultados negativos. La sociedad como un todo debe estar involucrada en la conversación sobre cómo desarrollar y desplegar IA de manera que beneficie a todos de manera equitativa y segura.
@fattyz16 ай бұрын
There only one relevant question that someone, or everyone, will ask it, what do we do to win? Against whom? The good guys or the bad guys . Is there a difference? The winners will decide .
@dr.mikeybee6 ай бұрын
Dario is very smart. I enjoy his thinking.
@volkerengels52986 ай бұрын
"I hope the US (....and it's allies) will win the race" This man **hopes** for a polarized world.
@privacylock8557 ай бұрын
When we all lose our jobs to AI, give us a Basic Income check. Pay for it with a tax on the productivity on the AI.
@TheMrCougarful7 ай бұрын
Not doing. Get ready.
@Niblss7 ай бұрын
It's shocking how the only thing you people can think of in a scenario where humans are obsolete is to keep going with capitalism, because crumbs are all you should ge You people terrify me
@senethys6 ай бұрын
The scaling laws are not exponential at all. Quite opposites. We are hitting the limits transformers and that is why we are now focusing on making inference a lot cheaper.
@Alice_Fumo6 ай бұрын
Name one model significantly larger than GPT-4 which does not exhibit significantly higher capabilities to support your claim. If anything, we are hitting the limits of our power grids.
@collins43597 ай бұрын
how does this still have only 12k views
@jannichi64317 ай бұрын
Do TOTAL votes get added up when syphoned off by KZbin type middlemen? Obviously I don't have a Podcast to know how viewers are calculated worldwide⁉️
@Saliferous7 ай бұрын
Ai fatigue.
@msabedra16 ай бұрын
How do we know this isn’t just two AI agents talking to each other and gaslighting us?
@vamps30006 ай бұрын
CEO of AI company hype his product, in another news water is wet
@skylark88287 ай бұрын
AI is limited by the chip hardware it uses, so until the chip fab plants can be made obsolete somehow there won't be exponential increases in AI progress. GPT4 was released a year ago but there is no perceived exponential jump in capabilities, instead we are seeing multi-modal AI's and the refining of AI training methods along with throwing ever larger amounts of compute at it.
@ManicMindTrick6 ай бұрын
This is not true. The algorithms are clunky and poorly optimized and you have a lot of hardware overhang available to be exploited to its full power by something much more sophisticated and intelligent.
@skylark88286 ай бұрын
@@ManicMindTrick LLM's are still using brute force approaches, and throwing ridiculous amounts of compute at the problems they cannot overcome is not going to achieve anything meaningful let alone exponential growth in AI performance. The hype bubble is about to burst.
@ProteusTG6 ай бұрын
All AI learning is fair use. People learn from others Why is an AI learning from people a problem? We all learn from work done by others.
@tommoody7286 ай бұрын
I think super human intelligence is a good thing, in fact it may be essential for our continued survival as an advanced civilisation.
@andym46955 ай бұрын
If we're talking about giant computational/training of models, this sounds an awful lot like a distributed computing problem. Seti@home or something of the such.
@pchasco5 ай бұрын
Model training is not done with general purpose CPUs. It is done with GPU’s and likely ASICs. It’s unlikely that training models with the SETI@home distributed method would be extremely inefficient if it works at all.
@kokomanation7 ай бұрын
This sounds like an AI generated conversation 😂
@Ben_D.7 ай бұрын
You should find an interview where you can see Dario as he speaks. He is quirky. Not at all a bot.
@dr.mikeybee6 ай бұрын
Llama 3 may be big enough already with the right agents.
@hklinkerАй бұрын
The questions/musings/assertions etc from the podcast host are long. Really long.
@benmurray29316 ай бұрын
The problem is that he is compromised by his role. He has to hype the technology in order to justify the capital being poured into his company. Same for every CEO of every LLM/diffusion model startup out there. There are many researchers who disagree with this take, and have arguments as to why more scale is unlikely to have a transformative effect compared to where we are now. What if the function learned by AI is exponential in complexity, and so adding extra zeros doesn't dramatically increase the scope of problems that can be solved by it?
@berniemadoff96887 ай бұрын
I'll save everyone here some time. An A.I. Guy likes A.I.
@ralphtoivonen20714 ай бұрын
If data storage and processing is going up exponentially then energy costs must go up exponentially. Laws of thermodynamics are a bummer. Also what will all the humans do when all the fun creative and intelligence intensive jobs are disappearing. Hmmm.
@brett70777 ай бұрын
If AGI pans out (scaling laws hold), all of Ezra’s small minded questions will be laughable
@raoultesla22926 ай бұрын
cute channel. Lockheed/MIT grad students/DARPA have surpassed your most Sci-Fi considerations 5+yrs ago.
@tristan72167 ай бұрын
When does the exponential curve get us to AI that doesn't need so much compute and data to learn? We have an existence proof that an agent can learn to do things without a billion dollars worth of compute - us. But our brains are millions of times as energy efficient as GPUs.
@DavenH6 ай бұрын
Where is the accounting for the evolution, and world simulation that it required? We do not have an existence proof.
@DavenH6 ай бұрын
Also the millions of times as energy efficient - lets see some actual numbers. That's not passing the sniff test.
@garyjohnson14667 ай бұрын
Why not use AI robots to explore other planets and moons in our solar system, even construct bases, as well as stations orbiting planets, as they can operate for extended periods in outer space without oxygen or special suits, to do any number work, repair or construction of the station…
@naomieyles2107 ай бұрын
We are using AI robots to explore Mars already. The rovers, and the Ingenuity copter, and even the landers, are Al robots. Their limitations show us the current forefront of AI robots working in hazardous environments.
@garyjohnson14667 ай бұрын
@@naomieyles210 yes, true, but in a limited capacity, many advancement have been made, I imagine someday they will be used onboard space station to perform hazardous repair work outside or to be part of the crew etc etc..
@naomieyles2107 ай бұрын
@@garyjohnson1466 specialised little AI robots for specialised jobs in the vacuum of space. Totally agree, and much safer if astronaut spacewalks are limited to training exercises or as Plan B if the AI robot can't do something. The AI robots would also respond to danger alerts by hurrying to a predetermined safe invacuation (lockdown) point. Invacuation rather than evacuation. 🙂
@skylineuk14857 ай бұрын
Look what happened in Blade Runner!
@garyjohnson14667 ай бұрын
@@skylineuk1485 yes, an like all created beings, they wanted to live, in the end he save the blade runner, showing his humanity, but Rachael was created without a termination date, but died giving birth, something that was supposed to be impossible..
@Bronco5416 ай бұрын
33:00 on "being better at persuasion by lying than telling the truth". Once again this should not be a surprise; humans are the same. People believe what they want to hear, not the truth.
@Bronco5416 ай бұрын
I disagree that it is very hard to bullshit. Actually im inclined to think its easier for less intelligent people to bullshit. Its kind of what they do; they have a weaker understanding of truth and necessarily a different respect/relationship to it versus smater and more mature people.
@BrianMosleyUK7 ай бұрын
38:50 I've wondered for a while, instinctively if a sense of the truth will be an emergent ability of next generation LLMs.
@AB-wf8ek7 ай бұрын
I think it's all about phrasing. At this point all they really need to do is place a confidence metric, which is simply based on how much of the training data correlates with the output. If developers simply included that, then people could judge better for themselves whether the information is accurate or not. Though this also needs to be taken with a grain of salt, because even the underlying training data can be manipulated by public relations campaigns, i.e. private sector propaganda, which is an older problem that's been around since mass media was invented.
@DanFrederiksen6 ай бұрын
You claim exponential advancement but notice that the attention mechanism is 10 years old and they are still using that and it still falls very short. What GPT4 can do is amazing but also very limited and there isn't a massive leap between 3 and 4. Indeed with the updated 3, it's hard to tell them apart. Despite 10x brain size. So progress isn't exponential at all, more like asymptotic. Also notice that Dario talks about how it might be as opposed to he knows how it will be. Meaning they don't have anything that will advance it. Just ever bigger models will only make nvidia very rich and drain some big companies. Progress in AI is predicated on a breakthrough and there hasn't been one in 10 years. A breakthrough is possible at any time but this current wave could well be a dead end. A very costly dead end. Altman has also expressed frustration and I don't think that's an act. Even if we say progress happens I'm pretty sure that anything that could fit the scifi term AGI will require vastly greater complexity and sophistication in the learning methods. Like transformers being a hack that doesn't even count as 1 and you might need 20 real potent techniques. And maybe 40 polishing heuristics. But I also suspect that we will see more narrow scope AI with quite potent capability. It doesn't need the well rounded versatility of human level intelligence to be significant. If the sophistication present in the human brain was given unlimited resources it would blow everything away. You see a glimmer of that in savants. What the early illusion of success with LLMs doesn't tell you is that real intelligence is quite complex. When I say 20 or 40 times more complex I could be underestimating it. Some of it might be compensated for with just brute force. It will look like AGI before it is close to the sophistication of human intelligence. And to some it already does.
@anatoly.ivanov6 ай бұрын
@01:16:46 - So Dario Amodei avoids replying to _the_ question about IP rights, twice? Including the very direct “hey, you’ve used my text” one from Ezra?! What’s the deal, then? As a director-producer, am I supposed to tell my actors, DP, VFX guys, costume, makeup, cooks, logistics… “You know what, work for free, cause you got UBI”?! And who’s going to pay for that UBI, which is supposed to be “basic”, not covering extra “discretionary” spending on stuff like “going to the cinema” or “paying Anthropic”? All that after taking all the planet’s electricity we’d might need to desalinate ocean water to drink and keep the AC on to survive? 😮🤯
@dovekie34376 ай бұрын
You could tell this interviewer was just DYING to take the moral highroad-I am surprised he held off for over an hour before showing his true colors: "I have written so many wonderful things and the AI is stealing my exceptional and unique prose, where is the compensation myself and other great people like myself?" It's already bad enough to stare at just a still photo of this guy. Nobody is using a variation on this guy's prose other than himself, that's for sure. He should be thankful he had enough publicly available writing that he doesn't need to pretrain a model to write a little for him.
@canadiangemstones76367 ай бұрын
How many billion will it take to just give me good results on a google search, instead of 99% garbage?
@Jasper_the_Cat7 ай бұрын
All I want is for it to generate a list of my availability in Outlook for a week and not come up with a hallucination. But yeah, they could start with improving Google search.
@jannichi64317 ай бұрын
Ironically, the Wawei phone from years ago gave me much better Google searches! Now that.was.before KZbin heavy usage and my algorithm certainly wouldn't be what.it.is now. !?!?!? D
@jeffkilgore63207 ай бұрын
Ridiculous comment. Ask it smarter search questions.
@rstray48017 ай бұрын
Need a Time Machine to send you back to 2011
@Saxafruge6 ай бұрын
@@superresistant0 I actually using an llm 90% of the times where i used to use search engines
@maxheadrom30887 ай бұрын
C'mon! Making AI and Oracle equal concepts is not only wrong but can also be dangerous! Dangerous because it dumbs down the listeners and also because Larry Ellison could end up suing!
@eSKAone-6 ай бұрын
We are not in control. Humanity is its own animal. This is inevitable. Biology is only 1 step of evolution. So just chill out and enjoy life 💟🌌☮️
@Claire-cs3gl7 ай бұрын
You still work there? kzbin.info/www/bejne/pIHFiqd5mtWee8Usi=deklNWmuEAAEJDva
@klausgartenstiel45864 ай бұрын
the next "3 to 18 months" sounds a lot like 3d6 to me.
@megavide06 ай бұрын
26:21 ".. how persuasive these systems/ your systems are getting as they scale..."
@maxheadrom30887 ай бұрын
11:15 The Apple Newton could do that in .. .I don't know ... late 1980s or early 1990s.
@volkerengels52986 ай бұрын
I feel like listening to an Petro-CEO in the seventieth
@brett70777 ай бұрын
I don’t think Ezra gets it
@dprggrmr7 ай бұрын
It's all fun and games until the great ai war
@AnthonyBurback7 ай бұрын
he's not even right about haircuts...
@scottharrison8127 ай бұрын
If AI can help me to connect to my car bluetooth for navigation and music - I’ll be happy.
@seanharbinger6 ай бұрын
I doubt the autogenerated pile of words was very good.
@artificialintelligencechannel7 ай бұрын
Amodei is talking about the exponential curve and investing in more compute. But surely there must be a way to reach human-level AI more efficiently? Using hybrid systems?
@Tayo396 ай бұрын
a month old AI vid ??? tf is is wrong witchu, algorythm ?
@RodCornholio6 ай бұрын
Dario should train AI on his fluency in buzzword-corporate-speak. He sounds as if he’s selling AI stock or is shilling for it.
@quanchi69727 ай бұрын
this was an incredible interview, however i doubt you'll ever get Amodei back on simply because your attitude (not your questions) was rather catty and combative
@Gee3Oh6 ай бұрын
These AI people are selling pipe dreams. LLMs are just the predictive text on your phone keyboard except trained on larger internet-scrapped data instead of data proprietary to the company. Yes this has the effect of generating coherent sentences but it’s a parlor tricks. There’s no intelligence at play at all and they use warehouses full of low paid 3rd world workers to feed the model human answers to further disguise the parlor trick. The most useful AI development of recent will be Adobe’s generative fill. They actually have the license for the training data and the industry experience to integrate the machine learning tools where they’ll be most useful. Chatbots aren’t productive. They wont be setting up birthday parties anytime soon. They’ll always just generate plausible sounding but unreliable text.
@stephenboyington6307 ай бұрын
AI is not going to be used to solve problems. It will be used to identify ways to make money. Make money in, make money out.
@nomadv78607 ай бұрын
It’s already being used to catch cancer early so it’s definitely being used to solve problems
@privacylock8557 ай бұрын
... and many of them will be criminal
@Cagrst7 ай бұрын
Now Stephen, this may blow your mind but one of the best ways to reliably make money is solving other peoples problems.
@haiderameer94737 ай бұрын
It’s already being used to accelerate drug discovery