#063

  Рет қаралды 37,916

Machine Learning Street Talk

Machine Learning Street Talk

Күн бұрын

We are now sponsored by Weights and Biases! Please visit our sponsor link: wandb.me/MLST to get started creating a centralised, system of record for your team's machine learning work.
Patreon: / mlst
Discord: / discord
For Yoshua Bengio, GFlowNets are the most exciting thing on the horizon of Machine Learning today. He believes they can solve previously intractable problems and hold the key to unlocking machine abstract reasoning itself. This discussion explores the promise of GFlowNets and the personal journey Prof. Bengio traveled to reach them.
Pod version (with no music): anchor.fm/machinelearningstre...
Our special thanks to:
- Alexander Mattick (Zickzack)
References:
Yoshua Bengio @ MILA (mila.quebec/en/person/bengio-...)
GFlowNet Foundations (arxiv.org/pdf/2111.09266.pdf)
Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation (arxiv.org/pdf/2106.04399.pdf)
Interpolation Consistency Training for Semi-Supervised Learning (arxiv.org/pdf/1903.03825.pdf)
Towards Causal Representation Learning (arxiv.org/pdf/2102.11107.pdf)
Causal inference using invariant prediction: identification and confidence intervals (arxiv.org/pdf/1501.01332.pdf)
A simple introduction to Markov Chain Monte-Carlo sampling
link.springer.com/content/pdf...
[00:00:00] Housekeeping
[00:01:20] Weights and Biases sponsor clip
[00:03:26] GFlowNets Introduction
[00:16:24] Interview kick off
[00:19:18] Galton Board Analogy
[00:22:20] Free Energy Principle Connection
[00:26:37] Diversity Preservation and Evolutionary Algorithms
[00:28:25] The multi-armed bandit perspective
[00:30:37] Avoiding Deception, Finding Unknown Unknows
[00:33:53] Where GFlowNets Find Free Lunch
[00:36:20] AlphaZero vs FlowZero (GFlowNets on Chess)
[00:40:08] Using GFlowNets for Interactive Search
[00:42:55] Learning Casaul Models as Graphs
[00:46:39] Learning Abstract World Models
[00:51:05] Can Machines Meta-Learn Categories
[00:54:22] The Consciousness Prior. Is GPT-3 Conscious?
[00:58:18] A Question For David Chalmers
[01:01:25] Why are linear models dominating? They are abstraction!
[01:05:23] Prof. Bengios Personal Journey (with Gary Marcus reference)
[01:10:02] Debrief: A Dream Come True!
[01:17:21] Abstraction is a Key
[01:21:27] A Funny Definition of Causal
[01:25:04] Arguing Semantics with a Semanticist
[01:30:07] Human Learning Over Evolutionary Time Scales

Пікірлер: 82
@MachineLearningStreetTalk
@MachineLearningStreetTalk 2 жыл бұрын
Oh yes! This one is a BANGER folks -- ENJOY!!
@alexfoo_dw
@alexfoo_dw 2 жыл бұрын
"Oh that's what I've been thinking for almost 20 years." I've been following Bengio's work for awhile now and this was truly an incredible conversation, in terms of both answers by someone patiently tackling fundamental problems in the field and questions by a group that has been patiently studying the field and connecting the ideas. Mad props!
@AICoffeeBreak
@AICoffeeBreak 2 жыл бұрын
OMG, is this real? 🤯 Awesome, can't wait to watch it whole!
@brokensymmetry307
@brokensymmetry307 2 жыл бұрын
Can't even describe how excited I am about this one! MLST strikes again, thanks guys!
@JackSPk
@JackSPk 2 жыл бұрын
This one was truly amazing! Imagine being Bengio, studying these topics for decades, replying to every question in hundreds of talks around the world, working with groups of bright people, and one day having a conversation for a youtube channel where he clearly enjoys and smiles at every question, to the point where he needs to say it out loud. It was like seeing Goku enjoying an amazing fight. The good vibes on this one were flowing both ways (pun intended), not only uncovering a better understanding of the topic but also motivating the viewer to learn more about it. Congratulations guys! This talk was pure delight.
@MachineLearningStreetTalk
@MachineLearningStreetTalk 2 жыл бұрын
Thanks Jack, we really appreciate it!
@nomenec
@nomenec 2 жыл бұрын
Thank you so much! We are very proud of this one and so very fortunate to have had him on the show.
@lenyabloko
@lenyabloko 2 жыл бұрын
This was really satisfying in terms of both questions and answers. So I can't even think of any question (or answer) not already given. We have our work cut out for all of us. Now you must have Jeff Hinton to complete the "Deep Learning Trio".
@michaeltraynor5893
@michaeltraynor5893 2 жыл бұрын
Geoff
@robbiero368
@robbiero368 2 жыл бұрын
Great to hear you have loads planned for this year
@osteinh12
@osteinh12 2 жыл бұрын
the best podcast out there... Keep it up!
@WilliamDye-willdye
@WilliamDye-willdye 2 жыл бұрын
I definitely need to read the papers about what they are calling "causality" here. It sounds very promising. Thanks for taking the time to post links in the description.
@marc-andrepiche1809
@marc-andrepiche1809 3 ай бұрын
if this is not one of the best episodes, I don't know what is.
@betoprocopio
@betoprocopio 2 жыл бұрын
Hell yeah! You have a merch store and patreon! Sweet that I don’t need to feel guilty about enjoying these insane quality videos and only supporting you with likes and comments hahaha
@user-xs9ey2rd5h
@user-xs9ey2rd5h 2 жыл бұрын
You guys and yannic Kilchers channel are really such good sources to get to know awesome research topics. You guys are the best!
@abby5493
@abby5493 2 жыл бұрын
Wow! Epic video! 😍
@jasonabc
@jasonabc Жыл бұрын
Amazing video introduction very well done and I really dig the music it put me into a meditative state
@Fordance100
@Fordance100 2 жыл бұрын
Another great show. I liked the intro a lot, that put me in the right frame of the mind for the following discussions. Prof. Bengio has scientific ways of thinking and explains various problems in AI and machine learning. I learned a lot.
@icriou
@icriou 2 жыл бұрын
What a wonderful interview. Prof. Bengio is honest and brave with class.
@daniilbelikau1247
@daniilbelikau1247 2 жыл бұрын
Wooow, the production value is impressive! Listening to the audio version is not the same
@MyU2beCall
@MyU2beCall 2 жыл бұрын
It might be interesting to invite Judea Pearl and ask his opinion about the possibilities for AI to grasp causal inference.
@lotfullahandishmand4973
@lotfullahandishmand4973 Жыл бұрын
the explanation in the begging was great, and prof. Bengio made it more discernable.
@glassrocketstair
@glassrocketstair 2 жыл бұрын
"we had a paper, i think it was at Neurips" ... lol he's so successful he can't even remember what he's published at neurips
@MachineLearningStreetTalk
@MachineLearningStreetTalk 2 жыл бұрын
🤣
@user-ut4zh3pw7l
@user-ut4zh3pw7l 5 ай бұрын
Thanks for great conversation. Hope someday understand it fully and make it actually work >D I hear same things over and over again. Feels good to know direction of the field in some sense.
@quebono100
@quebono100 2 жыл бұрын
Awesome episode :)
@Tuasmanque
@Tuasmanque 2 жыл бұрын
On a roll!
@MyU2beCall
@MyU2beCall 2 жыл бұрын
Well done. I'm looking forward to your interview with David Chalmers and his reaction on the 'awareness and experience' of an AI algorithm.
@oncedidactic
@oncedidactic 2 жыл бұрын
You guys deserve the props from bengio, well done! I would love to hear more discussion and dissection of this “free lunch if there’s structure” notion. How exactly are we beating dimensionality and combinatorial vastness? The answer has to be abstraction arising from, or baked into, the architecture + algorithm. But if this is truly effective at scale, then it implies weighting for exploration converges on parsimonious modeling- so that you can get powerful and versatile composition of abstractions. This is interesting: it associates “casting about” with “finding the best explanation” (which presumably tends to generalize or transfer well). Sort of turning exploitation on its head, no? (The conversation kept circling around the attractors of GOFI theme of composing abstractions, and information as a reward.) Honestly I deeply appreciate the irl oracle question search aspect, but I think Keith is on to something with the flowZero line of thought. It would be informative to understand learning rate and policy space in a setting where we already have some kind of grasp.
@LiaAnggraini1
@LiaAnggraini1 2 жыл бұрын
That title though. Bet it would be a great talk!
@binjianxin7830
@binjianxin7830 2 жыл бұрын
Mindblowing!
@louis3195
@louis3195 2 жыл бұрын
Awesome work guys 😋. Please interview Max Tegmark!
@Kartik_C
@Kartik_C 2 жыл бұрын
MLST is the best!!
@MachineLearningStreetTalk
@MachineLearningStreetTalk 2 жыл бұрын
Thanks! We really appreciate our amazing audience 😎
@welcomeaioverlords
@welcomeaioverlords 2 жыл бұрын
Quicker than a what now?
@MachineLearningStreetTalk
@MachineLearningStreetTalk 2 жыл бұрын
"Whippet with a bum-full of dynamite" ;)
@mattizzle81
@mattizzle81 2 жыл бұрын
lol @ the twitter reference. As somewhat of a Twitter addict myself recently, imagine... to just *imagine* the productivity improvement from shunning Twitter. It is unfathomable.
@sapito169
@sapito169 2 жыл бұрын
joshua benshi is a rock star of ml
@ZachDoty0
@ZachDoty0 2 жыл бұрын
On the topic of consciousness, I just finished The Case Against Reality by Donald Hoffman. It would be great if you could invite him to the MLST show.
@MachineLearningStreetTalk
@MachineLearningStreetTalk 2 жыл бұрын
Great idea! We would love to have him on
@rogermarin1712
@rogermarin1712 Жыл бұрын
@@MachineLearningStreetTalk also bernardo kastrup
@SLAM2977
@SLAM2977 2 жыл бұрын
How can you have abstractions without understanding first? A NLP system doesn't know what the text mean as the text is referring to something inaccessible to it(external world and its entities and dynamics), Can you recover the full information from just text? Obviously not.
@lenyabloko
@lenyabloko 2 жыл бұрын
Actually, you can! But that depends on, well - semantics, that is how you define what abstraction is. For example, so called abstract art does not (always) builds on understanding. Rother it interacts with understanding, some times changes it and vise versa. It is more constructive to speak about generalization and systematization since both have semantics at least partially defined. This why I always respond to people saying that LLM like GPT-3 do generalization - not to confuse generalization with degeneration - that is simplification.
@mgostIH
@mgostIH 2 жыл бұрын
Current models already aren't forced to get information from text alone, but a recent paper titled "One model for the learning of language" shows you can indeed learn semantics from observing a language and do that quite fast.
@SLAM2977
@SLAM2977 2 жыл бұрын
@@lenyabloko You can do it a degree, can you understand smell from text on the web?
@alighahramani2347
@alighahramani2347 Жыл бұрын
Thats me :D:D:D:D.
@marilysedevoyault465
@marilysedevoyault465 2 жыл бұрын
A simpler idea for making longer predictions from videos and chronology with what already exists ! - by Marilyse Devoyault, 1. Make time lapses from big data of available videos. For each image obtained, link : shooting number & real time date-hour of image : Ex: from shooting #789 2022-04-22 20:38:01 2. Use classifier to identify main elements of the identified context at the beginning (some pictures at the beginning with detailed elements to tell what to look for) Let say you have a DallE2 give you a first image of a dog, a cat and a tree. You take this picture to input in your new predictor with this classifier. It could also be a robot taking a picture of a new situation in front of him. 3. Have this predictor find every picture in the data that is close to your picture 4. Have some type of transformer find the next picture (using number of shooting date-hour) of every picture you found in 3. 5. Use some king of GFlow net to regroup similar next pictures found in 4 and keep the main probabilities. 6. For example, two possibilities : step 3, a picture with a dog seeing a cat and a tree nearby. Step 4, many next images possibles : cat runs toward tree and dog chases; cat turns toward dog and spit; cat lay down and dog happy, dog go away and cat don’t move; bird comes from the tree and land on the dog head… Step 5, the GFlownet regroup all the images of the cat running toward the tree, and all the images of the cat facing dog and spitting. Since they are numerous, they are the main probabilities kept for the predictions and to go on with next prediction 7. Take the images from the main flows of 5 and find the next picture (using number of shooting date-hour) of every picture you found in 5. 8. Use the GFlow net to regroup similar next pictures found in 7 and keep the main probabilities. 9. Go on as long as you can to make a longer plausible prediction with main probabilities of what can happen.
@gaceladri
@gaceladri 2 жыл бұрын
Hell yeah!
@jamesbuchanan27
@jamesbuchanan27 Жыл бұрын
Tim, can you square how you seem to react positively to YB's proposal at the one hour mark that this mini world model could give us the "illusion of Cartesian Dualism" but then are very negative on Ilya's "a little bit conscious" comment on his Transformer enable architecture. They sound like similar ideas, no?
@snippletrap
@snippletrap 2 жыл бұрын
MLST: Bengio, you're the best! Bengio: No U! What a lovefest
@DhruvMetha
@DhruvMetha 2 жыл бұрын
Wow
@koudbi7941
@koudbi7941 2 жыл бұрын
wondeful
@godbennett
@godbennett 2 жыл бұрын
1:09:11, I have the same thoughts on the matter
@sarah-lp2oc
@sarah-lp2oc 2 жыл бұрын
The financial market has tough one this past months, but I watch interview on CNBC where the anchor kept mentioning "...CATELYN MORRIS...". This prompted me to touch with her, and from October 2021 till now we have been working together, and I boast now of €35k in my trading wallet.
@harriswealthers1579
@harriswealthers1579 2 жыл бұрын
this is the miss conception going around, you call it love for money while some see it as receiving good information, which can still be a miracle in the making by God, how has the father worked miracles since the days of abraham it has always been through men
@lucyweilbel6681
@lucyweilbel6681 2 жыл бұрын
Your English may be poor even if your intuition to deduce money methods are impeccable.. .how were you able to get a meet .this is rare
@daytradingaddict5632
@daytradingaddict5632 2 жыл бұрын
@@harriswealthers1579 lots of people forget this, they are still waiting on manner to fall out of the skies personally I pity such people they have been brainwashed by society that things are meant to given to them which is wrong you find things and work for them either by getting cautious with sensitive info and knowing which info to act on
@sarah-lp2oc
@sarah-lp2oc 2 жыл бұрын
@@lucyweilbel6681 with norristrades as the user name .. we talk on the t e l e g ram better to ask madam you'rself
@helenp9085
@helenp9085 2 жыл бұрын
@@sarah-lp2oc your english is funny 😅😅
@dinoscheidt
@dinoscheidt 2 жыл бұрын
1:16:13 well than we‘re both suffering 😂 oh boy
@robbiero368
@robbiero368 2 жыл бұрын
Do any language models take their own output and feed it back in with the next input from the human? The thought here being that currently these bots are having a weird sort of interaction where they are only fully aware of one half of the conversation
@robbiero368
@robbiero368 2 жыл бұрын
Maybe it would require something like a GAN to pull it off, and also maybe if it did the network would have some notion if self
@futurisold
@futurisold 2 жыл бұрын
Discord link?
@MachineLearningStreetTalk
@MachineLearningStreetTalk 2 жыл бұрын
discord.gg/HNnAwSduud
@SimonJackson13
@SimonJackson13 Жыл бұрын
Sounds like a high gradient is well changy.
@SimonJackson13
@SimonJackson13 Жыл бұрын
Estimated maximal possible differential between d2/dx2 maxima and minima, for placement of secandary "randomization" betterment?
@SimonJackson13
@SimonJackson13 Жыл бұрын
Fixing a set of dimension reductions? Eventually expanding a split into a small dimension which would split any min/max. So fix y, solve, set y=0.0000x ... and flow from a 1D to a 2D ...?
@SimonJackson13
@SimonJackson13 Жыл бұрын
A hyper cube in 1D is just a line.
@SimonJackson13
@SimonJackson13 Жыл бұрын
Add the last latent layer 1 everything neuron to many delta neurons one by one at a time?
@SimonJackson13
@SimonJackson13 Жыл бұрын
Is delta a latent causality? Changing one dimension changes all the latent delta nuron outputs?
@gustafa2170
@gustafa2170 2 жыл бұрын
I don't think neuroscience will ever get an explanation for how consciousness arises. Matter, this abstract, contour that can be fully described by quantities (completely devoid of qualities) and subjective experience are two incommensurable categories. The best you will get is "and poof, consciousness arises".
@Addoagrucu
@Addoagrucu 2 жыл бұрын
Feel sad for the guy on Yannic's discord who watched 1/3rd the way through and stopped. At just about that point it goes from 0 to 100 real quick.
@rufex2001
@rufex2001 Жыл бұрын
Diversity, baby!
@mainsdor
@mainsdor 2 жыл бұрын
Who is the guy at the bottom right?
@TimScarfe
@TimScarfe 2 жыл бұрын
Keith Duggar
@nomenec
@nomenec 2 жыл бұрын
Uh oh ... it's me. Dare I ask why?
@swayson5208
@swayson5208 2 жыл бұрын
pog
@alexijohansen
@alexijohansen 2 жыл бұрын
Suggesting today that a machine or system can be conscious to any degree is the equivalent of suggesting the earth revolves around the sun hundreds of years ago. What it means to be human needs to be re-evaluated if true, we (humans) tend not to like to do that.
@marilysedevoyault465
@marilysedevoyault465 2 жыл бұрын
Oh no! Mr Stanley stired my artistic fiber… A question to GFlowNets The head as a materialised Plasma ball Every impulse from the outside Creates a pattern in the grey matter, chronologically impregnated. Monday 7 o’clock in the morning, I see my pineapple on the counter. This creates a pattern looking like a Dryopteris in my visual cortex. Monday 7 o’clock at night, I taste some new basswood herbal tea. This creates a pattern looking like an Osmunda in my gustatory cortex. Tuesday 7 o’clock in the morning, I see my pineapple on the counter. It isn’t ripe. This creates another pattern looking just like a Dryopteris with a slightly different shape in my visual cortex, This pattern is infinitely close to the Monday 7 o’clock am pattern of a Dryopteris, but absolutely not the same. It is impregnated almost at the same place in my grey matter, but infinitely slightly more inside. The discrepancy is infinitely small. Tuesday night the taste of my basswood herbal tea and its Osmunda pattern will slip right next to my Monday night slightly different Osmunda pattern. More inside. I am grasping the taste of basswood herbal tea. Could it be how everything is chronologically impregnated by electric impulse? Trillion of trillion of microscopic layers? Could it be how we can make predictions? Since everything is chronological in my materialised plasma ball? Could it be how we generalise with infinitely small layers (chronological layers) of tiny hair of patterns that are alike and almost merge, straw inside a straw inside a straw inside a straw, but when this grey matter area is visited by an impulse, flows the general concept of a pineapple or basswood herbal tea? GFlowNet, will you learn to consider time? Will you make chronological layers of flows? Will you learn the chronology of your encounters so that you may imagine the future? Will you use the flows with numerous layers to grasp the Platon Idea and it’s relation to time, in other words, its probabilities to exist following the happening of a previous Platon idea? Between two dense areas of layers of straws, are most of the straws of one area pretty much from the same impregnated impulse jet of the other area of dense layers of straws? Yes ? Then we have a dialectic!
@notgabby604
@notgabby604 2 жыл бұрын
He didn't really have an answer for: [01:01:25] Why are linear models dominating? They are abstraction! However there are answers out there, even if they require a certain human psychological reboot to move drastically to a new vantage point on the problem. Maybe poke around on archive dot org with ReLU as a switch.
@Fordance100
@Fordance100 2 жыл бұрын
The same way when we approximate derivative as the linear slope of two dots when they are very close to each other, also similarly in calculating integral. We can approximate a curve with a bunch of linear lines. Relu creates a kick of non-linearity on top of a linear function, makes it even easier for neural nets to carry out the task. It's much easier to learn linear functions than less defined non-linear functions.
KARL FRISTON - INTELLIGENCE 3.0
2:59:21
Machine Learning Street Talk
Рет қаралды 124 М.
#67 Prof. KARL FRISTON 2.0 [Unplugged]
1:42:11
Machine Learning Street Talk
Рет қаралды 11 М.
ПЕЙ МОЛОКО КАК ФОКУСНИК
00:37
Masomka
Рет қаралды 6 МЛН
Who enjoyed seeing the solar eclipse
00:13
Zach King
Рет қаралды 136 МЛН
I MADE A CARDBOARD SWING!#asmr
00:40
HAYATAKU はやたく
Рет қаралды 30 МЛН
This is why Deep Learning is really weird.
2:06:38
Machine Learning Street Talk
Рет қаралды 312 М.
GFlowNets for generative active learning | Amazon Science
29:18
Amazon Science
Рет қаралды 9 М.
#55 Dr. ISHAN MISRA - Self-Supervised Vision Models
1:36:22
Machine Learning Street Talk
Рет қаралды 22 М.
#69 DR. THOMAS LUX - Interpolation of Sparse High-Dimensional Data [UNPLUGGED]
50:39
Machine Learning Street Talk
Рет қаралды 14 М.
#036 - Max Welling: Quantum, Manifolds & Symmetries in ML
1:42:32
Machine Learning Street Talk
Рет қаралды 21 М.
AI AGENCY ISN'T HERE YET... (Dr. Philip Ball)
2:09:18
Machine Learning Street Talk
Рет қаралды 19 М.
AI Sentience, Agency and Catastrophic Risk with Yoshua Bengio - 654
59:40
The TWIML AI Podcast with Sam Charrington
Рет қаралды 2,8 М.
WE LIVE IN THE INFOSPHERE [Prof. LUCIANO FLORIDI]
1:06:52
Machine Learning Street Talk
Рет қаралды 17 М.
ICLR 2020: Yann LeCun and Energy-Based Models
2:12:12
Machine Learning Street Talk
Рет қаралды 21 М.
Prof. Chris Bishop's NEW Deep Learning Textbook!
1:23:00
Machine Learning Street Talk
Рет қаралды 75 М.
ПЕЙ МОЛОКО КАК ФОКУСНИК
00:37
Masomka
Рет қаралды 6 МЛН