Markov Decision Processes

  Рет қаралды 76,439

Bert Huang

Bert Huang

Күн бұрын

Пікірлер: 41
@hosamfikry2924
@hosamfikry2924 5 жыл бұрын
That is the best video I watched so far to understand this topic
@hobby_coding
@hobby_coding 4 жыл бұрын
very good lecture maybe the best introduction to this topic i've ever seen on youtube
@Pexers.
@Pexers. 3 жыл бұрын
Thank you, I spent hours in this algorithm, finally understood it !
@coeusmaze9413
@coeusmaze9413 5 жыл бұрын
The video provides intuitive but deep understanding in MDP
@syedrumman3920
@syedrumman3920 2 жыл бұрын
This is such a clear explanation!! Ty for this!! I wish I had taken your class while I was in VT!
@srujayop
@srujayop 2 жыл бұрын
Is the reward R(s) actually R(s')? And should that also be multiplied with the transition probability? max(over a) sum P(s', r|,s, a) [r + gamma*V(s')] ? I am trying to relate the equation presented in the video to standard notation 4 par notation.
@Ahmed.r.a
@Ahmed.r.a 7 ай бұрын
thank you for this brilliant explanation. I wished there was a Question with solution to practice on.
@jub8891
@jub8891 Жыл бұрын
thank you so much, you explain the subject very well and have helped me to understand..
@ryanflynn386
@ryanflynn386 5 жыл бұрын
This is a great explanation video, thanks so much. Your voice is easy to listen to too haha.
@berty38
@berty38 5 жыл бұрын
Ryan Flynn Thanks! I’m glad it’s helpful. My smooth voice is a huge disadvantage when I teach morning classes and my students all fall asleep.
@tarik8622
@tarik8622 4 жыл бұрын
Very interesting topic. And i think that you will make a fortune if you use your voice in publicity field. Best regards.
@jff711
@jff711 3 жыл бұрын
Thank you very much, very well explained.
@xruan6582
@xruan6582 4 жыл бұрын
can anyone explain (32:00) the switch between two modes (i.e. represented by green and red arrow). To me the green one seems like deterministic rule, the red one seems like stochastic rule. Can they exist simultaneously?
@behmandtirgar
@behmandtirgar 4 жыл бұрын
I have a question at time 8:30 : if we take an action to go to the left, why Pr(c | b, left) isn't 0.00? (we go to another side)
@seanxu6741
@seanxu6741 Жыл бұрын
Fantastic video! Thanks a lot!
@treegnome2371
@treegnome2371 4 жыл бұрын
at 17:35, why isn't it gamma = (0,1), instead of (0,1]...if gamma = 1, the influence of the actions farther down the road stays the same as all other actions, rather than shrinking the influence...right?
@JustinMasayda
@JustinMasayda 2 жыл бұрын
This was fantastic, thank you!
@sanskarshrivastava5193
@sanskarshrivastava5193 3 жыл бұрын
Best video for MDP on youtube
@Throwingness
@Throwingness 3 жыл бұрын
Around 34:00 when there are equations on the screen you should have had a pointer or something to point at what you are talking about. It's not clear.
@richardm5916
@richardm5916 4 жыл бұрын
Realy great explaintion on Machine learning
@consolesblow
@consolesblow 5 жыл бұрын
Thanks a lot! I found this very helpful.
@joshuasegal4161
@joshuasegal4161 5 жыл бұрын
What software are you using to make this?? It looks like you have like an infinite page which gives a really clean look
@berty38
@berty38 5 жыл бұрын
Nothing too fancy. This was done with Apple Keynote, and I'm faking that scrolling effect with "Magic Move" animations. I'm always looking for better tools to build useful visuals for lectures.
@quantlfc
@quantlfc 2 жыл бұрын
Absolutely amazing lecture!!!
@JebbigerJohn
@JebbigerJohn Жыл бұрын
This is so good!!!
@ismailasmcalskan2552
@ismailasmcalskan2552 4 жыл бұрын
Really good video about this topic. Thank you
@_brenda4975
@_brenda4975 3 жыл бұрын
much better than my lecturer
@linfrancis5204
@linfrancis5204 5 жыл бұрын
Great video. Thank you. Could you please make a similar video while we consider a two-dimensional Markov chain with more states?
@peterkimemiah9669
@peterkimemiah9669 3 жыл бұрын
Very good easy to understand.
@sander1426-2
@sander1426-2 4 жыл бұрын
Thanks for the explanation!
@rezadarooei248
@rezadarooei248 5 жыл бұрын
Thanks for your nice tutorial is it possible upload the slides?
@zenchiassassin283
@zenchiassassin283 4 жыл бұрын
What textbook ? thank you very much
@EdupugantiAadityaaeb
@EdupugantiAadityaaeb Жыл бұрын
What is the name of textbook
@jaideep_yes
@jaideep_yes 5 жыл бұрын
Thank you.
@y-3084
@y-3084 4 жыл бұрын
excellent
@dminn
@dminn 4 жыл бұрын
God bless
@ahmet9446
@ahmet9446 5 жыл бұрын
The best I find is [4, 1]. I couldn't achieve [4.2, 1.2]. Does anyone achieve [4.2, 1.2]?
@linfrancis5204
@linfrancis5204 5 жыл бұрын
YES, I GOT IT
@suvinaybothra8988
@suvinaybothra8988 4 жыл бұрын
honesty
@abdullahmoiz8151
@abdullahmoiz8151 4 жыл бұрын
33:27
@izazkhan1640
@izazkhan1640 5 жыл бұрын
jhk
Passive Reinforcement Learning
22:49
Bert Huang
Рет қаралды 13 М.
Markov Decision Processes Continued
33:04
Bert Huang
Рет қаралды 6 М.
Бенчик, пора купаться! 🛁 #бенчик #арти #симбочка
00:34
Симбочка Пимпочка
Рет қаралды 4 МЛН
Всё пошло не по плану 😮
00:36
Miracle
Рет қаралды 4,7 МЛН
Try Not To Laugh 😅 the Best of BoxtoxTv 👌
00:18
boxtoxtv
Рет қаралды 6 МЛН
Markov Decision Processes - Computerphile
17:42
Computerphile
Рет қаралды 170 М.
Markov Chains Clearly Explained! Part - 1
9:24
Normalized Nerd
Рет қаралды 1,2 МЛН
Lecture 8 MDPs
1:14:44
CS188 Spring 2014
Рет қаралды 41 М.
Lecture 8: Markov Decision Processes (MDPs)
1:07:10
CS188Spring2013
Рет қаралды 123 М.
Policy and Value Iteration
16:39
CIS 522 - Deep Learning
Рет қаралды 145 М.
RL Course by David Silver - Lecture 2: Markov Decision Process
1:42:05
Google DeepMind
Рет қаралды 633 М.
Generative Model That Won 2024 Nobel Prize
33:04
Artem Kirsanov
Рет қаралды 160 М.