You Need to Learn Importance Sampling NOW | Deep Out of the Money Options

  Рет қаралды 8,791

QuantPy

QuantPy

Күн бұрын

Пікірлер: 8
@Alexander-pk1tu
@Alexander-pk1tu 2 жыл бұрын
Hey man, I just wanted to say that you are very good at explaining those concepts and your code looks very good. Keep up the good work!!
@piepieicecream
@piepieicecream 2 жыл бұрын
Asked and delivered - great vid, thanks!
@MrEo89
@MrEo89 2 жыл бұрын
IIRC (and not totally sure if it’s applicable anymore with more recent versions of python & scipy, but ) I remember reading somewhere that using an internal function _pdf was Z times faster than the public function .pdf. Will Google later to see if it’s true, but I recall it was significantly faster.
@nicklashansen890
@nicklashansen890 2 жыл бұрын
Thanks for the great video (+1 like from me). I think that i do no fully understand the change of measure using the Radon-Nikodym derivative under the expected value. I know that you can define the Radon-Nikodym derivative of Q wrt. P (sometimes it's even called a Likehood process, L, but when?) as L(x) := dQ(x) / dP(x). However, how exactly does this work when taking expected values? I suppose it is somehow using the definition of the expected value being an integral but some details would be appreciated, thanks! My suggestion/ idea is something like this, where w (lowercase omega) is taken over the probability space W (uppercase Omega) E_Q[ X ] = int_W X(w) dQ(w) = Int_W X(w) dQ(w)/dP(w) * dP(w) = E_P[X dQ(w)/dP(w)]
@QuantPy
@QuantPy 2 жыл бұрын
The math is exactly how you’ve written it here, and what was showed in the importance sampling equation and change of measure for 25 std dev example. The expectation is the integral of the random variable over the domain space under the probability measure. When you change the probability measure to make your observed payoffs ect more likely, than as you’ve written in the 2nd line, you now are left with the ratio between the two probability distributions when integrating over the new probability measure space. When it comes to Monte Carlo, and taking the expectation, this translates as you’ve written in the 3rd line with the expectation with simulated variables under the new probability space multiplied by the likelihood of the simulated variables between the old distribution divided by the new one.
@natanalbuquerque1053
@natanalbuquerque1053 Жыл бұрын
Muito bom!
@jonasal-hadad7880
@jonasal-hadad7880 2 жыл бұрын
Hi, could you share the code with us?
@QuantPy
@QuantPy 2 жыл бұрын
Hi Jonas, yes working on updating my website, code tutorials will be out later next week
Risk Neutral Pricing of Weather Derivatives
27:14
QuantPy
Рет қаралды 7 М.
Importance Sampling
12:46
Mutual Information
Рет қаралды 69 М.
“Don’t stop the chances.”
00:44
ISSEI / いっせい
Рет қаралды 62 МЛН
Арыстанның айқасы, Тәуіржанның шайқасы!
25:51
QosLike / ҚосЛайк / Косылайық
Рет қаралды 700 М.
小丑女COCO的审判。#天使 #小丑 #超人不会飞
00:53
超人不会飞
Рет қаралды 16 МЛН
But what is the Central Limit Theorem?
31:15
3Blue1Brown
Рет қаралды 3,7 МЛН
2. Risk and Financial Crises
1:09:44
YaleCourses
Рет қаралды 419 М.
Perfect Shapes in Higher Dimensions - Numberphile
26:19
Numberphile
Рет қаралды 5 МЛН
Darts in Higher Dimensions (with 3blue1brown) - Numberphile
32:11
Numberphile
Рет қаралды 2 МЛН
How to calculate p-values
25:15
StatQuest with Josh Starmer
Рет қаралды 445 М.
Reinforcement Learning: Machine Learning Meets Control Theory
26:03
Steve Brunton
Рет қаралды 298 М.
But what are Hamming codes? The origin of error correction
20:05
3Blue1Brown
Рет қаралды 2,5 МЛН