Deterministic Image Editing with DDPM Inversion, DDIM Inversion, Null Inversion and Prompt-to-Prompt

  Рет қаралды 1,654

Gabriel Mongaras

Gabriel Mongaras

Күн бұрын

Пікірлер: 7
@EkShunya
@EkShunya 4 ай бұрын
great one, really liked it thanks
@FangruiHuang
@FangruiHuang 8 күн бұрын
why you can directly change x_t-1 to x_t+1 in 29:21.
@gabrielmongaras
@gabrielmongaras 2 күн бұрын
It depends on if you do a forward step or an inversion step in the opposite direction using the formulas above. Both come from x_t though.
@ml-ok3xq
@ml-ok3xq 2 ай бұрын
congrats on writing a paper! i notice that another recent paper from NVIDIA uses a unit vector for attention (nGPT) where the dot product is naturally equal to cosine as the lengths are one. are these two works related to each other in any way?
@gabrielmongaras
@gabrielmongaras 2 ай бұрын
Thanks!! I only read through the nGPT paper briefly, but I think nGPT was trying to make softmax attention/transformers more expressive and efficient by changing a few things. They do normalize before they apply the softmax function, making the logits a cosine similarity between -1 and 1. However they keep the softmax operation which forces the model to stay quadratic in terms of complexity. The paper I worked on removed the softmax function which allowed the attention mechanism to be changed into an RNN which is linear in complexity.
@陈兆伟-s5w
@陈兆伟-s5w 4 ай бұрын
How is the equality in DDPM established in 17:49?
@gabrielmongaras
@gabrielmongaras 4 ай бұрын
Looks like I forgot to write out the square root over the first term. As for the inner term that got turned into a fraction, I just multiplied sqrt{1-a_t} by the fraction sqrt{1-a_t}/sqrt{1-a_t}.
Quando A Diferença De Altura É Muito Grande 😲😂
00:12
Mari Maria
Рет қаралды 45 МЛН
She made herself an ear of corn from his marmalade candies🌽🌽🌽
00:38
Valja & Maxim Family
Рет қаралды 18 МЛН
Learning to (Learn at Test Time): RNNs with Expressive Hidden States
35:52
The Linear Algebra Trick that made Google (PageRank)
30:04
Google’s Quantum Chip: Did We Just Tap Into Parallel Universes?
9:34
KAN: Kolmogorov-Arnold Networks
37:09
Gabriel Mongaras
Рет қаралды 56 М.
Byte Latent Transformer: Patches Scale Better Than Tokens
45:05
Gabriel Mongaras
Рет қаралды 1 М.
WARP: On the Benefits of Weight Averaged Rewarded Policies
52:39
Gabriel Mongaras
Рет қаралды 758