Why You Shouldn't Trust Your ML Models (...too much)

  Рет қаралды 4,220

ritvikmath

ritvikmath

Күн бұрын

Whether you call it feedback loops, selection bias, etc, this pesky problem rears its head in almost every problem out there.
0:00 The Problem
9:35 The Solution

Пікірлер: 15
@mndhamod
@mndhamod 19 күн бұрын
I have a PhD in computer science with focus on deep learning and i still learn something new from your videos. I'm grateful for all the neat insights i get from your teaching!
@wenhanzhou5826
@wenhanzhou5826 23 күн бұрын
This phenomena occurs in Deep Q-learning and SARSA when you need the target Q-value in order to update the current Q-function. Especially in problems with continuous state spaces where the target Q-value is typically estimated using the same model. So the algorithm essentially try to predict a target and learn from that. One way to reduce this effect is to implement the epsilon-greedy policy which chooses a random action depending on the value of epsilon, which is conceptually similar to the concept in the video of keeping a small amount of randomness in the action of the model.
@climbscience4813
@climbscience4813 21 күн бұрын
I've had this same effect in some models I trained and I think there is one thing you can do that nearly gets completely rid of the effect is eliminating the effect of the model from your data. In my case it was possible to calculate the outcome if the model hadn't influenced the process. In the case you explained, what I would do is divide the numbers by the percentages of recommendations to compensate for the effect of the recommendations. It's essentially Bayesian statistics, where you try to determine the probability of the film getting watched given that it has been recommended. Hope this makes sense to everyone!
@aelloro
@aelloro 19 күн бұрын
Your channel is a treasure! Thank you!
@paull923
@paull923 22 күн бұрын
insightful video, thank you!
@mango-strawberry
@mango-strawberry 15 күн бұрын
hi ritivik ive been watching a lot of your videos. you explain very well. ive one request. can you do some videos purely on math topics that are required for ml? especially something like stats.
@Jack-cm5ch
@Jack-cm5ch 23 күн бұрын
Really unique video! I loved it. Do similar issues of a feedback loop occur with demand or price forecasting? And if so, how? I was thinking high demand on an item on one day put more bias on that item in the future?
@MorseAttack
@MorseAttack 20 күн бұрын
Missed a good opportunity to plug a “like and subscribe to train the model” 😂
@ritvikmath
@ritvikmath 20 күн бұрын
Haha good one!
@emmang2010
@emmang2010 23 күн бұрын
Thank you.
@user-sl6gn1ss8p
@user-sl6gn1ss8p 23 күн бұрын
Would you also take the diversity directly in account when training the next model? Like, say, if you can measure that the diversity had an effect different from the predicted, that indicates something, right?
@user-sb9oc3bm7u
@user-sb9oc3bm7u 20 күн бұрын
Probably should implement concepts from genetic algorithms to make sure you include in the next iteration's training set elements that were excluded in the i-1 output. Question: You start with N unique values. you sample with replacement N values (basically, bootstrapping). then you iterate this process, where the i input vector (of length N, always) is the output of iteration i-1. How many iterations would you need untill converging on a single value? Example: N=5 [1,2,3,4,5] i_1 = [1,2,2,4,5] i_2 = [1,2,4,5,5] . . . i_n-1 = [4,4,4,4,5] i_n = [4,4,4,4,4] Vector size: 5 Converged value: 4 (doesnt mean much. can be colors as well) # of iterations: n Answer: The larger the N (size of vector), the closer you get to number of iterations = 2 * N (so for vector of size 50, it'll take, on avg/expectancy, 100 times)
@zlucoblij
@zlucoblij 23 күн бұрын
Imagine you're paying out profit share to authors of your content based on popularity and your own recommendation model does this. Sucks to be the content creator... The way the diversity is implemented seems to be absolutely key...
@InfiniteQuest86
@InfiniteQuest86 22 күн бұрын
Yeah you shouldn't ever be training a new model based on a previous model. This is ignoring that you shouldn't just be recommending stuff that's popular. Which is already wrong. You should recommend related movies to what the user likes. Which avoids all of this.
@nononnomonohjghdgdshrsrhsjgd
@nononnomonohjghdgdshrsrhsjgd 21 күн бұрын
very unpleasant channel, starting from the loud background music.
Kernel Density Estimation : Data Science Concepts
25:52
ritvikmath
Рет қаралды 12 М.
I Used Data Science to Buy the Dip
19:32
ritvikmath
Рет қаралды 7 М.
Зу-зу Күлпәш. Стоп. (1-бөлім)
52:33
ASTANATV Movie
Рет қаралды 863 М.
Which one will take more 😉
00:27
Polar
Рет қаралды 80 МЛН
Surprise Gifts #couplegoals
00:21
Jay & Sharon
Рет қаралды 30 МЛН
The Most Important Algorithm in Machine Learning
40:08
Artem Kirsanov
Рет қаралды 187 М.
The most important skill in statistics
13:35
Very Normal
Рет қаралды 271 М.
TransformerFAM: Feedback attention is working memory
37:01
Yannic Kilcher
Рет қаралды 32 М.
The KL Divergence : Data Science Basics
18:14
ritvikmath
Рет қаралды 38 М.
Meta Announces Llama 3 at Weights & Biases’ conference
26:16
Weights & Biases
Рет қаралды 73 М.