Пікірлер
@user-yq8jp6bc5u
@user-yq8jp6bc5u 7 күн бұрын
Thank you so much Rachit, video is really awesome
@fightsatan2408
@fightsatan2408 10 күн бұрын
@venkyvenky4715
@venkyvenky4715 13 күн бұрын
but you can do getdummies before traintestsplit
@MRahdianEgakurnia
@MRahdianEgakurnia 29 күн бұрын
i have an already scaled data with powertransformer, can you really scaled a new data outside the scaled data with scaled data as standard using fit.? because ive tried this and the data seems dont mach the scaled data. Thank you
@DataAnalystVictoria
@DataAnalystVictoria Ай бұрын
Thanks! ❤
@umutg.8383
@umutg.8383 Ай бұрын
MICE part is good but the missingness definitions are all wrong.
@rishi1901
@rishi1901 Ай бұрын
excellent demonstration !I really appreciate your efforts. Very helpful for me as a beginner
@wtfashokjr
@wtfashokjr Ай бұрын
why pd.get_dummies not working for me ?
@nishantwhig7206
@nishantwhig7206 2 ай бұрын
Very clearly explained.Thank you.
@SodaPy_dot_com
@SodaPy_dot_com 2 ай бұрын
verey detailed with the parameters. love it
@DhirajSahu-ct1jp
@DhirajSahu-ct1jp 2 ай бұрын
Thank you so much!!
@r.s.572
@r.s.572 2 ай бұрын
thank you for explaining this! :) poor PhDs are thankful for people like you who use their free time to do such videos!
@KA00_7
@KA00_7 2 ай бұрын
in-depth and best explanation video
@KA00_7
@KA00_7 2 ай бұрын
learned something new today. Thank you so much
@rishidixit7939
@rishidixit7939 2 ай бұрын
If I want to use Simple Imputer on two different columns but with different strategies on each column then what should I do ?
@preethirathod6751
@preethirathod6751 2 ай бұрын
You have explained so clearly
@deeptimittal4552
@deeptimittal4552 3 ай бұрын
Wow now I completely understand pivoting. I was struggling to get the concepts, now its all clear. Thank you Rachit.
@longtuan1615
@longtuan1615 4 ай бұрын
That's the best video I've seen! Thank you so much. But in this video, the "purchased" column is ignored because this is fully observed. So what happens if missing values are only present in the "age" column, I mean the "experience", "salary" and "purchased" are fully observed and for the same reason, we will ignore them so we only have the "age" column that can not use the regression? Please help me!
@cadeepakgoyal7500
@cadeepakgoyal7500 5 ай бұрын
thanks a lot. really helpful
@jahnavinama8534
@jahnavinama8534 5 ай бұрын
well explanation bro..i am wathching 5 6 vedios about split method but the only one vedio is helpful for me and that's is yours
@DrizzyJ77
@DrizzyJ77 5 ай бұрын
Thanks Needed a clear explanation for my missed class😅
@dinushachathuranga7657
@dinushachathuranga7657 5 ай бұрын
Bunch of thanks for the clear explanation❤
@philcrom6299
@philcrom6299 6 ай бұрын
Wow, that really helped me in evaluating my master thesis!!!
@MrTau123
@MrTau123 7 ай бұрын
Eknumber.
@exanessa1234
@exanessa1234 7 ай бұрын
How imbalanced before AUPRC is preferred.
@focus72343
@focus72343 7 ай бұрын
very simple explanation, thank you and subscribed!
@ItzLaltoo
@ItzLaltoo 8 ай бұрын
Hey, the video was very helpful.. Can anyone explain me while implementing MICE in RStudio we get two columns Iteration & Imputation, how can we connect that with this video. Like in RStudio for each iteration we get 5 imputed dataset (by default). But from this video, we only get one dataset for a iteration.. It would be really helpful if anyone can explain me this. Thanks in advance
@cuoivelo8360
@cuoivelo8360 9 ай бұрын
Can you turn on subtitle for the videos? Im bad at English listening
@anonymeironikerin2839
@anonymeironikerin2839 9 ай бұрын
Thank your very much for this great explanation
@shoaibahmed5848
@shoaibahmed5848 9 ай бұрын
What about 1 row missing value and 4th row missing value is those values to be filled necessary?
@ShubhamKumar-xy6kj
@ShubhamKumar-xy6kj 10 ай бұрын
Great video bro...
@roshantonge1952
@roshantonge1952 10 ай бұрын
very good video
@modhua4497
@modhua4497 10 ай бұрын
Thanks, do you have example on how to incorporate LOG or SQRT transformation of features before modeling?
@zk321
@zk321 10 ай бұрын
The word “namaste" in Sanskrit means “bowing to you". Muslims believe that one can bow/prostrate only to Allah. We don't bow down to any human. It's important to note that religious beliefs and practices can vary among individuals and communities.
@ubaidghante8604
@ubaidghante8604 10 ай бұрын
Brother found some specific examples to explain MAR and MNAR 😅
@martinngobye3574
@martinngobye3574 11 ай бұрын
Great explanation regarding column-transformer and pipeline, however how do you have the data frame column names back instead of numbers? Thank you!!
@ishikaagarwal6945
@ishikaagarwal6945 11 ай бұрын
Nicely explained
@osoriomatucurane9511
@osoriomatucurane9511 11 ай бұрын
Namaste, Awesome, Sir. I must addmit the best tutorial by far in pandas groupby I have ever accross. Keep it up
@bellatrixlestrange9057
@bellatrixlestrange9057 11 ай бұрын
best explanation!!!
@mathewfernand8460
@mathewfernand8460 11 ай бұрын
Sir how can i get the dataset
@meenatyagi9740
@meenatyagi9740 11 ай бұрын
Very good explaination.I was struggling to get clearity on it .
@skyrayzor3693
@skyrayzor3693 Жыл бұрын
This tutorial is awesome!!
@subtlehyperbole4362
@subtlehyperbole4362 Жыл бұрын
(note: this is not an issue specific to your video, but something i have been getting confused by for a long time, this is just the first time I decided to stop and try ask about it in the comment section) It seems like it should be necessary (or maybe if not necessary, at least would be useful) to tell the model which column each imputed indicator is indicating for, right? But in the final dataset that you produce the imputed data indicators are all bunched up as the first four columns. How does the model know A) these features are imputed data indicators and, more importantly, B) which of the remaining 93 columns in the dataset each one is supposed to be for? They could be indicating for columns 5, 6, 7, and 8, or they could be indicating for columns 45, 72, 8, and 92, or any other combination of the remaining 93 feature columns. How does this not affect how the model trains? My brain is thinking that possibly the algorithm somehow susses that out on its own... but i don't understand how or why it can do that. Am i making much ado about nothing here?
@rachittoshniwal
@rachittoshniwal Жыл бұрын
At the very core of things, the computer only understands 0s and 1s haha. For human interpretability - yes, it might be necessary to label the columns to see what a specific missing indicator column is for, but for a computer, it doesn't matter. The column headers are just for us, as the model only cares about the data being in a 2D array format. For: A) the model doesn't know that a particular column is indicating missing values in some other column. It only cares about the values in it. B) Again, to reiterate, the model doesn't care what each of the other 93 odd columns stand for. It is only looking at their values You can shuffle the column ordering, pass in X.values instead of a dataframe X to the model. It will not affect performance
@subtlehyperbole4362
@subtlehyperbole4362 Жыл бұрын
@@rachittoshniwal Yeah I understand that but the 93 columns are all features in and of themselves, the 4 imputation indicators aren't really features of data in the same underlying way, right? They are more like features about other features, not features about the event whose label it is trying to train on. It seems like the imputation indicator data point's entire utility is essentially to point at a single data point in another column and say "don't take this data point too seriously because it was made up" -- wouldn't the entire weighting system need to treat that types of columns be different? It feels like it would be problematic (or at least, not useful) to treat those columns as if they were just additional features columns that could be treated like any other of the existing 93 features. (I mean, I guess it depends on the particular algorithm, like i imagine decision tree based algos would probably be able to handle that kinda thing, but others it feels like wouldn't be well served to treat those columns like they were just any other feature columns, no different than any of the others for its own starting purposes)
@rachittoshniwal
@rachittoshniwal Жыл бұрын
@@subtlehyperbole4362 although the missing indicator column is based off some column X, it is essentially a brand new column holding the information that "there is a column X which has missing values for these rows" so the model will test whether NOT having values in that column X is indicative of something or not.
@imranyounas4478
@imranyounas4478 Жыл бұрын
i did not understand if s = df['population'] then how remove outlier from all dataset instead of population column.
@prashu25925
@prashu25925 Жыл бұрын
Do we apply scaling techniques on categorical columns after encoding? Plz help
@rachittoshniwal
@rachittoshniwal Жыл бұрын
even if you apply scaling after encoding, the 0s and 1s will be converted to some new numbers, but all 0s will be the same number x and all 1s will be the same number y. So it is again essentially encoded, just that instead of 0s and 1s you have x and y
@amitblizer4567
@amitblizer4567 Жыл бұрын
Very well explained video. Thank you!
@kylehankins5988
@kylehankins5988 Жыл бұрын
I have also seen univariate imputation refer to a situation were you are only trying to impute one column instead of multiple columns that might more than one missing value
@olatheog
@olatheog Жыл бұрын
Great video, Rachit. Thank you. I also heard OneHotEncoding is not good for large categorical data in real world projects. Please which method do you advise or is there a video of you doing it that we can watch? Thank you so much
@tarun94060sharma
@tarun94060sharma Жыл бұрын
Sometimes we get videos having great explanation.
@olatheog
@olatheog Жыл бұрын
This is such a great video. I am just sad you did not end it with fitting a model and training after transforming as that is where I have problems. Is there another video of yours where you did that? I would really appreciate. Thank you
@rachittoshniwal
@rachittoshniwal Жыл бұрын
Thanks! I do have a couple of end to end project videos where I've fitted models after transforming. Hope they help!