Комментарии:
BAAAMM!!! JOSH
ОтветитьHey Josh , where is the part 1 video of decision tree ... Can you please give me the link
ОтветитьHi Josh. I am going over each of Your lessons in time-of-upload order (from oldest to newest) and I wonder why this lesson is called "Part 2" while earlier there was no Part 1. I assume that Part 1 is "Decision and Classification Trees, Clearly Explained!!!" uploaded after this video, correct? BTW: excellent work!
ОтветитьThanks!
ОтветитьThank you!
ОтветитьWhere is the part 1 of this video?. I couldn't find it.
ОтветитьVery good job making these videos, I can image how much time you spend on it. Wonderful job.
Ответить🏆
ОтветитьHello :) , thank you so much for this amazing video!
I have a query: to handle missing data, you used the column whose values is most correlated to it, as a guide. But generally, we drop those columns from our dataset and avoid high correlation due to its negative impact on model prediction. Should we really follow this practice?
Teacher Josh I have one question: if we apply feature selection techniques say Filter method ,wrapper method, embedded method on our dataset we may get different result from each method which features are relevant or which are not. Overall how we actually make evaluation analysing result coming from each method that which features should be chosen? if features and target variables both are numeric, hope you got my point
ОтветитьAwesome video! Thank you Josh. Would be very useful if you can make a video about "double dipping" after feature selection in random forest or machine learning in general?
ОтветитьDo features have to be stationary when applying ML models to time series data? Or any data for that matter?
Ответить2022 01 09
ОтветитьCan you make a video on how surrogate split is used in decision trees for handling missing data and computing feature importance?
ОтветитьHello!
First we will perform EDA,FEATURE ENGINEERING and then MODEL BUILDING right? so, my question is in EDA AND FEATURENGEERING we will handle missing values right?
anyway we'll handled missing values there, even after that will we get missing values?
You are genius!Respect from china
ОтветитьHey Man , U make ML look like a cake walk. Great work....
AND I am loving ur theme songs, so I am awesome😂 and decision tree algo is proving it🤣
Please make catboost video, i'm working on my thesis :')
ОтветитьI love the way you start your channel🤩
ОтветитьI love the intro.! Just so unique every time. :)
ОтветитьCan we use a decision tree or random forest to impute missing value? Great work as always!
ОтветитьYOU are the BEST!!!!!!!!
ОтветитьHi Prof Josh,
Suppose, I already have the selected features by using random forest algorithm. Then, I use this features for PLS-DA. Will the model I build by PLS-DA more valid?
Thanks
Sir, is it the same between missing data and outliers , noisy data ?
Ответитьselecting a second best gini impurity will reduce over fitting?
Ответитьhow dose feature selection helps in overcome of overfitting, please explain I'm not getting this?
ОтветитьHi Josh, I am a big fan!
I'd just like to ask something as I'm still in the midst of learning. In multiple linear regression, we are taught that multicollinearity is a big issue and a red flag. However, here it is mentioned that it can be used as a way to fill in missing data, if the missing data is of a variable that is highly correlated with another one that is known.
Is it because they are different models and thus the issue doesn't apply here?
Thanks, once again, a big fan!
I should redirect my tuition fees to Josh Starmer, because he deserves it more than my university.
ОтветитьClearly explained. but how to measure the correlation btw two binary columns.
ОтветитьYelling "oh no!!!" over and over is pretty irritating. I don't mind the other repeated yells as much because at least they aren't infantilizing... I get that you're trying to be relatable and stylized, but this is not pleasant.
ОтветитьI got StatQuest!
Ответитьplease can explain main difference between ID3,CHAID,CART
ОтветитьIs this how feature importances are assigned? Can you elaborate a little on this?
ОтветитьHello Josh!! I was thinking in a manner to reduce false negative in diagnosis, so is there some parameter to control the number of false negatives outcomes in a decision tree??
ОтветитьThis is the best statquest song so far.
ОтветитьYou are awesome Josh!!
ОтветитьJosh I am a huge fan of your videos!! You helped me understand all those complex ML concepts better than one year in grad school... I wonder if you can make some videos about how feature importances of random forest are calculated and DBSCAN clustering works (and how its parameters are chosen). Thank you so much!!
ОтветитьThank you!
ОтветитьHi, will you be able to do a video on how to numerically calculate the Random Forest Feature Importances? I couldn't find clear explanation anywhere on internet...It would be really appreciated!
ОтветитьNext level starting balad
ОтветитьYour the best ! thanks!
ОтветитьI love you Josh, you and your intros.
ОтветитьGreat video as always!!!
ОтветитьHi Sir, Josh Starmer... I hope you are good. Kindly make a video on pre pruning and post pruning. I have seen your videos and Information Gain 3 is also missing from the series.. Thanks in anticipation..Have a good day.
Ответитьif weight is highly correlated with height, why not remove weight column?
Ответить