Machine Learning Lecture 29 "Decision Trees / Regression Trees" -Cornell CS4780 SP17

Machine Learning Lecture 29 "Decision Trees / Regression Trees" -Cornell CS4780 SP17

Kilian Weinberger

5 лет назад

41,621 Просмотров

Ссылки и html тэги не поддерживаются


Комментарии:

Dominik Mazur
Dominik Mazur - 05.11.2023 01:50

Thank you 🙏

Ответить
Rahul Seetharaman
Rahul Seetharaman - 08.01.2023 19:13

Why do we do a weighted sum of the entropies ? What is the intuition behind weighting them and not simply adding the entropies of the splits ?

Ответить
Greg Makov
Greg Makov - 08.06.2022 16:50

giang bai ma long ghep tum lum het nha :D:D:D met ca nup lum bat bo tu bi gio :D:D:D thay hu qua diiii

Ответить
yashwardhan CHAUDHURI
yashwardhan CHAUDHURI - 22.05.2022 14:38

Hi can anyone please explain why equally likely events are a problem in decision trees? What I understood from it was that the model will need to be very comprehensive to tackle such cases but I am unsure of my insight.

Ответить
Jalil Sharafi
Jalil Sharafi - 21.12.2021 06:17

I'm watching this end of December 2021, I found the demos at the end starting roughly at 45 mins in the video very informative about the capabilities and limitations of a decision tree. Thanks.

Ответить
R. A.
R. A. - 29.10.2021 13:52

Thank you for sharing your content.

It is very interesting. Especially the discussion about why we do this ( computational problems, NP-hard, people tried many splits and found out it was the best in practice), the interactive examples at the end (very useful for learning) and all your work on trying to make it clear and simple. I like the point of view of minimizing the entropy from maximum the KL between two probability distributions.

In fact, it is also easy to see the Gini impurity loss function as an optimization problem in 3D also (you get a concave/convex function by computing the hessian matrix with two parameters as the third one is just 1 - p_1 - p_2) and you have to optimize it on a space (conditions on the p_i) and you can actually draw the function and the space. You get the maximum/minimum at 1/3 for p_1, p_2, p_3 (what we don't want) and it is diminishing as we move away this point (with the the best case for one which is 1 and the others 0).

Ответить
Orkun Aran
Orkun Aran - 27.07.2021 19:09

This is perfect. I am coming from a non-technical, non-math background; and this presentation really made me understand DT easily. Thank you very much.

Ответить
Shay Williams
Shay Williams - 26.04.2021 13:11

Forgive me if I'm wrong but if a pure leaf node with 3 classes that results in P1=1, P2=0, P3=0, the sum of Pk*log(Pk) would be 0, so the idea would be to minimize from the positive entropy equation?

Ответить
Kaushal Kishore
Kaushal Kishore - 18.04.2021 21:54

Pk is zero means k is infinity how it is possible, Q at 39.00

Ответить
Yunpeng Tai
Yunpeng Tai - 22.03.2021 13:41

I have some problems about regression.I wonder if I can discuss them with you.

Ответить
Utkarsh Trehan
Utkarsh Trehan - 29.01.2021 05:33

Machine Learning ~ Compression 💡

Ответить
Varun Jindal
Varun Jindal - 25.12.2020 11:20

Thanks Professor Kilian Weinberger. Examples in the end was really helpful to actually visualize how trees can look like.

Ответить
tran le
tran le - 19.10.2020 04:36

Thank you very much, Prof. Weinberger. I was reading The Elements of statistical Learning as my reading course, then I found your channel. I truly appreciate your lectures also your notes, I print all of your notes and watch your almost all of your videos, they are extremely helpful. Thank you, I really appreciate that you let us have access to your wonderful lectures.

Ответить
Amir Mohajerani
Amir Mohajerani - 18.09.2020 04:41

This was amazing. Thank you very much

Ответить