Unbiased Estimators (Why n-1 ???) : Data Science Basics

Unbiased Estimators (Why n-1 ???) : Data Science Basics

ritvikmath

3 года назад

43,224 Просмотров

Ссылки и html тэги не поддерживаются


Комментарии:

Neelabh Choudhary
Neelabh Choudhary - 14.11.2023 04:07

dude. this is amazingly clear

Ответить
Bobby from Yakima
Bobby from Yakima - 12.11.2023 02:20

Ah. Learning is i the details. You just skipped over "not interesting" that permits the logic to flow. Not good, Even mentioning the names of the quoted formulas you used but not explain be helpful.... variance decomposition formula or the deviation square formula

Ответить
yepitsodex
yepitsodex - 23.10.2023 19:44

the 'we need it to be slightly smaller to make up for it being a sample and not the population' argument isnt needed or realistic. Having n-1, regardless of the size of the sample, says that the one is completely arbitrary just to tweak it the smallest amount. in reality when you go to the sample space from the population space, you lose exactly one degree of freedom. It seems like thats why its n - 1 and not n-2 or something else. if you had all of the sample space numbers except for one of them, the value of the last one would be fixed, because it has to average out to the sample variance. Since it cant be just anything, that is a loss of a degree of freedom, which justifies the use of n-1

Ответить
Alexander Smith
Alexander Smith - 20.10.2023 23:44

Hi @ritvikmath, I want to understand those derivations in the red brackets. Do you have a good set of sources that will explain to me why those three expected values return their respective formulae?

Ответить
Tyrone Frielinghaus
Tyrone Frielinghaus - 11.10.2023 08:27

Good intuitive explantation,,,thanksd

Ответить
Matthew
Matthew - 24.09.2023 02:01

I am reading a book on Jim Simons, who ran the Medallion fund. I’ve gone down the rabbit hole of Markov chains and this is an excellent tutorial. Thank you.

Ответить
Cadence Goveas
Cadence Goveas - 03.09.2023 09:46

been trying to understand this for weeks now, this video cleared it all up. THANK YOU :))

Ответить
Asif Shikari
Asif Shikari - 07.08.2023 06:21

Why n-1...we could adjust even better by doing n-2

Ответить
EkShunya
EkShunya - 05.07.2023 17:58

good one

Ответить
Nguyễn Kim Quang
Nguyễn Kim Quang - 26.03.2023 17:47

Thank you for great content!!!❤❤❤

Ответить
June Chu
June Chu - 05.03.2023 14:57

Thanks!! I love the way of saying "boost the variance."

Ответить
Pranav Jain
Pranav Jain - 06.02.2023 03:46

You are awesome

Ответить
Nelson K
Nelson K - 07.10.2022 13:10

you are GREAT

Ответить
Prof. G
Prof. G - 19.07.2022 07:21

incorrect intuition.
this is more accurate: ideally the actual sample mean equals the population mean, however the actual sample mean is rarely ideal and there's an error amount. if the sample is more concentrated on lower values, then the sample mean will be lower than the population mean. since the sample is concentrated on lower values and the sample mean is also lower, the differences between the samples and the sample mean will mostly be lower than the samples and the population mean thus lowering the sample variance. if the sample is instead concentrated on higher values, then the sample mean will be higher than the population mean. since the samples are concentrated on higher values and the sample mean is higher than the population mean, the distance between the samples and the sample mean will mostly be higher than the differences between the samples and the population mean thus lower the sample variance. whether the sample is concentrated on lower or higher values (not concentrated is unlikely for small sample sizes), the sample variance (using n as denominator) will prob be lower than the population variance. therefore, we need to add a correction factor.

Ответить
Jeff Bezos
Jeff Bezos - 11.05.2022 05:34

you're hired!

Ответить
Chinmay Bhalerao
Chinmay Bhalerao - 01.03.2022 13:37

I guess second approach for n-1 explanation will be right when both population and sample will follow same distribution which is very rare case.

Ответить
Yass theta
Yass theta - 12.02.2022 15:50

Now it makes total sense. Thank you 👏👍

Ответить
Soumik Dey
Soumik Dey - 25.01.2022 08:28

just wow!

Ответить
Martin W.
Martin W. - 06.01.2022 23:13

Great explanation! Love your videos.

Ответить
Pltt J
Pltt J - 29.12.2021 23:57

Thank you for the video, can you help me how to prove that is unbiased in this question? Question: Compare the average height of employees in Google with the average height in the United States, do you think it is an unbiased estimate? If not, how to prove it is not mathced?

Ответить
DSK Chakravarthy
DSK Chakravarthy - 26.12.2021 16:13

Please do one lesson on the concept of ESTIMATORs. It would be good if the basics of these ESTIMATORs is understood before getting into the concept of being BIASED or not. Anyways, you are doing extremely good and you way of explanation is simply superb. clap.. clap ..

Ответить
Abdusattor Sattarov
Abdusattor Sattarov - 01.12.2021 04:21

Great video, thanks!

Ответить
Jingsi Xu
Jingsi Xu - 05.10.2021 05:03

Thanks for the explaination from this perspective. Can u talk more about why 'n-1'? I remember there is something with the degree of freedom but I never fully understand that when I was learning it.

Ответить
miss ghani
miss ghani - 28.09.2021 13:59

this is how we can understand stats not by just throwing some number to students

Ответить
Syed Zain Zaidi
Syed Zain Zaidi - 15.09.2021 13:42

good stuff!

Ответить
stelun56
stelun56 - 03.08.2021 01:44

The lucidity of this explanation is commendable.

Ответить
Matthaeus Muniz
Matthaeus Muniz - 17.07.2021 10:07

tks, great explanation

Ответить
C
C - 14.07.2021 11:08

Great video but still not convinced on the intuition. How do you know that the adjustment compensates for missing tail in sampling? And if so, why not n-2, etc? I guess, if anywhere there would be missing data, it would be in the tail.

Ответить
Abrar Ahmed
Abrar Ahmed - 29.06.2021 02:35

Exactly what I have been looking for.

Ответить
Richard Chabu
Richard Chabu - 06.06.2021 22:52

well explained very clear to understand

Ответить
kunal singh
kunal singh - 24.03.2021 19:13

Thanks for the great explanation! But one question! why minus 1? Why not 2? I know the DoF concept would come over here! but all the explanation I have gone through, they have fixed the value of the mean so as to make the last sample not independent!
but in reality as we take samples the mean is not fixed! It is itself dependent on the value of the samples! then DoF would be number of samples itslef!

Ответить
Gianluca L'Episcopia
Gianluca L'Episcopia - 04.02.2021 21:43

Never understood why "data science" and not "statistics"

Ответить
Alex Combei
Alex Combei - 22.01.2021 18:09

<3

Ответить
Lrome_1
Lrome_1 - 09.12.2020 18:39

I wanted to ask a question. For E(x bar), x bar is calculated using a sample of size n, so is E(x bar) the average value of x bar over all samples of size n? Other than that, I think this has been one of the more informative videos on this topic. Additionally, many times people tie in the concept of degrees of freedom into this, but usually they show why you have n-1 degrees of freedom and then just say "that's why we divide by n-1", I understand why it's n-1 degrees of freedom, but not how that justifies dividing by n-1. I was wondering if you had any input on this?

Ответить
Too irRational
Too irRational - 17.10.2020 07:29

Bias is not the factor that is used to deside the best estimates...its Mean Squares Error...n-1 is used because error is low not because its unbiased

Ответить
David Szmul
David Szmul - 17.10.2020 02:33

In order to be even more practical, I would simply say that:
- Mean: You only need 1 value to estimate it. (Mean is the value itself)

- Variance: You need at least 2 values to estimate it. Indeed the variance estimates the propagation between values (the more variance, the more spreaded around the mean it is). It is impossible to get this propagation with only one value.

For me it is sufficient to explain practicaly why it is n for mean and n-1 for variance

Ответить
Ashish Kumar
Ashish Kumar - 15.10.2020 18:01

I believe this is the best channel I have discovered in a long time. Thanks man.

Ответить
!____!%___%
!____!%___% - 15.10.2020 17:50

Thank you. Could you please do a clip on Expected value and it's rules and how to derive some results.

Ответить
Loafclotnit
Loafclotnit - 15.10.2020 03:06

Quality video , keep it up !

Ответить
YITONG CHEN
YITONG CHEN - 15.10.2020 02:45

is that because of we lose 1 degree of freedom when we used the estimated mean to calculate the estimated variance?

Ответить
Christos Koutkos
Christos Koutkos - 15.10.2020 01:41

Try explaining the above ideas using the degrees of freedom.

Ответить
Jamie Walker
Jamie Walker - 15.10.2020 00:28

How I think about it: suppose you have n data points: x1, x2, x3, x4.., xn. We don't really know the population mean, so let's just pick the data point on our list which is closest to the sample mean, and use this to approximate the population mean. Say this is xi

We can then code the data, by subtracting xi from each element - but this doesn't affect any measure of spread (including the variance). But then after coding we will have a ist x1', x2', ...., xn' but the i'th position will be 0. Then only the other n-1 data points will contribute to the spread around the mean, so we should take the average of these n-1 square deviations.

Ответить
Ni999
Ni999 - 14.10.2020 23:49

That last blue equation looks more straightforward to me as -

= [n/(n-1)] [σ²-σ²/n]
=[σ²n/(n-1)] [1-1/n]
=σ²[(n-1)/(n-1)] = σ²

... but that's entirely my problem. :D

Anyway, great video, well done, many thanks!

PS - On the job we used to say that σ² came from the whole population, n, but s² comes from n-1 because we lost a degree of freedom when we sampled it. Not accurate but a good way to socialize the explanation.

Ответить
DistortedV12
DistortedV12 - 14.10.2020 22:39

I watch all your vids in my free time. Thanks for sharing!

Ответить
Thomas Kim
Thomas Kim - 14.10.2020 21:45

You still are not clear why we use n-1 instead n in the sample variance, intuitively.

Ответить
Muse Vanced
Muse Vanced - 14.10.2020 20:14

Great video. But anyone else feel unsatisfied with the intuitive explanation? I've read a better one.

When calculating the variance, the values we are using are x_i from 1 to n and x_bar. Supposedly, each of these values represents some important information that we want to include in our calculations. But, suppose we forget about the value x_n and consider JUST the values x_i from 1 to (n-1) and x_bar. It turns out we actually haven't lost any information!

This is because we know that x_bar is the average of x_i from 1 to n. We know all the data points except one, and we know the average of ALL of the data points, so we can easily recalculate the value of the lost data point. This logic applies not just for x_n. You can "forget" any individual data point and recalculate it if you know the average. Note that if you forget more than one data point, you can no longer recalculate them and you have indeed lost information. The takeaway is that when you have some values x_i from 1 to n and their average x_bar, exactly one of those values (whether its x_1 or x_50 or x_n or x_bar) is redundant.

The point of dividing by (n-1) is because instead of averaging over every data point, we want to average over every piece of new information.

And finally, what if we were somehow aware of the true population mean, μ, and decided to use μ instead of x_bar in our calculations? In this case, we would divide by n instead of (n-1), as there would be no redundancy in our values.

Ответить
Machiavelli
Machiavelli - 14.10.2020 19:37

What about n-2 or n-p, howcome more estimators we have the more we adjust? How does it exactly transfer intro calculation and ehat is the logic behind it?

Ответить