Explaining the Chi-squared test
12:38
Explaining the ANOVA and F-test
11:51
The better way to do statistics
17:25
Explaining Power
12:36
6 ай бұрын
What haunts statisticians at night
16:34
How to do a t-test in R
7:57
8 ай бұрын
Explaining The Two-Sample t-Test
8:36
Explaining The One-Sample t-Test
16:23
What do statisticians research?
17:26
Explaining Parametric Families
15:09
What is functional data analysis?
6:21
Explaining Probability Distributions
12:54
What is an N-of-1 trial?
4:30
11 ай бұрын
Пікірлер
@xenoduck3189
@xenoduck3189 12 сағат бұрын
Bayesan probability still has the same definition as frequentist probability!!! What you are showing is not a "definition" of probability, it is just Bayes' rule, which says NOTHING of P(A), only of P(B|A). The law of large numbers gives the definition of probability, regardless of what field of maths you study. I feel like this was really misrepresented in the video.
@TwentyNineJP
@TwentyNineJP 15 сағат бұрын
Finally I have the vocabulary to describe my philosophical objections to the way that the topic of statistics is often discussed. Probabilities have no place in a world of perfect knowledge; to a hypothetical god, all probabilities would be either 1 or 0, and nothing in between. It is only our ignorance of outcomes that gives meaning to statistics. FWIW the Bayesian approach is what I studied in signal analysis. I just didn't realize that the whole of statistics was bifurcated like this.
@tuongnguyen9391
@tuongnguyen9391 16 сағат бұрын
When I use machine learning algorithm to predict stuff, is it the bayesian way or the frequentist way ? or something between both or does it really depends on the data distribution or depend of the specific machine learning algorithm ?
@very-normal
@very-normal 15 сағат бұрын
I think it depends on the model. For prediction, I don’t think the distinction matters all that much. I don’t work a lot with prediction but this has been my experience But for inference, it changes how you do statistics and interpret results
@XxRiseagainstfanxX
@XxRiseagainstfanxX 16 сағат бұрын
Read "A history of mathematical statistics (from 1750 to 1930)" by Anders Hald
@very-normal
@very-normal 16 сағат бұрын
it’s a good book, you should also try Stigler’s History of Statistics too
@PerishingTar
@PerishingTar 16 сағат бұрын
You got my butt with that ad transition 😅
@very-normal
@very-normal 15 сағат бұрын
gottem
@GenericInternetter
@GenericInternetter 17 сағат бұрын
Not a statistician, but I do have a take on this... The Bayesian method relies on priors which hamstrings the whole practical purpose of analysis. Instead of debating results, people instead debate priors. It just shifts the whole thing from one frying pan to the other. The simplistic frequentist approach you described is utterly naive. You completely missed the whole concept of random walk. In practice, the most reliable approach to probability is the non-naive version with a large dataset, or a large set of datasets. Random walk is critical to understand for the frequentist approach to make any sense. For example, imagine flipping a balanced coin 4 times (small example, easier to explain) The naive approach would assume that larger datasets tend towards 50% heads, but this doesn't make sense. The probabilities are: 0% heads -> 1/16 25% heads -> 4/16 50% heads -> 6/16 75% heads -> 4/16 100% heads -> 1/16 It's a bell curve centered at 50%. With large data sets, your chance of getting the expected 50% result is only around 6/16, but your chances of getting either 25% or 75% is 8/16... Which means the naive approach is more likely to give an inaccurate result! Random Walk (results steering away) is a huge topic in itself and definitely needs to be accounted for to rely on the frequentist method.
@very-normal
@very-normal 16 сағат бұрын
how would it accounting for it help us understand frequentist methods any better
@yahlimelnik4483
@yahlimelnik4483 19 сағат бұрын
Damn dude, what is the frequency of you hitting the gym? Your arms are BIG
@very-normal
@very-normal 15 сағат бұрын
weekdays and ty lol
@hughobyrne2588
@hughobyrne2588 21 сағат бұрын
It seems like having factions that say "Pi is determined by geometry: the ratio of a circle's circumference to its diameter" and "Pi is determined by calculus: the limit of an infinite sum (pick your favourite)".
@very-normal
@very-normal 21 сағат бұрын
and then some people in one of the factions can’t stand it that the other one says it differently, I won’t say who
@davidarredondo2106
@davidarredondo2106 22 сағат бұрын
Excellent video!! I’m almost done with Bernoulli’s Fallacy myself. I do want to add that, for what I’ll call “reasonable” priors, the choice of prior doesn’t matter in the long run, as the data will dominate the posterior through the likelihood. Basically, with Bayesian statistics, we’ll find the truth if we just keep on collecting more data. Again, thanks for this great summary! I teach both a high school stats course and a high school Bayesian data science course, and this is the best short explanation of the difference I’ve seen. Congrats!
@q-tuber7034
@q-tuber7034 Күн бұрын
Today I learned: some people pronounce Bayesian like “beige” rather than “Bayes”
@very-normal
@very-normal Күн бұрын
beigians
@ckq
@ckq Күн бұрын
I'm a big predictions markets and forecasting guy so I essentially take Bayes for granted. Typically you look at past frequencies then apply a prior to it, which can get kinda subjective. Bayesian as a word seems so fancy (aka objective) I thought you could never be truly Bayesian. But really its just fancy terminology ppl say to look cool and not be a frequentist dummy who doesn't understand that randomness exists. I never really saw it as a conflict.
@BrentVis
@BrentVis Күн бұрын
Can you talk more about bootstrapping?
@very-normal
@very-normal Күн бұрын
I have an earlier video about it, but I think my better explanation is in my “biggest prize in statistics” video. It’s in the first chapter on Bradley Efron
@barttrudeau9237
@barttrudeau9237 Күн бұрын
I really enjoy your videos and style of teaching. I found this video especially well done and I learned a lot about a subject that's hard to understand (I'm an architect, not a mathematician, but I love this subject). I have been trying to learn statistical concepts for years and from this video I am starting to understand that if you are a mathematician focused on the computation and not knowledgeable about the subject you are studying, a frequentist approach may be more appealing and appropriate. If you are a subject matter expert trying to predict possible future outcomes, a Bayesian approach would be a better fit. You could say it caters to your prejudices, that's fair, but it also allows you to employee your expertise. So perhaps a Bayesian approach is more corruptible, but done properly it seems to have higher potential.
@very-normal
@very-normal Күн бұрын
Thanks, I’m glad they could be helpful to you. And I think you’re right, both methods have their use. Something I took out of this video was the fact that both methods are necessary in my space of biostatistics. Frequentist statistics are very desirable because of error rate control. If a medicine is risky but possibly useful, we want to be as sure as possible it works. Type-I errors are a different beast when humans are involved. But when we’re trying to look for new drugs and there’s millions of candidates to vet, Bayesian methods are a little better at this because posterior updates are still valid with multiple looks/analyses of the data without having to finagle with our level each time. Sometimes I think people skip my conclusion that you need both but it is what it is lol
@barttrudeau9237
@barttrudeau9237 Күн бұрын
@@very-normal I wish I could do both, I try, but I don't have the depth of mathematics education to do it properly. I do the best I can and keep learning every day. Thank you again so much for sharing your knowledge. It's really appreciated.
@Aldotronix
@Aldotronix Күн бұрын
So now we are backing up probability in the single biggest flaw in the history of science, the human bias?
@very-normal
@very-normal Күн бұрын
yup sounds about right
@gileneusz
@gileneusz Күн бұрын
good video, but I'm lost near 10:00
@matthewb2365
@matthewb2365 2 күн бұрын
Nonsense. The biggest beef in statistics is the thirders vs the halvers. #thirdersftw!!!!
@WylieThompson
@WylieThompson 2 күн бұрын
#stopbayesianhate
@craigparker1410
@craigparker1410 2 күн бұрын
Do you use Manim to make your visualizations ? I love how you work through the concepts and keep the canvas as clean as possible. Keep up the great work 🎉
@very-normal
@very-normal 2 күн бұрын
Yee i am a manim novice
@tuongnguyen9391
@tuongnguyen9391 16 сағат бұрын
@@very-normal Where to learn manim from your bayesian prior ?
@very-normal
@very-normal 15 сағат бұрын
I’m self taught from reading documentation but I’m aware of tutorial videos on KZbin
@iamdigory
@iamdigory 2 күн бұрын
There are frequintists? Really?, i thought they were a story told to scare children.
@very-normal
@very-normal 2 күн бұрын
they all up in the comments with their 200 word essays
@sarimbinfarooq2739
@sarimbinfarooq2739 2 күн бұрын
Hi, did your Masters had wetlab?
@very-normal
@very-normal 2 күн бұрын
nope! biostats is all code
@spiderchopproductions8172
@spiderchopproductions8172 2 күн бұрын
Frequentism? Oh you mean Nazi race science math. (Bayes 4 lyfe boiiiiiii)
@ucchi9829
@ucchi9829 2 күн бұрын
Please burn your copy of the bernoulli's fallacy...
@jonathanlivengood767
@jonathanlivengood767 2 күн бұрын
Please, I'm begging you, distinguish between [1] Bayes' Theorem, which is the thing you talk about at ~4:45 and which frequentists fully endorse (since it follows from the axioms, the ordinary definition of conditional probability, and classical logic), and [2] Bayes' Rule, properly so-called, which is a claim about how degrees of belief or confidences should be updated in light of evidence -- namely, that belief update should be by way of conditioning on one's evidence, i.e. Pr_new(h) = Pr_old(h|e), where e is the new evidence.
@very-normal
@very-normal 2 күн бұрын
semantics
@jonathanlivengood767
@jonathanlivengood767 2 күн бұрын
@@very-normal You say that like it's a bad thing. Surely a healthy science will have well-defined terminology that is designed to be broadly useful and not introduce unnecessary confusions. Right? Anyway, it's true that the labels are incidental, in a sense. They could be called Equation 1 and Equation 2 if you like. But they're also terms with a history, and they are part of the longstanding dispute that your video is ostensibly about. Given that context, it seems important to be extra careful about the terminology. Moreover, the *distinction* has practical implications, even if the labels don't. Frequentists can happily accept Bayes' Theorem and reject Bayes' Rule. In fact, some historically important defenders of a personalist interpretation of probability, e.g. Ramsey, have rejected Bayes' Rule (of conditionalization), but of course, they don't reject Bayes Theorem. Being clear about what you take a Bayesian to be committed to is important for understanding the debate.
@ucchi9829
@ucchi9829 2 күн бұрын
@@very-normal I want to be charitable and say you're just trolling.
@danielkeliger5514
@danielkeliger5514 2 күн бұрын
Yes, I totally agree that for large sample sizes the two methods basically give the same asnwer. They are also compatible in the sense that Bayesians have their own interpetarion for maximum likelihood and Bayesion methods can be analysed via frequentist language. (Infact it is more natural to understand the limit theorem mention in the video in frequentist terms, in my opinion.) Still, I want to make some remarks. Firstly, I’m sort of a prularist. I don’t think probability stands for a single concept. Statements like “what is the probability that this previously unknowm sonnet was written by Shakespeare” can be interpeted in a Bayesian way much more generally, while physical problems (see belowe) makes more sense in the frequentist interpetation. Ultimately, there are many things that satisfied by the Kolmogorov axioms that has nothing to do with randomness. (Say the ratio of votes in an election.) It is possible to do probability theory without reffering to randomness at all. There are cases when we do actually talk about frequencies in the world. Ergodicity is s good example. Saying things like “if I know the exact inital conditions, I can calculate the exact ratio of times the coins will land on head” and therefore “probabilites are purely epistemic” kind of misses the point. I’m not interested in this very particular initial condition. I want to show that this behaviour when roughly 1/2 of the coin tosses lands on head is typical for most of the initial conditions. This 1/2 number is a property of the system, and it doesn’t describe the mental state of an idealised, rational observer. (With the obvious objection that of cource observations themselves are model dependent, etc.) Lastly, all the populat interpetations have their own philosophycal problems. I don’t know any interpetation that are not ultimately flawed under greater scrutiny. This is actually very typical when it comes to phylosophical problems. (Think about all the different schools of ethics.) I think I like the propensity interpetation of probability the most, but that is not perfect either.
@QuandaleDingle-bq1on
@QuandaleDingle-bq1on 2 күн бұрын
Bayesian propaganda 😂
@very-normal
@very-normal 2 күн бұрын
propaganda for great posteriors
@christophergame7977
@christophergame7977 2 күн бұрын
I think you haven't really got the essence of Bayesian theory. You write such things as P(A). That's against the basis of the Bayesian approach. The basis is that you must state specifically the data and grounds on which you rely for the the value of your probability. That means that you should write only such things as P(A|θ), where A denotes the event as usual, and θ denotes the data and grounds on which you rely for the value of your probability. For example, A could denote 'heads' in a single toss, and θ could denote the proposition that the coin is a fair coin so that only heads in a single toss 'A' and taiis in a single toss 'B' are possible and P(A|θ) = P(B|θ). It's called transparency, and just requires you to state you position explicitly. So-called 'Bayes Theorem' isn't the essence of the approach; it's just a theorem that is true for frequentists as well. The essence is the requirement for explicit transparency. The Bayesian approach has nothing to do with subjectivity. The allegation of subjectivity is just frequentist rhetoric intended to badmouth the Bayesian approach. Perhaps I have been unfair in the foregoing by under-recognising what you said, but I ask for a harder line on the inadmissibility of such expressions as P(A).
@cube2fox
@cube2fox 2 күн бұрын
For a subjective Bayesian, P describes his beliefs at a particular point in time, P(A) describes his belief in A, and P(A|B) describes what his belief in A under the assumption that B. So we can see, unconditional probabilities cannot be replaced by conditional ones, since they denote different things, and subjective Bayesians have no problems with unconditional probabilities. Conditional probabilities merely cover hypothetical reasoning.
@christophergame7977
@christophergame7977 2 күн бұрын
@@cube2fox Thank you for your response. I think it a mistake to ever allow the writing of such things as P(A). I think that the notation should be amended to outlaw such things as P(A) and to mandate such things as P(A|θ). In other words, if you insist on calling P(A|θ) a conditional probability, I say that the every probablility should be considered conditional on θ, which must be explicitly specified. I think that the notation should be amended to outlaw such things as P(A) and to mandate such things as P(A|θ). I think that such is the essence of the Bayesian approach. I guess that makes me a non-subjective Bayesian. I am demanding of the Bayesian that he specify the data and grounds every time. No hidden stuff such as in P(A).
@cube2fox
@cube2fox Күн бұрын
@@christophergame7977 Even if the data is specified every time, there is still "hidden stuff". Even when two people agree on P(A) and P(B), they can well have different values for P(A|B).
@christophergame7977
@christophergame7977 Күн бұрын
@@cube2fox While you insist on writing P(A) (as opposed to the reasonable P(A|θ)), you will run into messes.
@insertalias3587
@insertalias3587 2 күн бұрын
Great Video, thank you! Regarding 17:05: Would you have some references for the "some" that "have argued that the bootstrap form something similar" would be quite applicable for a current project of mine
@seriousbusiness2293
@seriousbusiness2293 2 күн бұрын
It's very arguable if the idea of probability itself is a fundamentally real thing. Like as a mini example the digits of pi behave random by every single metric we know, yet they are determenistic and nothing random is happening. The ultimate goal of probability is modeling unknown outcomes and that can be done in many ways. So there is no true right option, all we care for is how accurate we can predict things and how interpretable it is to us. (ps in my eyes Bayesian feels more true to real life and my thinking)
@weetabixharry
@weetabixharry Күн бұрын
I'm not sure what you mean by "real" here. Casinos make real profits. A digit of pi, selected at random, (it is believed, but not proven) has an equal probability of being any number. Meanwhile, a digit of 50/99 (in base 10), selected at random, will be either 0 or 5 with equal probability. These things seem real to me.
@seriousbusiness2293
@seriousbusiness2293 Күн бұрын
@@weetabixharry I meant the sequence is random but deterministic, if you pick random digits you introduce other randomness. My thinking is multiple things appear to us as something random but if we knew the underlying dynamics we could often agree that probability theory is the wrong approach. Lets imagine an event i can only measure a single time like "Alex immediately says yes if ask him on a date today.", the idea of doing repeated trials is not real unless i have access to parallel universes, and taking other variables into account to refine my guess like comparing with other people i asked gives confidence but doesn't fundamentally reflect Alex choice then. Even if we measured every atom interaction in Alex brain we get into discussions of quantum and chaos theories. So even if our best models say the probability was 50% we cant tangibly experience or measure that 50% since we only see one outcome.
@aakashparida2026
@aakashparida2026 2 күн бұрын
New found love for statistics....Thank you so much!!
@__-de6he
@__-de6he 2 күн бұрын
I guess ,probability is derived from geometric property of our microspace (like general relative theory is derived from timespace geometry). So frequentist approach is more relevant.
@Shantanu_Dixit
@Shantanu_Dixit 2 күн бұрын
You just gained a subscriber love your content 🌿🌿🌿
@John-zz6fz
@John-zz6fz 2 күн бұрын
One of the advantages of the Bayesian approach is it feels more "natural" to incorporate non-quantitative evidence into your calculations. For example it's pretty easy in a frequentists analysis to calculate the odds of rolling a 6 on a d6 if you have rolled it a hundred times and can see the distribution of prior outcomes (fair or loaded). If instead you tell them we have no prior data but there's double sided tape on the 1 side you can easily "swag" that prior with a Bayesian approach and get better results. I've never actually seen a calculable advantage to either view but if you start fudging some numbers using a frequentist approach it just feel like you are doing something wrong... I don't actually think there is a difference or if there is I clearly don't understand it.
@Ethan13371
@Ethan13371 2 күн бұрын
One option for avoiding Bayesian prior-rigging is to simply publish the calculation itself, lacking any prior. Then, the probability is simply some function of the prior, which you could graph or something to visualize it better
@andrewharrison8436
@andrewharrison8436 2 күн бұрын
That seems a really interesting approach - as much for the psychology as for the mathematics. It might make it easier to soften a dogmatic prior into an introspection on the evidence/knowlege/belief that underlay the prior.
@danielkeliger5514
@danielkeliger5514 2 күн бұрын
It is kind of what is the frequentist interpetation of Bayesian methods. They are “just a fancy estimate” for which you can proove things like asymltotic unbiasedness and normality, etc. Fun fact, but apart from degenerate cases, Bayesian estimates are not unbiased.
@piwi2005
@piwi2005 2 күн бұрын
As usual for a bayesian video, there is much bias towards complexification. First, the test should be one sided: you requested at least 85%, so please have the courtesy to do the correct one. That divides p value by 2 from scratch. Then, you do not need a confidence interval at all, you have p value. What the test tells you is that from a sample of 1074 people, there was a probability of 0.27% to get the data you got if anyone was puting 4 or 5 stars _less_ than 85% of the time (by the way, this is how you got the 99.7% "that only Bayesian gives you", supposedly....). This is the frequentist approach, and it deals with facts and makes two assumptions: independance of choices from users and validity of CLT. Then from that p value, you can do what you want, you are not even obliged to do anything, because so far you only collected data and did maths. Once the computation is done, you can _finally_ go philosophical and decide you do not live in a universe where you got unlucky to be in the 0.27%. There is no binomial, no beta, no prior, no "I don't have an idea of my prior, so I'll use uniform distribution but I will call it Beta(1,1)", no some god of philosophy told me that "no idea" meant the existence of a uniform distribution in the realm of ideas, etc... Frequentist works with facts and try, at least when they're not psychologists or marketers, to be rigorous, not forgetting the assumptions they made. They uses stats to falsify theories, and they don't put probabilities on theories which rermain true or false. Bayesians do decision making, using a tool that always works, always getting an answer whatever was the question they had. It is very good for investors who want to use some maths and have a magical tool that allow them to propose a strategy with some appearance of seriousness, and it will work whenever they were lucky with their priors. But at the end of the day, either your posterior "probability" depends a lot on your priors, and you only put a number on your feelings, or it doesn't and you didn't need to go Bayesian. Frequentists don't deal with philosophy. Bayesian do and must.
@very-normal
@very-normal 2 күн бұрын
wow
@mousev1093
@mousev1093 18 сағат бұрын
It's a shame this isn't higher but that's probably to be expected in a channel/comment section so heavily biased to one approach. The number one suggested video to follow this is literally called "the better way to do statistics" His entire interpretation of the "frequentist perspective" was purposefully limited and he tried to divorce it from reality and naturally occurring events. I'd go as far and argue that his interpretation of how to report a confidence interval was bordering on incorrect. It can, and should be, phrased practically identically to the way he talked about credibility intervals later. The entire point is that you can't know something perfectly to arbitrary confidence and the estimation of true probability can only be refined. A confidence interval is the way of quantifying this spread of uncertainty. He even contradicted himself on the definition of "repeated experiment". First he defines experiments as events that produce individual data points and then he's purposely obtuse and redefines repeating the experiment to gather another 1074 reviews. Really should have partnered with someone else to present the other side. An entire video with straw men is boring
@very-normal
@very-normal 18 сағат бұрын
feel free to make that video with more correct frequentist teachings, more good statistics videos wouldn’t hurt on KZbin
@mousev1093
@mousev1093 18 сағат бұрын
@@very-normal I think that might hurt my current content algorithm stuff yknow ;)
@Sammysapphira
@Sammysapphira 2 күн бұрын
Frankly a ridiculous theorem. Human interpretation should have no merit in the actual probability of an event; otherwise simple probabilistic tricks such as the Monty Hall Problem would never be solved. "Frequentist" interpretation doesn't mean you need to throw all context related to the event to the side and calculate it in a vacuum. That is simply bad definitions of statistical problems.
@very-normal
@very-normal 2 күн бұрын
lol
@localidiot4078
@localidiot4078 3 күн бұрын
Im not a statiatician, but it feels like you are strawmanning their argument. Instead of a second 1000 review experiment, you could have made 2 different selections of half and you do have 2 different experiments. Thw repeated experiment is the 1000 different reviews. It makes you lose credibility when you do this.
@very-normal
@very-normal 3 күн бұрын
which part was the straw man
@localidiot4078
@localidiot4078 2 күн бұрын
@@very-normalso when you were explaining frequentist approach to statistics, you said something like, we have 100 reviews, my math says i shouldnt go here based on those 100 reviews, if i repeated the experiment, i.e. got another 100 reviews, I should have the same result. but wouldnt a frequentist think that every individual review was an experiment? the probability that the restaurant is good be the ratio between 4-5 star reviews over total reviews, not between collections of 100 reviews over the number of those collections. when frames that way it makes frequentists sound stupid, but like thats cause the framing is bad.
@very-normal
@very-normal 2 күн бұрын
Yeah you’re right, each person is thought to be an independent observation, but that’s not the only thing frequentist statistics count on. It’s the test statistic that needs to be replicated in multiple experiments. The test statistic depends on the sample size being the same in each repeated experiment, that’s why frequentist statistics have problems with repeated looks at the data. They’re looking a fundamentally different p-values if the sample size changes
@ucchi9829
@ucchi9829 2 күн бұрын
​@@very-normalYou can do Frequentist inference without repeated sampling actually.
@very-normal
@very-normal 2 күн бұрын
what method are you referring to
@jkid1134
@jkid1134 3 күн бұрын
Oh boy, we're gonna talk about Laplace's rule of succession! (<- wrong) Anyway. I really don't like the underlying assumption that all the data is equally relevant - maybe, for example, the cafe changed beans six months ago, and 50 of your 1000 reviews are very important, and 950 are pulling the review in an unrelated direction - is there an approach that would be able to infer things like this mathematically, or is it data weighting always part of the input? When I think about trying to pull valuable metadata from, say, a survey, it seems like it would be critically important not to let filtering and weighting be human decisions.
@very-normal
@very-normal 3 күн бұрын
That’s a good question. I’m aware of Bayesian models that take into account weighing of the observations, but I’m not too familiar with the method. From what I understand, a way to handle it is though a hierarchical model that incorporates the weights
@tunneloflight
@tunneloflight 3 күн бұрын
Btw - my arguments with statistics do not mean they are useless. Rather, they are frequently and all too easily abused (intentionally or unintentionally). In every instance, the use of statistics as applied to real world analyses must be critically analyzed and scrutinized, starting with the assumptions, presumptions and desires and relative ignorance of those involved. Even when all of that is fair, statistics often goes wildly away from truth or reality. And researchers often fail to apply even the most basic critical analyses to the results. Is the population a single population? Is the population linearly, triangularly, normally, poison or other distributed? Are there hidden variables? Is the data the result of stochastic events acting on stochastic events? Do the results violate sanity? Do the results suggest results outside the bounds of the analysis? Is the thesis or hypothesis that resulted in the data gathered biased in its own rights? Etc...
@tunneloflight
@tunneloflight 3 күн бұрын
Both frequently lead to bogus answers by never understanding the problem in the first place (in dozens of different ways), and by wrongly conflating chance, risk, and probability of many different flavors, and by presuming assumptions that have no basis in reality. Lies, damnable lies, and statistics - the most damnable lies of all. Beyond all of this, statistical analysis is subject to thousands of human flaws in cognition, in emotion, in biases, and more. Statistical analysis often then gets subsumed into bogus hierarchical fault tree analysis that pile error upon error, and/or supporting bogus and erroneous multi-function weighted attribute analysis, and/or conflating cost-benefit analysis as dispassionate, fair and/or meaningful, and/or truncating high-consequence low-probability events, and/or utterly missing consequence analysis and vulnerability assessment, and/or utterly failing to understand and properly apply "Safety Culture", and wrongly substituting a wildly erroneous belief that a misunderstanding about the meanings of the two words is a substitute fir understanding and properly application of the phrase. Oh yes, AND confusing or reversing cause and effect, or assuming causal linkage when no linkage exists. Said more simply: a pox on both.
@very-normal
@very-normal 3 күн бұрын
i was waiting for someone to bring up the “lies” quote, congrats on being the first here 🥇🎊
@tunneloflight
@tunneloflight 3 күн бұрын
@@very-normal It isn't and wasn't simply a matter of a quote. Having spent an entire career in engineering and science, I had the opportunity to watch as statistics was repeatedly abused as a replacement for wisdom and reality, often concealing powerful truths. Among these, I got to watch in real-time the development and spread of p-hacking in hundreds of wilful and inadvertent ways. I also got to see risk assessments pruned of true errors to focus on central value - which no meaning at all, and to see cases where cause and effect were reversed for those to allow maximal harm, while asserting little to no harm, etc... In all of that, I came to appreciate just how true that aphorism is. Add to this the hierarchical approach in science and academia, where an Italian researcher over three decades ago defined uncertainty in an analysis if a real system to be numerical variation in a chosen model. I cannot speak to whether this was a logical or intentional error, or simple ignorance. What ever it was, by being the first major author in the field, other researchers have relied on that error to narrow their analysis of unrelated subjects to achieve desired results (whether they understood the invalidity of that or not). Meanwhile other experts attempting to reach truer results using tools like the Jupiter Suite are repeatedly rejected as they broaden the distributions, while the researchers desire is to (wrongly) narrow their analysis distribution. Bayesian analysis can bring new insights. More often in my experience they serve to obfuscate the relationships resulting in embedded hidden errors being buried in the process. Also add to this the rejection of other tools, like "inconfidence" analysis that more easily identify vulnerabilities in the analyses.
@tunneloflight
@tunneloflight 3 күн бұрын
"These concocted analyses don't get very far." Quite to the contrary in my experience they rule the roost, and are lost to most analysts in the complexity or obtuseness of the analysis.
@ejrupp9555
@ejrupp9555 3 күн бұрын
So when you say the probability of this happening, you are a frequentist and if you say it you are bayesian.
@personal-qs6dz
@personal-qs6dz 3 күн бұрын
The "subjective" thing is entirely your invention. I've never heard of inventing the prior probabilities when using Bayes' rule. You compute the probabilities as best as you can and then use those in Bayes' formula.
@ejrupp9555
@ejrupp9555 3 күн бұрын
If versus when. One says 'if' this happens - the other 'when' this happens. If, is more quantized superposition, while when & superposition are incompatible.
@very-normal
@very-normal 3 күн бұрын
now you’ve heard of it
@12nites
@12nites 3 күн бұрын
You did a two-sided test when a one-sided test was needed. The actual pvalue was a half of what you got. Your H0 should have been pi<0.85 and H1 should've been pi>=.85
@very-normal
@very-normal 3 күн бұрын
lol even if the pvalue would have been halved, it doesn’t change the conclusion
@peterbonucci9661
@peterbonucci9661 3 күн бұрын
My statistics teacher refused to teach Bayesian inference because you got a supposedly improved probably estimate with no additional information. Information theory is important in my work. Generally, a frequencist approach works for me. Adding information that isn't there by using an a priori model can be a problem.
@madeofmarble8514
@madeofmarble8514 3 күн бұрын
choosing not to teach bayesian inference at all is wild
@sriharsha580
@sriharsha580 3 күн бұрын
"C.I does n't tell us if it contains the true value of PI or not, you can only know that if you repeated the experiment multiple times then most of them will. " Can you explain this statement I didn't get it.
@very-normal
@very-normal 3 күн бұрын
The definition of confidence is the proportion of intervals that contain the true parameter value. Different experiment repetitions will produce different datasets, so the ends of the intervals will change depending on the data. In the same way that choosing a 5% level means you only get a type-I error in 5% of experiments, the confidence interval will contain/cover the value of the true parameter in 95% of experiments. There’s no guarantee that you know the one you calculated actually contains it or not
@adnankhrais3207
@adnankhrais3207 3 күн бұрын
the only thing i didn't understand is the 0.05 i mean why we chose it ? what is special about 0.05 exactly? is it used for all problems or it can be calculated to fit the situation better ? using one constant assumption for all situations does't make sense i hope anyone answer my question
@very-normal
@very-normal 3 күн бұрын
Nothing is inherently special about 5%, it just happens to be pretty small. It’s often chosen based on how safe or dangerous a Type-I error is. If a Type-I error is safe, then we might even tolerate a higher rate like 10%. But if we really want to decrease the chance of one happening if it’s really dangerous, we could go for something like 1%.
@f1f1s
@f1f1s 3 күн бұрын
The idea of a repeated experiment showcases the inherent variability in the parameter estimate, i.e. the sampling distribution. A frequentist assumes that there could be a different data set borne by the same invisible data-generating process (law). Bayesians tend to jump onto data matrices as if those n=200 observations were the one and only realisation possible, without other hypothetical scenarios occurring, as if there were no Heisenberg principle or quantum uncertainty. The frequentist approach reflects the randomness of Nature and unobservability of hypothetical outcomes better: ‘it could have been otherwise’. Finally, Bayesians often make ridiculous distributional claims: ‘assuming the prior normal distribution, the posterior distribution of the linear regression slope estimator is precisely Student with n=198 degrees of freedom', whilst frequentists are much more careful about heteroskedasticity, calibration, coverage probability, and Bartlett correction, which are essential to control the false discovery rate: ‘there is some unknown law, but we can compute some functionals thereof regardless of the joint and marginal distributions, as long as enough finite moments exist for the WLLN and CLT to work’.
@ucchi9829
@ucchi9829 2 күн бұрын
Finally, a Frequentist defense.
@Acbelable
@Acbelable 3 күн бұрын
O love this guy
@Zxymr
@Zxymr 3 күн бұрын
I told my Asian parents that I was Bayesian. They disowned me.
@nunkatsu
@nunkatsu 21 сағат бұрын
Dude, worst time possible for me to read that comment. I just found out that my Asian crush (who reminds me of my Asian ex) at statistics classes in college has a boyfriend. Everything you wrote gave me PTSD.
@Skeleman
@Skeleman 3 күн бұрын
To understand the beta distribution: 1. Imagine you are sending rockets to an alien planet to see what portion of the surface is covered in water. 2. You can send probes that hit the ground, are destroyed instantly, but send back whether they landed on water or land. 3. Let's say you send down a probe and it says it hit land. 4. If the entire planet were water, there is a 0% chance the probe would say land. If the planet were 100% land, there is a 100% chance of it happening. If you plot the percent of land on the planet on the x axis, and the relative probability that the probe says land, you get a line from 0,0 to 1,1. You can image the opposite being true if it said water: a line from 0,1 to 1,0 5. If you send another probe and it says water, then you can combine two plots. multiply the land plot by the water plot because at each possible percent of land on the planet, the probabilities are being combined. You'll end up with a parabola. 6. Keep multiplying the right plot if the probe says water or land, slowly you'll get a bell curve where the peak is at the ratio of land and water probes. This is what the beta distribution is.
@ClementinesmWTF
@ClementinesmWTF Күн бұрын
Except that this interpretation only works for positive integer parameters α, β, whereas the full scope of the distribution works for any positive real numbers α, β. An actual description of the beta distribution would require a bit more complicated “probe sampling” than described here-having troubling thinking up any alterations to the described example at 3a tho.
@zak3744
@zak3744 3 күн бұрын
20:04 - "You can have strange priors, but you're going to have to justify them with evidence." But in that case, they're not really a subjective prior at all, are they? If they're properly evidence-based, then they're objective, surely? And in that case, the fundamental objective/subjective difference that you'd previously described is no longer there. An objective prior followed by an objective analysis gives an objective result, and a subjective prior followed by an objective analysis gives a result that is to some extent not evidence-based!
@Skeleman
@Skeleman 3 күн бұрын
If anyone is interested in if there is an objective way to pick a prior probability distribution, you do it with something called "maximum entropy". And the entropy they refer to is the same one the physicists talk about.
@danielkeliger5514
@danielkeliger5514 2 күн бұрын
I disagree. In the case of the p parameter for Bernoulli would be the uniform distribution. That is, however, depends on the coordite system you choose, as opposied to other methods like Jefferey’s prior. Maximal entropy arguments in general rely on some assumption of a unoform distributions even in physics. (Think about the whole combinatoric derivation with Stirling’s formula.) Ultimately, all models depends on assumptions let it be frequentist or Bayesion. There is no such thing as “purely leting the data talking for itself”.
@Skeleman
@Skeleman 2 күн бұрын
@@danielkeliger5514 I agree that there is never a way to "let the data talk for itself". I think i misused the term "objective". There are reasons to use the maxent distribution to ensure you aren't adding any "hidden" assumptions to your analysis.
@danielkeliger5514
@danielkeliger5514 2 күн бұрын
@@Skeleman I totally agree that uninformative priors are great tools for mitigating subjectivity. I just don’t belive in logical positivism :)