Three Sigma Rule in the empirical sciences express a conventional heuristic that “nearly all” values are taken to lie within three standard deviations of the mean, i.e. that it is empirically useful to treat 99.7% probability as “near certainty”.The rule states that even for non-normally distributed variables, at least 88.8% of cases should fall within properly-calculated three-sigma intervals. It follows from Chebyshev’s Inequality. For unimodal distributions, the probability of being within the interval is at least 95%. There may be certain assumptions for a distribution that force this probability to be at least 98%. These numerical values “68%, 95%, 99.7%” come from the cumulative distribution function of the normal distribution. The “68–95–99.7 rule” is often used to quickly get a rough probability estimate of something, given its standard deviation, if the population is assumed to be normal. It is also as a simple test for outliers if the population is assumed normal, and as a normality test if the population is potentially not normal. To use as a test for outliers or a normality test, one computes the size of deviations in terms of standard deviations and compares this to expected frequency.

Was the above useful? Please share with others on social media.

If you want to look for more information, check some free online courses available at coursera.org, edx.org or udemy.com.

Recommended reading list:

Data Science from Scratch: First Principles with Python Data science libraries, frameworks, modules, and toolkits are great for doing data science, but they’re also a good way to dive into the discipline without actually understanding data science. In this book, you’ll learn how many of the most fundamental data science tools and algorithms work by implementing them from scratch. If you have an aptitude for mathematics and some programming skills, author Joel Grus will help you get comfortable with the math and statistics at the core of data science, and with hacking skills you need to get started as a data scientist. Today’s messy glut of data holds answers to questions no one’s even thought to ask. This book provides you with the know-how to dig those answers out. Get a crash course in Python Learn the basics of linear algebra, statistics, and probability—and understand how and when they're used in data science Collect, explore, clean, munge, and manipulate data Dive into the fundamentals of machine learning Implement models such as k-nearest Neighbors, Naive Bayes, linear and logistic regression, decision trees, neural networks, and clustering Explore recommender systems, natural language processing, network analysis, MapReduce, and databases |