Law of Large Numbers is a principle of probability according to which the frequencies of events with the same likelihood of occurrence even out, given enough trials or instances. As the number of experiments increases, the actual ratio of outcomes will converge on the theoretical, or expected, a ratio of outcomes. For example, if a fair coin is tossed 1,000,000 times, the heads-to-tails ratio will be extremely close to 1:1. However, if the same coin is tossed only 10 times, the ratio will likely not be 1:1. The law of large numbers is sometimes referred to as the law of averages and generalized, mistakenly, to situations with too few trials to illustrate the law of large numbers. This error in logic is known as the gambler’s fallacy. If for example, someone tosses a fair coin and gets several heads in a row, that person might think that the next toss is more likely to come up tails than heads because they expect frequencies of outcomes to become equal. But, because each coin toss is an independent event, the true probabilities of the two outcomes are still equal for the next coin toss and any coin toss that might follow.
Was the above useful? Please share with others on social media.
Recommended reading list:
|Data Science from Scratch: First Principles with Python
Data science libraries, frameworks, modules, and toolkits are great for doing data science, but they’re also a good way to dive into the discipline without actually understanding data science. In this book, you’ll learn how many of the most fundamental data science tools and algorithms work by implementing them from scratch.
If you have an aptitude for mathematics and some programming skills, author Joel Grus will help you get comfortable with the math and statistics at the core of data science, and with hacking skills you need to get started as a data scientist. Today’s messy glut of data holds answers to questions no one’s even thought to ask. This book provides you with the know-how to dig those answers out.
Get a crash course in Python
Learn the basics of linear algebra, statistics, and probability—and understand how and when they're used in data science
Collect, explore, clean, munge, and manipulate data
Dive into the fundamentals of machine learning
Implement models such as k-nearest Neighbors, Naive Bayes, linear and logistic regression, decision trees, neural networks, and clustering
Explore recommender systems, natural language processing, network analysis, MapReduce, and databases