Big data: Why does it matter?

You might be surprised to learn that the theory behind many AI applications we see today was developed a long time ago. Here are a few key milestones:

So why is AI only becoming such a big deal recently? The answer comes down to two things: computing power and data. We now have access to huge amounts of data (big data) and the computing power needed to process it. In fact, the rise in computing power is largely due to the need to process this growing data.

Why is Big Data So Important?

To understand why big data matters, think about how we make sense of the world around us. We believe in gravity, not because there’s a perfect mathematical proof, but because we’ve seen things fall countless times. Newton’s equation F = ma is based on those observations, forming a model of how gravity works. Later, Einstein introduced general relativity, showing that Newton’s model only works in everyday situations, but breaks down under extreme conditions like in space. This was because Newton’s observations weren’t enough to capture the full picture.

This is called inductive reasoning [5], where we use a set of observations to come up with a general principle. The more observations we have, the better our model. That’s why big data is so powerful. If you’ve only seen 10 cats in your life, your idea of what a cat looks like might be biased. But if you’ve seen millions of cats, it’s unlikely a new cat will surprise you. Big data gives us a clearer, more accurate model.

The Limits of Inductive Reasoning

But inductive reasoning has its limits. For example, some of my French friends think that most Moroccans are great at coding. This belief is based on their experience, but it’s biased. Most of the Moroccans they’ve met are students or professionals in tech fields, so the sample is not representative of the whole population. This is called survivor bias [6].

Many philosophers question the reliability of inductive reasoning. If you’re interested, look up the Problem of Induction [7].

Big Data and Patterns

Big data helps us spot patterns in what might seem random. For example, Benford’s Law [8] shows that in many real-world datasets, the first digits of numbers follow a specific distribution: 1: 30% 2: 18% 3: 12% 4: 10% 5: 8% 6: 7% 7: 6% 8: 5% 9: 4%

This pattern is useful in detecting fraud. In 1972, Hal Varian suggested that Benford’s Law could help spot fraud in economic data. People who make up numbers often distribute digits evenly, which doesn’t match the natural pattern.

To see this in action, I picked a dataset from Paris’s open data platform: “Subsidies Paid: Appendix Administrative Account from 2018.” I checked the distribution of the first digits in the “Prestation en nature” column. As I increased the sample size, the numbers followed Benford’s Law more closely.

What Does Big Data Tell Us?

Big data gives us a better understanding of the world. It helps us see the mechanisms behind the patterns. Some might even ask, does everything follow patterns? Could what we think of as free will [9] just be the result of not understanding these patterns?

References

  1. Gradient DescentMIT article on Gradient Descent and Cauchy’s contributions.

  2. Classification and Regression Tree (CART)CART by Leo Breiman and Overview of CART in machine learning.

  3. AdaBoostAdaBoost original paper and AdaBoost tutorial.

  4. Viola-Jones AlgorithmViola-Jones original paper and Simplified explanation.

  5. Inductive ReasoningStanford article on inductive reasoning and Explorable guide.

  6. Survivor BiasHarvard Business Review on survivor bias and Farnam Street guide.

  7. Problem of InductionStanford Encyclopedia of Philosophy on Problem of Induction and Britannica summary.

  8. Benford’s LawBenford’s Law scientific paper and Popular explanation.

  9. Free WillStanford article on free will and Discussion of determinism and free will.