Back to Basics
ELI5 maths refreshers. Short, visual explainers for the building blocks that keep showing up in machine learning — logarithms, exponents, summation notation, and more.
8 parts
Logarithms — What Are They?
Part 1 of the Back to Basics series. Logarithms from first principles — what they are, why they exist, and why they show up everywhere in machine learning.
Exponents, Powers, and the Number e
Part 2 of the Back to Basics series. What exponents actually are, why anything to the power of zero is 1, and how the mysterious number e powers everything from …
Summation and Product Notation (Σ and Π)
Part 3 of the Back to Basics series. Breaking down sigma and pi notation from first principles — every piece explained, concrete examples, and the Python …
Functions and Graphs
Part 4 of the Back to Basics series. What a function actually is, how to read a graph, and the key functions that power machine learning — all visualised with …
Fractions, Ratios, and Percentages
Part 5 of the Back to Basics series. Fractions, ratios, and percentages — the language every ML metric is built on, from accuracy to F1 score.
Inequalities and Absolute Values
Part 6 of the Back to Basics series. Greater than, less than, absolute value as distance, and why these simple ideas power thresholds, error metrics, and …
Coordinate Systems and Plotting
Part 7 of the Back to Basics series. From the number line to n-dimensional feature space — how coordinate systems work and why every dataset is a cloud of …
Sets and Set Notation
Part 8 of the Back to Basics series. Sets and set notation from first principles — the foundation of probability theory, data splitting, and half the notation …