The maximum likelihood estimation (MLE) and maximum a posterior (MAP) are two ways of estimating a parameter given observed data. Here I’ll compare them, by applying both methods to a really simple problem in 1-dimension (based on the univariate Gaussian distribution). The same basic methodology forms the concepts used to derive more powerful Bayesian methods like Bayesian linear regression and even Gaussian processes.

Imagine we buy a small box of \(N\) pencils. The box says the pencils are 6cm long, but as there seems to be some variation you want a better estimate and so you measure each pencil. This gives us some data, i.e.

We’ll call our best estimate \(\theta\), which is commonly used to denote the parameter in probability theory. One way to find an estimate of \(\theta\) is to recognise that we have a distribution of measured pencil lengths in \(D\), and to find a value for \(\theta\) that maximises the likelihood of this distribution. This is known as the maximum likelihood estimate,

By assuming that our measurements in \(D\) are i.i.d (independent and identically distributed) then we can rewrite the above as,

We can now take each \( p(x_i|\theta) \) to be normally distributed with mean and variance , i.e. , and so,

To find the MLE, we need to find the value of \(\theta\) that maximises the above expression. First we’ll take the log of everything to simplify the maths (this is valid as taking the log is a monotonic function), and then we’ll differentiate to find the stationary points.

Solving for \(\theta\) then gives us the result which is just the arithmetic mean of the measured samples.

Maximum a posteriori estimate (MAP)

The MLE assumes no prior knowledge of the expected length of a pencil, however we said above that the box of pencils states they are around 6cm long, so surely we should use this? Using a prior like this is a Bayesian approach to the problem, and involves using the data to find a distribution for \(\theta\).

The MAP estimate is written as,

Using Bayes’ theorem (and recognising that p(D) is independent of \theta) we get something that looks pretty close to the MLE case,

The \(p(\theta)\) term represents our prior belief, i.e. what we believe before we start making measurements. Here we will assume it is normally distributed with mean and variance : . The approach will be very similar: we take the log, and then differentiate to find the stationary points.

Find stationary points,

After some algebra, we get the result,

So the MAP estimate is just a linear combination of the sample and prior means (actually it is a convex combination). It’s also useful to see that as \(N\to\infty\), the MAP estimate for \(\theta\) reduces to the MLE estimate. Referring back to our pencils in a box example, this makes intuitive sense: we were told the pencils should be 6cm long (our prior belief), but after we’ve observed enough data we’ll start having a stronger belief in our measured sample mean.