maximum a posteriori estimate is a parameter learning scheme that uses Beta Distribution and Baysian inference to get a distribution of the posterior of the parameter, and return the argmax (i.e. the mode) of the MAP. This differs from MLE because we are considering a distribution of possible parameters:
Calculating a MAP posterior, in general:
We assume that the data points are IID, and the fact that the bottom of this is constant, we have:
Usually, we’d like to argmax the log:
where, g is the probability density of \theta happening given the prior belief, and f is the likelyhood of x_{i} given parameter \theta. You will note this is just Maximum Likelihood Parameter Learning function, plus the log-probability of the parameter prior. MAP for Bernoulli and Binomial p To estimate p, we use the Beta Distribution: The MODE of the beta, which is the MAP of such a result:
now, for a Laplace posterior Beta(2,2), we have:
MAP for Poisson and Exponential \lambda We use the gamma distribution as our prior
where \alpha-1 is the prior event count, and \beta is the prior time periods. Let’s say you have some data points x_1, …x_{k}, the posterior from from those resulting events: