Viewing a Lagrangian inverse problem as a MAP estimator

\[\newcommand{\bs}{\boldsymbol} \newcommand{\ud}{\mathrm{d}} % Infinitesimal part for integration \newcommand{\Rbb}{\mathbb{R}} \newcommand{\Cbb}{\mathbb{C}} \newcommand{\tinv}[1]{\frac{1}{#1}} \DeclareMathOperator{\argmin}{argmin} \DeclareMathOperator{\argmax}{argmax} \newcommand{\indep}{\perp \!\! \perp} \newcommand{\norm}[2]{\| #1 \|_{#2}}\]

In the following derivations, we show that the typical regularised inverse problem to recover a quantity $\bs x \in \Rbb^N$ from noisy observations of the form $\bs y = \bs{Hx} + \bs n \in \Cbb^M$ with additive noise $\bs n \in \Cbb^M$ corresponds to the maximum a posteriori (MAP) estimator when a prior distribution $p(\bs x) = e^{-\theta \norm{\bs{\Psi x}}{1}}$ is assumed on the data:

\[\begin{align} \hat x_{\text{MAP}} &= \argmax_{\bs x} p(\bs x|\bs y) = \argmax_{\bs x} \frac{p(\bs y|\bs x) p(\bs x)}{p(\bs y)} \underset{p(\bs y) \indep \bs x}{=} \argmax_{\bs x} p(\bs y|\bs x) p(\bs x) \\ &= \argmax_{\bs x} e^{-\tinv 2 \norm{\bs y-\bs{Hx}}{2}^2} e^{-\theta \norm{\bs{\Psi x}}{1}} = \argmax_{\bs x} e^{-\tinv 2 \norm{\bs y-\bs{Hx}}{2}^2 -\theta \norm{\bs{\Psi x}}{1}} \\ &= \argmax_{\bs x} \log \big(e^{-\tinv 2 \norm{\bs y-\bs{Hx}}{2}^2 -\theta \norm{\bs{\Psi x}}{1}} \big) \\ &= \argmin_{\bs x} \underbrace{\tinv 2 \norm{\bs y-\bs{Hx}}{2}^2}_{\text{Data fidelity}} + \underbrace{\theta \norm{\bs{\Psi x}}{1}}_{\text{Regularization}}. \end{align}\]

Additionnally, we define Bayesian point estimators that arise from the decision ``what point $\hat{\bs x} \in \Rbb^N$ summarises $\bs x|\bs y$ best?’’. The optimal decision under uncertainty is

\[\begin{equation*} \hat{\bs x}_L = \argmin_{\bs u \in \Rbb^N} \mathbb{E}[L(\bs u,\bs x)|\bs y] = \argmin_{\bs u \in \Rbb^N} \int L(\bs u,\bs x) p(\bs x|\bs y) \ud \bs x \end{equation*}\]

where the loss $L(\bs u,\bs x)$ measures the “dissimilarity” between $\bs u$ and $\bs x$.

General desiderata:

  • $L(\bs u,\bs x) \ge 0,~ \forall \bs u, \bs x \in \Rbb^N$,
  • $L(\bs u,\bs x)=0 \Leftrightarrow \bs u=\bs x$,
  • $L$ strictly convex w.r.t. $\bs u$ (for estimator uniqueness).

Example:

The minimum mean square error (MMSE) estimator, taking $L(\bs u,\bs x) = \norm{\bs u-\bs x}{2}^2$, is obtained as:

\[\begin{align*} \begin{split} \hat{\bs x}_{\text{MMSE}} &= \argmin_{\bs u} \int \norm{\bs u-\bs x}{2}^2 p(\bs x|\bs y) \ud \bs x \\ &\Leftrightarrow \int (\hat{\bs x}_{\text{MMSE}}-\bs x) p(\bs x|\bs y) \ud \bs x=0 \\ &\Leftrightarrow \hat{\bs x}_{\text{MMSE}} \underbrace{\int p(\bs x|\bs y) \ud \bs x}_{=1} = \int \bs x p(\bs x|\bs y) \ud \bs x \\ &\Leftrightarrow \hat{\bs x}_{\text{MMSE}} = \mathbb{E}[\bs x|\bs y]. \end{split} \end{align*}\]



Enjoy Reading This Article?

Here are some more articles you might like to read next:

  • Google Gemini updates: Flash 1.5, Gemma 2 and Project Astra
  • Displaying External Posts on Your al-folio Blog
  • Quantum bit teleportation
  • Quantum Pauli matrices
  • Adding the LaTeX font on your Windows computer