| %!TEX root = ceres-solver.tex |
| \chapter{Non-linear Least Squares} |
| \label{chapter:tutorial:nonlinsq} |
| Let $x \in \reals^n$ be an $n$-dimensional vector of variables, and |
| $F(x) = \left[f_1(x); \hdots ; f_k(x)\right]$ be a vector of residuals $f_i(x)$. |
| The function $f_i(x)$ can be a scalar or a vector valued |
| function. Then, |
| \begin{equation} |
| \arg \min_x \frac{1}{2} \sum_{i=1}^k \|f_i(x)\|^2. |
| \end{equation} |
| is a Non-linear least squares problem~\footnote{Ceres can solve a more general version of this problem, but for pedagogical reasons, we will restrict ourselves to this class of problems for now. See section~\ref{chapter:overview} for a full description of the problems that Ceres can solve}. Here $\|\cdot\|$ denotes the Euclidean norm of a vector. |
| |
| Such optimization problems arise in almost every area of science and engineering. Whenever there is data to be analyzed, curves to be fitted, there is usually a linear or a non-linear least squares problem lurking in there somewhere. |
| |
| Perhaps the simplest example of such a problem is the problem of Ordinary Linear Regression, where given observations $(x_1,y_1),\hdots, (x_k,y_k)$, we wish to find the line $y = mx + c$, that best explains $y$ as a function of $x$. One way to solve this problem is to find the solution to the following optimization problem |
| \begin{equation} |
| \arg\min_{m,c} \sum_{i=1}^k (y_i - m x_i - c)^2. |
| \end{equation} |
| With a little bit of calculus, this problem can be solved easily by hand. But what if, instead of a line we were interested in a more complicated relationship between $x$ and $y$, say for example $y = e^{mx + c}$. Then the optimization problem becomes |
| \begin{equation} |
| \arg\min_{m,c} \sum_{i=1}^k \left(y_i - e^{m x_i + c}\right)^2. |
| \end{equation} |
| This is a non-linear regression problem and solving it by hand is much more tedious. Ceres is designed to help you model and solve problems like this easily and efficiently. |