Errors-in-variables model

Errors-in-variables (EIV) is a robust modeling technique in statistics, which assumes that every variable can have error or noise. Errors-in-variables is also referred to as total least squares (TLS), in a broad sense, in the literature of computational mathematics and engineering. However, TLS in a strict sense implies the application of EIV or orthogonal regression to a linear model $$\mathbf{A x} = \mathbf{b}$$.

Robust linear regression
In linear regression, the least squares (LS) attributes all error to the dependent variables. It has variant versions according to other error configurations, including total least squares (i.e. orthogonal error), data least squares (DLS), and constrained or structured TLS.

Given an observation vector $$\mathbf{b} \in \reals^n$$ and a data matrix $$\mathbf{A} \in \reals^{n \times m}$$, consider the solution of the overdetermined system of equations


 * $$\mathbf{Ax \approx b}$$.

The ordinary least squares method (OLS) yields the solution $$\mathbf{x}$$ that minimizes the Euclidean norm of the residuals $$\|{\mathbf{Ax-b}}\|_2$$, where $$\|\cdot\|_2$$ is also known as the two-norm. The residual is an estimate of the error. Equivalently, the OLS problem can be paraphrased by


 * $$ \min_{\mathbf{x}}\|\Delta\mathbf{b}\|_2 \quad

\mbox{ subject to }\quad \mathbf{Ax}=\mathbf{b}+\Delta\mathbf{b}. $$

If the data matrix $$\mathbf{A}$$ is also noisy (i.e. error in both the dependent and the explanatory variables), the OLS solution is no longer optimal. In cases where orthogonal optimization is acceptable, TLS offers a proper formulation:


 * $$ \min_{\mathbf{x}} \|{[{\Delta\mathbf{A}\,\Delta\mathbf{b}}]}\|_F \quad

\mbox{ subject to }\quad (\mathbf{A}+\Delta\mathbf{A})\mathbf{x}=\mathbf{b}+\Delta\mathbf{b},$$

where $$\|{\cdot}\|_F$$ is the Frobenius norm (or colloquially the "length" of the vector); and the perturbations $$\Delta\mathbf{A}$$ and $$\Delta\mathbf{b}$$ are used to compensate for the noisy signals $$\mathbf{A}$$ and $$\mathbf{b}$$, respectively. This formulation of TLS also implies that the noises are assumed to be independent and identically distributed (i.i.d.) both in $$\mathbf{A}$$ and $$\mathbf{b}$$. Note that the objective can have a weighting matrix according to the distribution of errors if the distribution is known or well-estimated, which is called the constrained or structured TLS.

In the other case, where the noise is only in $$\mathbf{A}$$, DLS can be used alternatively as


 * $$ \min_{\mathbf{x}} \|{[{\Delta\mathbf{A}}]}\|_F \quad \mbox{ subject to } \quad (\mathbf{A}+\Delta\mathbf{A})\mathbf{x}=\mathbf{b}.$$

The solution of the OLS problem can be obtained by using the (pseudo-)inverse of the data matrix. Solutions to the TLS and DLS problems have been shown to be closely connected to a set of singular vectors of the (augmented) system-related matrix corresponding to the minimum singular value.