-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathintroduction.tex
78 lines (56 loc) · 6.52 KB
/
introduction.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
\chapter{Introduction}
Traditional signal processing usually follows Shannon and Nyquist's sampling theorem. It states that a signal can be recovered perfectly if the sampling frequency is at least twice the highest frequency present in the signal. However, this theorem assumes that the samples are taken with a uniform time interval, and that the reconstruction happens by interpolating the samples with sinc functions.
Compressive sensing is a new scheme for sampling and reconstructing which abandons these assumptions. It turns out that by placing different assumptions on the signal, and by using a different recovery strategy, we can recover our signal with less samples than the traditional theory suggests.
The field of compressive sensing has exploded in the last decade, after the initial publications by Cand{\`e}s, Tao, Romberg and Donoho \cite{candes2006near, candes2006robust, donoho2006compressed}. Some of the ideas and concepts have roots further back in time, but these publications mark the beginning of compressive sensing as a field of study. The key idea is to assume that the signal is \textit{sparse}, and then look for the sparsest possible signal which matches the sampled signal. A lot of new theory has been developed for recovering such signals.
During the development of this theory, one has found several applications of compressive sensing techniques. In this report, we will work more closely with one of these applications, namely face recognition. Face recognition is perhaps one of the most studied problems in machine learning and statistical classification. Possibly because the traditional approaches still show several weaknesses, as we will discuss more in \cref{sec:applications}, but also because of the human mind's extraordinary ability to recognize faces, even when key elements such as skin tone, hair color or length, facial hair or glasses, change.
In this report we will first give a general introduction to the basic concepts of compressive sensing, and then look at how these techniques can be applied to the face recognition problem.
\section{Preliminaries and Notation}
In this section we will introduce some of the necessary notation and concepts for this report.
Throughout we will denote vectors by boldface lower case letters, and matrices by boldface upper case letters. For a vector $ \x $, we will by $ x_{i} $ refer to the $ i $'th element in the vector. Similarly, for a matrix $ \A $, we will by $ \a_{i} $ refer to the $ i $'th column of $ \A $, and by $ a_{i, j} $ denote the element of $ \A $ at column $ j $ and row $ i $.
Sets will be denoted by italic upper case letters, and the cardinality of a set $ S $ is denoted as $ \abs{S} $. The complement of a set $ S $ will be written as $ \overline{S} $.
\subsection{Norms}
As we soon will see, a central part of compressive sensing is a minimization problem involving different norms of a vector. Therefore, we will begin with the definition of norms. Even though the author suspect this to be known material, it is included for ease of reference later.
\begin{definition} \label{def:norm}
Let $ V $ be a vector space over a field $ \K $. A norm $ \norm{\cdot} $ on $ V $ is a function $ \norm{\cdot}\colon V \to\R $ such that
\begin{subdef}
\item $ \norm{\v} \geq 0 $ for all $ \v \in V $ with $ \norm{\v} = \0 $ if and only if $ \v = \0 $
\item $ \norm{c\v} = \abs{c}\norm{\v} $ for all $ \v \in V $ and $ c \in \K $
\item $ \norm{\u + \v} \leq \norm{\u} + \norm{\v} $ for all $ \u, \v \in V $
\end{subdef}
\end{definition}
This definition is quite broad, and very general. In this report we will be working more closely with a family of norms called the $ \ell_{p} $ norms:
\begin{definition} \label{def:lpnorm}
Let $ p \geq 1 $ and $ p \in \R $. The $ \ell_{p} $ norm $ \norm{\cdot}_{p} $ is defined as
\[
\norm{\v}_{p} = \left( \sum_{i=1}^{n} \abs{v_{i}}^{p} \right)^{\dfrac{1}{p}}
\]
\end{definition}
We will not prove here that the $ \ell_{p} $-norms actually are norms, but this can be proven for all $ p\geq 1 $. We observe that for $ p=1 $ we get the Manhattan norm, and for $ p=2 $ we get euclidean norm. If we let $ p\to\infty $ we arrive at the supremum norm. Even though \cref{def:lpnorm} does not allow for $ p < 1 $, it can be proven that if we let $ p\to0 $, accept that $ 0^{0} = 0 $, and ignore the $ 1/p $-exponent, we get what is sometimes called the $ \ell_{0} $ norm:
\begin{definition} \label{def:l0norm}
The $ \ell_{0} $ norm $ \norm{\cdot}_{0} $ is the number of non-zero entries in $ \v $.
\end{definition}
It is worth noting that the $ \ell_{0} $ norm is strictly speaking not a norm, since $ \norm{\cdot}_{0} $ does not fulfill axiom \textit{(ii)} of \cref{def:norm}. In fact, for all $ q<1 $, $ \norm{\cdot}_{q} $ is not a norm, as $ \norm{\cdot}_{q} $ does not fulfill axiom \textit{(iii)} for any $ q \in (0, 1) $. Despite this, it is customary to refer to $ \norm{\cdot}_{0} $ as the $ \ell_{0} $ norm.
\subsection{Support and Sparsity}
As compressive sensing deals with the recovery of sparse vectors, we will need to define sparsity. Before we do that, we will introduce \textit{support}:
\begin{definition} \label{def:support}
Let $ \v \in \Cx^{N} $. The support $ S $ of $ \v $ is defined as the index set of its non-zero entries, that is:
\[
\supp \v = \set{j \in \indexset{N} \st v_{j} \neq 0}
\]
\end{definition}
The notion of support yields a new formulation of \cref{def:l0norm}: The $ \ell_{0} $ norm is simply the cardinality of the support: $ \norm{\v}_{0} = \abs{\supp \v} $.
For a vector $ \v\in\Cx^{N} $ and a set $ S \subset \indexset{N} $, we denote by $ \v_{S} $ either the subvector in $ \Cx^{\abs{S}} $ consisting of the entries in $ \v $ indexed by $ S $, that is:
\begin{equation}
\label{eq:vSalt1}
(\v_{S})_{i} = v_{i} \for i \in S
\end{equation}
Or the vector in $ \Cx^{N} $ which coincides with $ \v $ on the indices in $ S $, and is zero otherwise, that is:
\begin{equation}
\label{eq:vSalt2}
(\v_{S})_{i} = \twopartdef{v_{i}}{\text{if } i \in S}{0}{\text{Otherwise}}
\end{equation}
It should always be clear from context which of these is used. Similarly, for a matrix $ \A\in\R^{m \times n} $ we will by $ \A_{S} \in \R^{m \times \abs{S}} $ refer to the matrix consisting of the columns of $ \A $ indexed by $ S $.
The final concept we will introduce is the notion of \textit{sparsity}:
\begin{definition} A vector $ \v \in \Cx^{N} $ is said to be $ s $-sparse if it has no more than $ s $ non-zero entries. That is, $ \norm{\v}_{0} \leq s $
\end{definition}
Note that any vector supported on a set $ S $ with $ \abs{S} = s $ is $ s $-sparse.