- Published at
Local Features for Medical Applications
Lifting of raw data to higher dimensional feature space while maintaining localily.
- Authors
- Name
- Dr. Dmitrij Sitenko
- Mathematical Image Processing at Ruprecht Karls University Heidelberg
Table of Contents
The Manifold of Symmetric Positive Definite Matrices
For a given positive integer we next carry over the differential geometric objects to the specific open set of symmetric positive definite matrices that play an important role in mathematics, physics, numerical analysis and imaging that is given by the linear subspace of
The tangent space at consists of symmetric matrices
As the set is embedded in the Euclidean space the
most
simple metric on is naturally given in terms of the
Frobenius norm
We will use as a building backbone for prototype
extraction in connection with the labeling task
More
specifically, we adopt various
geometries on by utilizing
alternative metrics and clarify their
influence on the resulting mean properties defined by optimization
problem along with main
computational tools for mean approximation.
Affine Invariant Riemannian Metric
Geometrically the set of symmetric positive define matrices
comprises the interior of a convex cone in the Euclidean space
with zero curvature. Therefore, one major limitation of metric
is the finite distance to points on the
boundary, i.e. matrices with zero
eigenvalues. Instead, equipping the tangent space with the
Riemannian metric.
commonly known as the affine invariant
Riemannian metric alleviates the problem and turns the set into a
negatively curved
Riemannian manifold with a natural Riemannian distance
where are the generalized eigenvalues of the pencil that is
For this particular case the exponential map reads
In the map is denoted as the matrix exponential with the inverse given by matrix logarithm
Finally, given a smooth objective function , the Riemannian gradient is given by
where the symmetric matrix denotes the Euclidean gradient of at . Since is a simply connected, complete and nonpositively curved Riemannian manifold, the exponential map is globally defined and bijective, and the Riemannian mean always exists and is uniquely defined as minimizer of the objective function , after substituting the Riemannian distance .
Given a set of positive definite matrices
together with positive weights , we next focus on the solution of the
problem for specific geometry ,
with the distance given by . Due to
invertibility of
, each tangent vector is
represented
at the base point by
As a result, optimality condition reads
Apart from the specific case in
where the Riemannian mean is directly given by , the solution to nonlinear equation
is not explicitly known. A remedy is the corresponding basic fixed iteration initialized at
where the step size can be defined heuristically or by following more
sophisticated
line search strategies, see and .
However, as pointed out in applying
is
limited in the sense that convergence is not
theoretically guaranteed and if the
iteration converges, than at a linear rate only.
This can be remedied by restricting the set of matrices in
to
pairwise commute resulting in the
following variant of Riemannian mean recovery proposed by
that comes with guarantees to converge
at a quadratic rate. Using the
parametrization by means corresponding to the Cholesky decomposition
along with replacing the map of fixed point iteration with its linearization leads to the following fixed point iteration
with damping parameter . Comparing to shows that the basic idea is to compute the Riemannian mean as fixed point of the iteration
Algorithm provides a refined variant of this iteration including adaptive stepsize selection.
Fixed Point Iteration for Computing the Riemannian Matrix Mean
Initialization
-
(termination threshold)
-
, , with solving.
-
,
-
(condition number and step size selection parameters)
-
-
(iterative step)
-
While :
- If , stop
Log-Euclidean Metric
Leveraging the specific properties of the mappings and an alternative metric was proposed by (among several other ones). More precisely, introducing the operations
prescribes on a Lie group structure where set
is isomorphic to the vector space with
playing the role of addition. The Log-Euclidean metric for two
symmetric positive definite matrices is then defined by
the Euclidean inner product on the flat Riemannian space of zero curvature
after applying the mapping ,
(cf. ) and provides a lower bound to the (AIRM) metric via
The corresponding Riemannian mean of the set has the closed form expression
which has analogous form as the arithmetic mean. While computing the mean is considerably cheaper than integrating the flow using approximation Algorithm, the critical drawback of relying on is not taking into account the structure (curvature) of the manifold . Therefore, in the next section, we additionally consider another approximation of the Riemannian mean that better fits the underlying geometry and avoids the expensive computation of generalized eigenvalues resulting in a more efficient mean evaluation than the Riemannian mean associated to metric.
-Divergence
In this section we make use of divergence functions introduced in Section and present a general approach of approximating the objective function by instead replacing the intractable problem specific squared Riemannian distance by a divergence function
Property says that, for any feasible , the Hessian with respect to the first argument is positive definite. In fact, suitable divergence functions recover in this way locally the metric tensor of the underlying manifold , in order to qualify as a surrogate for the squared Riemannian distance .
Properties of mean | Riemann. | Log-Eucl. | Stein div. | |
---|---|---|---|---|
Invariance by reordering of | ✔ | ✔ | ✔ | |
Congruence invariance | ||||
for | ✔ | ✔, | ✔ | |
Self-duality | ✔ | ✔ | ✔ | |
Joint homogeneity | ||||
✔ | ✔ | ✖ | ||
Determinant identity | ||||
✔ | ✔ | ✔ | ||
If commute pairwise then: | ||||
✔ | ✔ | ✔ |
For the present case of interest, proposed the divergence function, called Stein divergence. For it is defined as
which is not a metric on due to the lack of triangle inequality. However, as shown by taking the square root in yield a metric on . Thus replacing the Riemannian distance in the second term of problem the Riemannian distance by provides an approximate objective
which allows mean recovery while avoiding to solve the numerically involved generalized eigenvalue problem as opposed to . The resulting Riemannian gradient flow reads
with
Discretizing the flow using the geometric explicit Euler scheme with step size ,
and using the log-Euclidean mean as initial point , defines Algorithm as listed below.
(termination threshold)
solves
(any value )
While :
for
The key advantage of Algorithm over Algorithm is that it respects the geometry of
while remaining numerically efficient.
In Section~ we will provide an experimental
evaluation and a qualitative
comparison of the mean properties with respect to the aforementioned metrics
while extracting prototypes for the labeling task.
Application
details are reported in Chapter .
We conclude by listing the main properties of the resulting means
when relying on the Riemannian, log-Euclidean and Stein divergence summarized
in Table.