Proximity Problems
From Wikimization
(→Comparing strain and sstress) |
|||
(6 intermediate revisions not shown.) | |||
Line 6: | Line 6: | ||
= Introduction = | = Introduction = | ||
- | We consider an <math>\,n\times n\,</math> matrix <math>\,D=[d_{ij}]\,</math> defined as a real symmetric matrix that is ''hollow'' <math>\,d_{ii}=0\,</math> for <math>\,i=1\ldots n\,</math> and nonnegative <math>\,d_{ij}\ | + | We consider an <math>\,n\times n\,</math> matrix <math>\,D=[d_{ij}]\,</math> defined as a real symmetric matrix that is ''hollow'' <math>\,d_{ii}=0\,</math> for <math>\,i=1\ldots n\,</math> and nonnegative <math>\,d_{ij}\geq 0\,</math> for all <math>\,i,j\,</math>. |
<math>D\,</math> is said to be Euclidean distance matrix of dimension <math>\,p\,</math> if there exists a list of points <math>\,\{z_1\ldots z_n\}\,</math> in | <math>D\,</math> is said to be Euclidean distance matrix of dimension <math>\,p\,</math> if there exists a list of points <math>\,\{z_1\ldots z_n\}\,</math> in | ||
- | <math>\,\mathbb R^p\,</math> <math>\,(p\ | + | <math>\,\mathbb{R}^p\,</math> <math>\,(p\leq n-1)\,</math> such that |
<math>d_{ij}=\|z_i-z_j\|^2 \quad\forall\,i,j=1\ldots n</math> | <math>d_{ij}=\|z_i-z_j\|^2 \quad\forall\,i,j=1\ldots n</math> | ||
Line 27: | Line 27: | ||
<math>W_s(D)=-\frac{1}{2}(I-es^{\rm T})D(I-se^{\rm T})\qquad(1)</math> | <math>W_s(D)=-\frac{1}{2}(I-es^{\rm T})D(I-se^{\rm T})\qquad(1)</math> | ||
- | is positive semidefinite with <math>\,\text | + | is positive semidefinite with <math>\,{\text rank}(W_s(D))\leq p\,</math>, where <math>\,e\,</math> is a vector of ones and <math>\,s\,</math> is any vector such that |
<math>\,s^{\rm T}e=1\,</math>. | <math>\,s^{\rm T}e=1\,</math>. | ||
Line 66: | Line 66: | ||
= Classical Multidimensional Scaling = | = Classical Multidimensional Scaling = | ||
- | Given <math>\,p\ | + | Given <math>\,p\leq n\,</math>, let <math>\,\mathbb{S}_+^n(p)\,</math> denote the closed set of symmetric <math>\,n\times n\,</math> matrices that are positive semidefinite and have rank no greater than <math>\,p\,</math>. |
- | Let <math>\, | + | Let <math>\,||.||_{\rm F}\,</math> denote the Frobenius norm and <math>\,\Delta\,</math> a given symmetric <math>\,n\times n\,</math> matrix of squared dissimilarities. Let <math>\,W=W(\Delta)\,</math> and <math>\,W_s=W_s\,(\Delta)</math>. |
Classical MDS can be defined by the optimization problem | Classical MDS can be defined by the optimization problem | ||
<math>\begin{array}{rl} | <math>\begin{array}{rl} | ||
- | \text{minimize}_B& | + | \text{minimize}_B&||W-B||_{\rm F}^2\\ |
- | \text{subject to}&B\in\mathbb | + | \text{subject to}&B\in\mathbb{S}_+^n(p) |
\end{array}~~~~~~~~~~~~~~~~~~~~~~~~\textbf{(P)}</math> | \end{array}~~~~~~~~~~~~~~~~~~~~~~~~\textbf{(P)}</math> | ||
Line 80: | Line 80: | ||
<math>\begin{array}{rl} | <math>\begin{array}{rl} | ||
- | \mbox{minimize}_B& | + | \mbox{minimize}_B&||W_s-B||_{\rm F}^2\\ |
- | \text{subject to}&B\in\mathbb | + | \text{subject to}&B\in\mathbb{S}_+^n(p) |
\end{array}~~~~~~~~~~~~~~~~~~~~~~~~(\textbf{P}_\textbf{s)}</math> | \end{array}~~~~~~~~~~~~~~~~~~~~~~~~(\textbf{P}_\textbf{s)}</math> | ||
The following explicit solution to problem '''(P)''' (respectively problem '''(P<sub>s</sub>)''') is well known: | The following explicit solution to problem '''(P)''' (respectively problem '''(P<sub>s</sub>)''') is well known: | ||
- | let <math>\,\lambda_1\ | + | let <math>\,\lambda_1\geq\ldots\geq\lambda_n\,</math> denote the eigenvalues of <math>\,W\,</math> (respectively of <math>\,W_s\,</math>) |
and <math>\,v_1\ldots v_n\,</math> denote the corresponding eigenvectors. | and <math>\,v_1\ldots v_n\,</math> denote the corresponding eigenvectors. | ||
Line 104: | Line 104: | ||
In Section '''5''', we will prove that for any squared dissimilarity matrix <math>\,\Delta\,</math> we have | In Section '''5''', we will prove that for any squared dissimilarity matrix <math>\,\Delta\,</math> we have | ||
- | <math>f\ | + | <math>f\leq f_s</math> |
that is, at the minimum, the strain criterion always gives smaller value than criterion '''(P<sub>s</sub>)'''. | that is, at the minimum, the strain criterion always gives smaller value than criterion '''(P<sub>s</sub>)'''. | ||
Line 110: | Line 110: | ||
;Lemma | ;Lemma | ||
- | Let <math>\,\lambda(C)\in\ | + | Let <math>\,\lambda(C)\in\mathbb{R}^n\,</math> denote the eigenvalues of any symmetric <math>\,n\times n\,</math> matrix in nonincreasing order. |
* For all <math>\,A,B\,</math> | * For all <math>\,A,B\,</math> | ||
- | <math>\lambda_i(A+B) \ | + | <math>\lambda_i(A+B) \leq \lambda_i(A)+\lambda_1(B)</math> |
* For all positive semidefinite <math>\,A,B\,</math> | * For all positive semidefinite <math>\,A,B\,</math> | ||
- | <math>\lambda_i(A\,B)\ | + | <math>\lambda_i(A\,B)\leq \lambda_i(A)\lambda_1(B)\qquad\diamond</math> |
Line 126: | Line 126: | ||
In this section we recall a result (see [2]) that relate the strain and sstress criteria. The sstress criterion is given by: | In this section we recall a result (see [2]) that relate the strain and sstress criteria. The sstress criterion is given by: | ||
- | <math>\begin{array}{rl}\text{minimize}&S(D)= | + | <math>\begin{array}{rl}\text{minimize}&S(D)=||\Delta-D||^2\\ |
\text{subject to}&D\in\mathbb{EDM}^n(p) | \text{subject to}&D\in\mathbb{EDM}^n(p) | ||
\end{array}</math> | \end{array}</math> | ||
'''Result.''' The following inequality holds: | '''Result.''' The following inequality holds: | ||
- | Given <math>\,p\ | + | Given <math>\,p\leq n-1\,</math>, for any <math>\,B\in\mathbb{S}_+^n(p)\,</math>, let <math>\,D={\text diag}(B)e^{\rm T}+e\;{\text diag}(B)^{\rm T}-2B\,</math>. |
Then | Then | ||
- | <math> | + | <math>||\Delta-D||^2 \geq 4||W-B||^2</math> |
- | <br>'''Proof.''' Let <math>\,B\in\mathbb | + | <br>'''Proof.''' Let <math>\,B\in\mathbb{S}_+^n(p)\,</math>; we have |
<math>\begin{array}{rcl} | <math>\begin{array}{rcl} | ||
Line 145: | Line 145: | ||
Writing <math>\,a_{ij}=w_{ij}-b_{ij}\,</math> we get | Writing <math>\,a_{ij}=w_{ij}-b_{ij}\,</math> we get | ||
- | <math>\sum_i\sum_j (\delta_{ij}-d_{ij})^2=2n\,\sum_ia_{ii}^2+4\sum_i\sum_j a_{ij}^2\ | + | <math>\sum_i\sum_j (\delta_{ij}-d_{ij})^2=2n\,\sum_ia_{ii}^2+4\sum_i\sum_j a_{ij}^2\geq 4\sum_i\sum_ja_{ij}^2\qquad\diamond</math> |
= Main result = | = Main result = | ||
Line 151: | Line 151: | ||
;Theorem. | ;Theorem. | ||
- | For any <math>\,s\in\mathbb R^n\,</math> such that <math>\,s^{\rm T}e=1\,</math> and for any <math>\,p\,</math> we have | + | For any <math>\,s\in\mathbb{R}^n\,</math> such that <math>\,s^{\rm T}e=1\,</math> and for any <math>\,p\,</math> we have |
- | <math>\,f \ | + | <math>\,f \leq f_s\qquad(18)</math> |
- | '''Proof.''' We show, for all <math>\,i\,</math>, that <math>\,|\lambda_i(W)|\ | + | '''Proof.''' We show, for all <math>\,i\,</math>, that <math>\,|\lambda_i(W)|\leq|\lambda_i(W_s)|\,</math>. |
Toward that end, we consider two cases: | Toward that end, we consider two cases: | ||
- | * If <math>\,W\,</math> is PSD then <math>\,W_s\,</math> is PSD and the inequality becomes <math>\,\lambda_i(W)\ | + | * If <math>\,W\,</math> is PSD then <math>\,W_s\,</math> is PSD and the inequality becomes <math>\,\lambda_i(W)\leq \lambda_i(W_s)\,</math>. But |
- | <math>\lambda_i(W)=\lambda_i(JW_sJ)\ | + | <math>\lambda_i(W)=\lambda_i(JW_sJ)\leq \lambda_i(W_s)\lambda_1(J)=\lambda_i(W_s)</math> |
because <math>\,\lambda_1(J)=1\,</math>. | because <math>\,\lambda_1(J)=1\,</math>. | ||
Line 168: | Line 168: | ||
<math>\begin{array}{rcl} | <math>\begin{array}{rcl} | ||
\lambda_i(W^2) &=& \lambda_i(W_sJW_s-\frac{1}{n}ee^{\rm T}W_sJW_s) \\ | \lambda_i(W^2) &=& \lambda_i(W_sJW_s-\frac{1}{n}ee^{\rm T}W_sJW_s) \\ | ||
- | &\ | + | &\leq& \lambda_i(W_sJW_s)+\lambda_1(-\frac{1}{n}ee^{\rm T}W_sJW_s) |
\end{array}</math> | \end{array}</math> | ||
Line 177: | Line 177: | ||
because <math>\,J\,</math> and <math>\,W_s^2\,</math> are PSD we have | because <math>\,J\,</math> and <math>\,W_s^2\,</math> are PSD we have | ||
- | <math>\lambda_i(W_sJW_s)\ | + | <math>\lambda_i(W_sJW_s)\leq\lambda_i(W_s^2)\lambda_1(J)=\lambda_i(W_s^2)\qquad\diamond</math> |
= Modified Gower problem = | = Modified Gower problem = | ||
Line 190: | Line 190: | ||
where <math>\,\lambda_i(W_s)\,</math> denotes the <math>\,i^{\rm th}\,</math> eigenvalue of <math>\,W_s\,</math>. | where <math>\,\lambda_i(W_s)\,</math> denotes the <math>\,i^{\rm th}\,</math> eigenvalue of <math>\,W_s\,</math>. | ||
- | But by a well known result... we have, for <math>\,X\!\in\ | + | But by a well known result... we have, for <math>\,X\!\in\mathbb{R}^{n\times p}\,</math> |
<math>\sum_{i=1}^p\lambda_i(W_s)=\max_{X^{\rm T}X=I}\;\mathrm{tr}(X^{\rm T}W_sX)</math> | <math>\sum_{i=1}^p\lambda_i(W_s)=\max_{X^{\rm T}X=I}\;\mathrm{tr}(X^{\rm T}W_sX)</math> |
Current revision
by M. Bennani Dosse
Contents |
Abstract
The aim of this short paper is to give an algebraic result that relate two criteria in multidimensional scaling...
- Key-words
- Euclidean distance, Multidimensional scaling, strain, sstress, comparing criteria.
Introduction
We consider an matrix
defined as a real symmetric matrix that is hollow
for
and nonnegative
for all
.
is said to be Euclidean distance matrix of dimension
if there exists a list of points
in
such that
where denotes Euclidean norm.
Denote by
the set of all
Euclidean distance matrices of dimension
.
A problem common to various sciences is to find the Euclidean distance matrix closest, in some sense, to a given predistance matrix
defined to be any symmetric hollow nonnegative real matrix.
There are three statements of the closest-EDM problem prevalent in the literature, the multiplicity due primarily to choice of projection on the EDM versus positive semidefinite (PSD) cone and vacillation between the distance-square variable
versus absolute distance
.
During the past two decades a large amount of work has been devoted to Euclidean distance matrices and approximation of predistances by an
in a series of works including Gower[6-8], Mathar..., Critchley..., Hayden et al..., etc.
Mathematical preliminaries
It is well known that if and only if the symmetric
matrix
is positive semidefinite with , where
is a vector of ones and
is any vector such that
.
This result was proved by Gower... as a generalization of an earlier result of Schoenberg...
Later Gower considered the particular choices and
where
is the
vector from the standard basis.
In what follows, when
then matrix
will be denoted by
:
We see no compelling reason to prefer one particular over another. Each has its own coherent interpretation. Neither can we say any particular problem formulation produces generally better results than another. Dattorro...
The aim of this short paper is to clarify that point...
We shall also use Wolkowicz' notation:
so the equations (2) and (3) can be written:
It is easy to verify the following properties:
Classical Multidimensional Scaling
Given , let
denote the closed set of symmetric
matrices that are positive semidefinite and have rank no greater than
.
Let denote the Frobenius norm and
a given symmetric
matrix of squared dissimilarities. Let
and
.
Classical MDS can be defined by the optimization problem
Problem (P) can be viewed as a particular case of a more general optimization problem
The following explicit solution to problem (P) (respectively problem (Ps)) is well known:
let denote the eigenvalues of
(respectively of
)
and
denote the corresponding eigenvectors.
Assume that the largest eigenvalues are positive.
Then
is a global minimum of problem (P) (respectively of problem (Ps)). Furthermore, the minimum value for problem (P) is
and for problem (Ps)
In Section 5, we will prove that for any squared dissimilarity matrix we have
that is, at the minimum, the strain criterion always gives smaller value than criterion (Ps). In order to show this result we shall use
- Lemma
Let denote the eigenvalues of any symmetric
matrix in nonincreasing order.
- For all
- For all positive semidefinite
Proof. see, for instance, Wilkinson...
Comparing strain and sstress
In this section we recall a result (see [2]) that relate the strain and sstress criteria. The sstress criterion is given by:
Result. The following inequality holds:
Given , for any
, let
.
Then
Proof. Let ; we have
Writing we get
Main result
In this section we show an inequality involving the criteria in (12) and
in (13).
- Theorem.
For any such that
and for any
we have
Proof. We show, for all , that
.
Toward that end, we consider two cases:
- If
is PSD then
is PSD and the inequality becomes
. But
because .
- If
is not PSD then, using the definition of
:
But
because and
are PSD we have
Modified Gower problem
In this Section we consider the following problem:
Given a nonEuclidean matrix, can we find an that maximizes the total squared real distances from the points to the centroid given by
in the fitted configuration. What is this choice of
?
This problem can be written as an optimization problem in the following manner. First note that if is not Euclidean, then the number of negative eigenvalues of
does not depend on
; call that number
.
The total squared-real distances from the points to the centroid given by in the fitted configuration can be written as
where denotes the
eigenvalue of
.
But by a well known result... we have, for
So the final optimization problem can be written as
where
- Question
Is it true that at the optimum, the problem (20) is equivalent to the problem...
References
[1] Critchley, F., 1986. On certain linear mappings between inner-product and squared-distance matrices. Linear Algebra Appl. 105, 91-107.
[2] De Leeuw, J., Heiser, W., 1982. Theory of multidimensional scaling. Krishnaiah, P.R., Kanal, I.N.(Eds.), Handbook of Statistics, vol. 2. North-Holland, Amsterdam, pp. 285-316 (chapter 13).
[3] Gower, J.C., 1966. Some distance properties of latent root and vector methods in multivariate analysis. Biometrika 53, 315-328.
[4] Gower, J.C., 1982. Euclidean distance geometry, Math. Scientist 7, 1-14.
[5] Schoenberg, I.J., 1935. Remarks to Maurice Fréchet's article Sur la définition axiomatique d'une classe d'espaces distanciés vectoriellement applicable sur l'espace de Hilbert. Ann. of Math. 38, 724-738.
[6] Torgerson, W.S., 1952. Multidimensional scaling: I. Theory and method. Psychometrika 17, 401-419.