
Multidimensional scaling (MDS) is a means of visualizing the level of
similarity of individual cases of a dataset. MDS is used to translate "information about the pairwise 'distances' among a set of
objects or individuals" into a configuration of
points mapped into an abstract
Cartesian space.
More technically, MDS refers to a set of related
ordination techniques used in
information visualization, in particular to display the information contained in a
distance matrix. It is a form of
non-linear dimensionality reduction
Nonlinear dimensionality reduction, also known as manifold learning, refers to various related techniques that aim to project high-dimensional data onto lower-dimensional latent manifolds, with the goal of either visualizing the data in the low-d ...
.
Given a distance matrix with the distances between each pair of objects in a set, and a chosen number of dimensions, ''N'', an MDS
algorithm places each object into ''N''-
dimensional space (a lower-dimensional representation) such that the between-object distances are preserved as well as possible. For ''N'' = 1, 2, and 3, the resulting points can be visualized on a
scatter plot
A scatter plot (also called a scatterplot, scatter graph, scatter chart, scattergram, or scatter diagram) is a type of plot or mathematical diagram using Cartesian coordinates to display values for typically two variables for a set of data. ...
.
Core theoretical contributions to MDS were made by
James O. Ramsay
James O. Ramsay (born 5 September 1942) is a Canadian statistician and Professor Emeritus at McGill University, Montreal, who developed much of the statistical theory behind multidimensional scaling (MDS). Together with co-author Bernard Silverm ...
of
McGill University, who is also regarded as the founder of
functional data analysis.
Types
MDS algorithms fall into a
taxonomy, depending on the meaning of the input matrix:
Classical multidimensional scaling
It is also known as Principal Coordinates Analysis (PCoA), Torgerson Scaling or Torgerson–Gower scaling. It takes an input matrix giving dissimilarities between pairs of items and outputs a coordinate matrix whose configuration minimizes a
loss function
In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cost ...
called ''strain.'',
which is given by
where
denote vectors in ''N''-dimensional space,
denotes the scalar product between
and
, and
are the elements of the matrix
defined on step 2 of the following algorithm, which are computed from the distances.
: Steps of a Classical MDS algorithm:
: Classical MDS uses the fact that the coordinate matrix
can be derived by
eigenvalue decomposition from
. And the matrix
can be computed from proximity matrix
by using double centering.
:# Set up the squared proximity matrix