**Multidimensional scaling (MDS)** is a way to reduce the dimensionality of data to visualize it. We basically want to project our (likely highly dimensional) data into a lower dimensional space and preserve the distances between points. The first time that I learned about MDS was by way of STATS 202 with John Taylor, although to be honest, I got through the class by watching old lectures by David Mease on youtube. Thank goodness for youtube!

If we have some highly complex data that we project into some lower N dimensions, we will assign each point from our data a coordinate in this lower dimensional space, and the idea is that these N dimensional coordinates are ordered based on their ability to capture variance in the data. Since we can only visualize things in 2D, this is why it is common to assess your MDS based on plotting the first and second dimension of the output. It’s not that the remaining N-2 axis don’t capture meaningful information, however they capture progressively less information.

If you look at the output of an MDS algorithm, which will be points in 2D or 3D space, the distances represent similarity. So very close points = very similar, and points farther away from one another = less similar. MDS can also be useful for assessing correlation matrices, since a correlation is just another metric of similarity between two things.

### How does MDS Work?

Let’s implement very simple MDS so that we (and when I say we, I mean I!) know what is going on. The input to the MDS algorithm is our proximity matrix. There are two kinds of classical MDS that we could use:

- Classical (metric) MDS is for data that has metric properties, like actual distances from a map or calculated from a vector
- Nonmetric MDS is for more ordinal data (such as human-provided similarity ratings) for which we can say a 1 is more similar than a 2, but there is no defined (metric) distance between the values of 1 and 2.

For this post, we will walk through the algorithm for classical MDS, with metric data of course! I’m going to use one of Matlab’s random datasets that is called “cities” and is made up of different ratings for a large set of US cities.

```
% Multidimensional scaling (MDS) Example
% Load matlab cities data
load cities
% This data has cities in rows, and different categories for ratings in
% columns. We will implement MDS to assess city similarity based on
% ratings.
% Step 1: Set up our proximity matrix
% First let's create our similarity (proximity) matrix by calculating the
% euclidean distance between pairwise cities
proximities = zeros(size(ratings,1));
for i=1:size(ratings,1)
for j =1:size(ratings,1)
proximities(i,j) = pdist2(ratings(i,:),ratings(j,:),'euclidean');
end
end
% Step 2: Create our centering matrix
% Now we need to perform double centering. I'm going to define these
% matrices explicitly so it's clear
% This is an identity matrix of size n, where n is the size of our nxn
% proximity matrix that we just made
n = size(proximities,1);
identity = eye(n);
% This is an equally sized matrix of 1's
one = ones(n);
% Here is our centering matrix, commonly referred to as "J"
centering_matrix = identity - (1/n) * one;
J = centering_matrix;
% Step 3: Apply double centering, meaning we square our proximity matrix,
% and multiply it by J on each side with a -.5 coefficient
B = -.5*J*(proximities).*(proximities)*J;
% Step 4: Extract some M eigen values and vectors, where M is the dimension
% we are projecting down to:
M = 2; % so we can plot in 2D
[eigvec,eigval] = eig(B);
% We want to get the top M...
[eigval, order] = sort(max(eigval)','descend');
eigvec = eigvec(order,:);
eigvec = eigvec(:,1:M); % Note that eigenvectors are in columns
eigval = eigval(1:M);
% Plop them back into the diagonal of a matrix
A = zeros(2);
A(1:3:end) = eigval;
% If we multiply eigenvectors by eigenvalues, we get our new representation
% of the data, X:
X = eigvec*A;
% We can now look at our cities in 2D:
plot(X(:,1),X(:,2),'o')
title('Example of Classical MDS with M=2 for City Ratings');
% Grab just the state names
% NOTE: doesn't work perfectly for all, but it's good enough!
% ALSO NOTE: The labels were hard to see, so they are not
% included in the graphic below.
cities = cell(size(names,1),1);
for c=1:size(names,1)
cities{c} = deblank(names(c,regexp(names(c,:),', ')+2:end));
end
% Select a random subset of 20 labels
y = randsample(size(cities,1),20);
% Throw on some labels!
text(X(y,1), X(y,2), cities(y), 'VerticalAlignment','bottom', ...
'HorizontalAlignment','right')
```

Here are our cities, with MDS applied, closer dots = more similar cities, as assessed by ratings pertaining to climate, housing, health, crime, transportation, education, arts. recreation, and economics.

This isn’t the greatest example, because showing the labels wouldn’t give you much insight into if the method works. (Better would have been to use a dataset of actual distances then projected into 2d, oh well!) The end of the script does format and display labels, however the picture doesn’t show them, because they didn’t add much. For more details (and a better example that uses simple cities distances) see here.

Remember that this is an incredibly simple form of MDS. See page 589 of Elements of Statistical Learning for more.