A demo code in Matlab for S-WMD [Supervised Word Mover's Distance, NIPS 2016] [Oral presentation video recording by Matt Kusner].
Demo code runs on the bbcsport dataset. Usage: run swmd.m
in MATLAB. Dataset is preprocessed to contain the following fields:
X
is a cell array of all documents, each represented by a dxm matrix where d is the dimensionality of the word embedding and m is the number of unique words in the documentY
is an array of labelsBOW_X
is a cell array of word counts for each documentindices
is a cell array of global unique IDs for words in a documentTR
is a matrix whose ith row is the ith training split of document indicesTE
is a matrix whose ith row is the ith testing split of document indices
Here is a Dropbox link to the datasets used in the paper: https://www.dropbox.com/sh/nf532hddgdt68ix/AABGLUiPRyXv6UL2YAcHmAFqa?dl=0
They're all matlab .mat files and have the following variables (note the similarity to the demo dataset):
for bbcsport, twitter, recipe, classic, amazon
X [1,n ne]
: each cell corresponds to a document and is a[d,u]
matrix whered
is the dimensionality of the word embedding,u
is the number of unique words in that document,n
is the number of training points, andne
is the number of test points. Each column is the word2vec vector for a particular word.Y [1,n ne]
: the label of each documentBOW_X [1,n ne]
: each cell in the cell array is a vector corresponding to a document. The size of the vector is the number of unique words in the document, and each entry is how often each unique word occurs.words [1,n ne]
: each cell corresponds to a document and is itself a{1,u}
cell where each entry is the actual word corresponding to each unique wordTR [5,n]
: each row corresponds to a random split of the training set, each entry is the index with respect to the full dataset. So for example, to get the BOW of the training set for the third split do:BOW_xtr = BOW_X(TR(3,:))
TE [5,ne]
: same as TR except for the test set
for ohsumed, reuters (r8), 20news (20ng2_500)
The only difference with the above datasets is that because there are pre-defined train-test splits, there are already variables BOW_xtr
, BOW_xte
, xtr
, xte
, ytr
, yte
.
In the paper, we used cross-validation to set k for each dataset and tried these k's [1,3,5,7,9,11,13,15,17,19]. We also implemented a KNN function that given a k (or a list of k's) would only classify a point if the majority of the k nearest neighbors voted on the same class. If not, then we would reduce k (by 2) and consider if for this smaller k there was a majority vote for a class. This would continue this way until either a majority was reached or k=1 (in which case we just use the nearest neighbors vote). This function is in the file knn_fall_back.m