@arilotter/tsne-js
v1.0.3
Published
t-SNE implementation in JavaScript
Downloads
23
Maintainers
Readme
t-SNE.js
t-distributed stochastic neighbor embedding (t-SNE) algorithm implemented in JavaScript
Runs in the browser (also runs in Web Workers)
Runs in node.js
Uses efficient in-place matrix operations via ndarray
Follows closely the API of scikit-learn, allowing specification of perplexity and early exaggeration factor, among other parameters.
Background
t-SNE is a powerful manifold technique for embedding data into low-dimensional space (typically 2-d or 3-d for visualization purposes) while preserving small pairwise distances or local data structures in the original high-dimensional space. In practice, this results in a much more intuitive layout within the low-dimensional space as compared to other techniques. The low-dimensional embedding is learned by minimizing the Kullback-Leibler divergence between the pairwise-similarity probability distribution over the original data space and distribution over the embedding space.
An important note is that the objective function is non-convex with numerous local minima, and thus the results are non-deterministic. There are a few model parameters which influence the learning and optimization process. Selecting appropriate parameters for the input data can significantly improve the chances the model converge on good solutions.
Currently implemented is the exact fomulation, which has computational complexity O(dN^2), where d is the original dimensionality of the data and N is the number of samples. Implementation of the O(dN*logN) Barnes-Hut approximation variant is planned (contributions welcome!).
Usage
Can be run in node.js or the browser. In the browser, should ideally be run in a web worker.
node.js
$ npm install tsne-js --save
import TSNE from 'tsne-js';
let model = new TSNE({
dim: 2,
perplexity: 30.0,
earlyExaggeration: 4.0,
learningRate: 100.0,
nIter: 1000,
metric: 'euclidean'
});
// inputData is a nested array which can be converted into an ndarray
// alternatively, it can be an array of coordinates (second argument should be specified as 'sparse')
model.init({
data: inputData,
type: 'dense'
});
// `error`, `iter`: final error and iteration number
// note: computation-heavy action happens here
let [error, iter] = model.run();
// rerun without re-calculating pairwise distances, etc.
let [error, iter] = model.rerun();
// `output` is unpacked ndarray (regular nested javascript array)
let output = model.getOutput();
// `outputScaled` is `output` scaled to a range of [-1, 1]
let outputScaled = model.getOutputScaled();
browser
<script src="tsne.min.js"></script>
Then it's the same API as above. A browser example using Web Workers is in the example/
folder.
Model Parameters
dim
: number of embedding dimensions, typically 2 or 3perplexity
: approximately related to number of nearest neighbors used during learning, typically between 5 and 50earlyExaggeration
: parameter which influences spacing between clusters, must be at least 1.0learningRate
: learning rate for gradient descent, typically between 100 and 1000nIter
: maximum number of iterations, should be at least 200metric
: distance measure to use for input data, currently implemented measures include'euclidean'
'manhattan'
'jaccard'
(boolean data)'dice'
(boolean data)
You can also pass a distance function to
metric
import cwise from 'cwise'; // Operates on an n-dimensional array using the cwise module let euclidean = cwise({ args: ['array', 'array'], pre: function(a, b) { this.sum = 0.0; }, body: function(a, b) { var d = a - b; this.sum += d * d; }, post: function(a, b) { return Math.sqrt(this.sum); } }); let model = new TSNE({ metric: euclidean });
Build
To run build yourself, for both the browser (outputs to build/tsne.min.js
) and node.js (outputs to dist/
):
$ npm run build
To build for just the browser, run npm run build-browser
, and to build for just node.js, run npm run build-node
.
Tests
$ npm test
References
The original paper on t-SNE:
L.J.P. van der Maaten and G.E. Hinton.
Visualizing High-Dimensional Data Using t-SNE.
Journal of Machine Learning Research 9(Nov):2579-2605, 2008.
Paper on Barnes-Hut variant t-SNE:
L.J.P. van der Maaten.
Accelerating t-SNE using Tree-Based Algorithms.
Journal of Machine Learning Research 15(Oct):3221-3245, 2014.