Automated optimized parameters for T-distributed stochastic neighbor embedding improve visualization and analysis of large datasets

Anna C. Belkina, Chris Ciccolella, R. Anno, Richard L. Halpert, Josef Spidlen, J. Snyder-Cappione
2019
1 reference

Abstract

Accurate and comprehensive extraction of information from high-dimensional single cell datasets necessitates faithful visualizations to assess biological populations. A state-of-the-art algorithm for non-linear dimension reduction, t-SNE, requires multiple heuristics and fails to produce clear representations of datasets when millions of cells are projected. We develop opt-SNE, an automated toolkit for t-SNE parameter selection that utilizes Kullback-Leibler divergence evaluation in real time to tailor the early exaggeration and overall number of gradient descent iterations in a dataset-specific manner. The precise calibration of early exaggeration together with opt-SNE adjustment of gradient descent learning rate dramatically improves computation time and enables high-quality visualization of large cytometry and transcriptomics datasets, overcoming limitations of analysis tools with hard-coded parameters that often produce poorly resolved or misleading maps of fluorescent and mass cytometry data. In summary, opt-SNE enables superior data resolution in t-SNE space and thereby more accurate data interpretation. Visualisation tools that use dimensionality reduction, such as t-SNE, provide poor visualisation on large data sets of millions of observations. Here the authors present opt-SNE, that automatically finds data set-tailored parameters for t-SNE to optimise visualisation and improve analysis.

1 repository
1 reference

Code References

â–¶ scikit-learn/scikit-learn
1 file
â–¶ doc/modules/manifold.rst
1
Belkina et al. (2019) is to set the learning rate to the sample size
Link copied to clipboard!