Neighbor Embeddings Using Unbalanced Optimal Transport Metrics
By: Muhammad Rana, Keaton Hamm
Potential Business Impact:
Makes computers learn better from messy data.
This paper proposes the use of the Hellinger--Kantorovich metric from unbalanced optimal transport (UOT) in a dimensionality reduction and learning (supervised and unsupervised) pipeline. The performance of UOT is compared to that of regular OT and Euclidean-based dimensionality reduction methods on several benchmark datasets including MedMNIST. The experimental results demonstrate that, on average, UOT shows improvement over both Euclidean and OT-based methods as verified by statistical hypothesis tests. In particular, on the MedMNIST datasets, UOT outperforms OT in classification 81\% of the time. For clustering MedMNIST, UOT outperforms OT 83\% of the time and outperforms both other metrics 58\% of the time.
Similar Papers
Structured Matching via Cost-Regularized Unbalanced Optimal Transport
Machine Learning (Stat)
Matches different data types, even if they don't fit.
Sharp Convergence Rates of Empirical Unbalanced Optimal Transport for Spatio-Temporal Point Processes
Statistics Theory
Measures how well data points match patterns.
Unsupervised Learning for Optimal Transport plan prediction between unbalanced graphs
Machine Learning (CS)
Makes computers compare big networks much faster.