We then conducted preliminary numerical experiments based on our theoretical results in the context of a simple hypothesis testing example. Our theory now allows users to specify preferred characteristics using an infinite-dimensional hypothesis class.
The experimental outcome shows that our method, based on our theoretical results, can efficiently distinguish between two distributions with user-preferred characteristics while providing good accuracy. These experimental results provide further insights and help to clarify why, despite the curse of dimensionality, the Wasserstein distance metric still performs well. And that, independent of the dimension of the hypothesis class, and across a wide range of machine learning and statistical applications.
From a theoretical perspective, we’ve established a general and fairly complete set of mathematical results to solve a fundamental gap between theory and practice in machine learning and statistical applications. At the same time, we are interested in studying statistical convergence for general function classes and in developing efficient algorithms to compute .
Our next big step is also to expand the preliminary numerical experiments in the paper – which were intended to be an initial demonstration of our theoretical results — to a broad set of applications that leverage our theoretical insights.
It may have started with military barracks — but there is an absolutely infinite universe of possibilities to explore with the wonderful notion of optimal transport.
Mathematical Sciences: We’re currently focused on optimization, probability, complexity, geometry of data, as well as linear and multi-linear algebra, to deliver tools that are fundamental to big data and AI.
Si, N., Blanchet, J., Ghosh, S. & Squillante, M. Quantifying the Empirical Wasserstein Distance to a Set of Measures: Beating the Curse of Dimensionality. Advances in Neural Information Processing Systems 33, 21260–21270 (2020) ↩