MBI Videos

Optimal Transport Workshop

  • video photo
    Sayan Mukherjee
    It has been a longstanding challenge in geometric morphometrics and medical imaging to infer the physical locations (or regions) of 3D shapes that are most associated with a given response variable (e.g. class labels) without needing common predefined landmarks across the shapes, computing correspondence maps between the shapes, or requiring the shapes to be diffeomorphic to each other. In this talk, we introduce SINATRA: the first machine learning pipeline for sub-image analysis which identifies physical shape features that explain most of the variation between two classes without the aforementioned requirements. We also illustrate how the problem of 3D sub-image analysis can be mapped onto the well-studied problem of variable selection in nonlinear regression models. Here, the key insight is that tools from integral geometry and differential topology, specifically the Euler characteristic, can be used to transform a 3D mesh representation of an image or shape into a collection of vectors with minimal loss of geometric information.

    Crucially, this transform is invertible. The two central statistical, computational, and mathematical innovations of our method are: (1) how to perform robust variable selection in the transformed space of vectors, and (2) how to pullback the most informative features in the transformed space to physical locations or regions on the original shapes. We highlight the utility, power, and properties of our method through detailed simulation studies, which themselves are a novel contribution to 3D image analysis. Finally, we apply SINATRA to a dataset of mandibular molars from four different genera of primates and demonstrate the ability to identify unique morphological properties that summarize phylogeny.
  • video photo
    Samir Chowdhury
    Geometric and topological data analysis methods are increasingly being used in human neuroimaging studies to derive insights into neurobiology and behavior. We will begin by describing a novel application of optimal transport toward predicting task performance, and go on to explain why reproducing such insights across clinical populations requires statistical learning techniques such as averaging and PCA across graphs without known node correspondences. We formulate this problem using the Gromov-Wasserstein (GW) distance and present a recently-developed Riemannian framework for GW-averaging and tangent PCA. This framework permits derived network representations beyond graph geodesic distances or adjacency matrices. As an application, we show that replacing the adjacency matrix formulation in state-of-the-art implementations with a spectral representation leads to improved accuracy and runtime in graph learning tasks. Joint work with Caleb Geniesse, Facundo Mémoli, Tom Needham, and Manish Saggar.
  • video photo
    Claire Brecheteau
    I will introduce proxies for the distance function to the support of a distribution, whose sublevel sets are unions of balls or ellipsoids. I will consider rates of approximation of these proxies by their empirical counterpart, built from sample points. I will also explain how to use such estimators to cluster geometrically structured datasets.
  • video photo
    Marcel Klatt
    In recent years, the theory of optimal transport (OT) has found its way into data analysis. Especially regularized OT methods have encountered growing interest, as the routine use of OT in applications is still hampered by its computational complexity. Among others, the most prominent proposal is entropy regularization that serves to define an entropy regularized OT plan and a corresponding divergence also known as Sinkhorn divergence. This talk invites to a small trip through distributional limit theory for certain empirical (regularized) OT quantities defined for distributions supported on finite metric spaces. In particular, we will explore the statistical differences between asymptotic distributions for empirical non-regularized OT quantities and their regularized counterparts. Specific focus is set to the empirical regularized OT plan for which we can prove that it asymptotically follows a Gaussian law. As a consequence we discuss applications in colocalization analysis of protein interaction networks based on regularized OT. In the final part of the talk, we consider the non-regularized OT plan that is a solution to a finite dimensional basic linear program. In fact, distributional limit theory for such a quantity is not as straightforward and brings into play the combinatorial nature and the concept of degeneracy inherent in linear programming. This is joint work with Carla Tameling, Axel Munk and Yoav Zemel.
  • video photo
    Radmila Sazdanovic
    A multitude of knot invariants, including quantum invariants and their categorifications, have been introduced to aid with characterizing and classifying knots and their topological properties. Relations between knot invariants and their relative strengths at distinguishing knots are still mostly elusive. We use Principal Component Analysis (PCA), Ball Mapper, and machine learning to examine the structure of data consisting of various polynomial knot invariants and the relations between them. Although of different origins, these methods confirm and illuminate similar substructures in knot data. These approaches also enable comparison between numerical invariants of knots such as the signature and s-invariant via their distribution within the Alexander and Jones polynomial data. This is joint work with P. Dlotko, J. Levitt, and M. Hajij.
  • video photo
    Justin Curry, Justin Curry
    This talk will begin with a review of elementary constructions in topological data analysis (TDA), such as merge trees and the Elder Rule. A brief introduction to inverse problems in TDA is considered before turning to recent joint work with Catanzaro, Fasy, Lazovskis, Malen, Reiss, Wang and Zabka in https://arxiv.org/abs/1909.10623. The talk will conclude with recent combinatorial results obtained by my PhD student, Jordan DeSha, who provided a closed-form formula for the number of unbraided height equivalence classes (HECs) of embedded two-spheres with a prescribed level-set barcode. This formula establishes a conjecture outlined in the earlier paper with Catanzaro, et al.
  • video photo
    Katy Craig
    Over the past ten years, optimal transport has become a fundamental tool in statistics and machine learning: the Wasserstein metric provides a new notion of distance for classifying distributions and a rich geometry for interpolating between them. In parallel, optimal transport has led to new theoretical results on the stability and long time behavior of partial differential equations through the theory of Wasserstein gradient flows. These two lines of research recently intersected in a series of works that characterized the dynamics of training neural networks with a single hidden layer as a Wasserstein gradient flow. In this talk, I will briefly introduce the mathematical theory of Wasserstein gradient flows and describe recent results on discrete to continuum limits. In particular, I will show how passing from the discrete to continuum limit by introducing an appropriate regularization can lead to faster rates of convergence, as well as novel, deterministic particle methods for diffusive processes.
  • video photo
    Julie Delon
    In this talk we will introduce a Wasserstein-type distance on the set of Gaussian mixture models. This distance is de�ned by restricting the set of possible coupling measures in the optimal transport problem to Gaussian mixture models. We derive a very simple discrete formulation for this distance, which makes it suitable for high dimensional problems. We also study the corresponding multimarginal and barycenter formulations. We show some properties of this Wasserstein-type distance, and we illustrate its practical use with some examples in image processing.
  • video photo
    Varun Jog
    Modern machine learning algorithms are surprisingly fragile to adversarial perturbations of data. In this talk, we present some theoretical contributions towards understanding fundamental bounds on the performance of machine learning algorithms in the presence of adversaries. We shall discuss how optimal transport emerges as a natural mathematical tool to characterize "robust risk", which is a notion of risk in the adversarial machine learning literature that is analogous to Bayes risk in hypothesis testing. We shall also show how, in addition to tools from optimal transport, we may use reverse-isoperimetric inequalities from geometry to provide theoretical bounds on the sample size of estimating robust risk.
  • video photo
    Xianfeng Gu
    This work introduces an optimal transportation (OT) view of generative adversarial networks (GANs). Natural datasets have intrinsic patterns, which can be summarized as the manifold distribution principle: the distribution of a class of data is close to a low-dimensional manifold. GANs mainly accomplish two tasks: manifold learning and probability distribution transformation. The latter can be carried out using the classical OT method. From the OT perspective, the generator computes the OT map, while the discriminator computes the Wasserstein distance between the generated data distribution and the real data distribution; both can be reduced to a convex geometric optimization process. Furthermore, OT theory discovers the intrinsic collaborative—instead of competitive—relation between the generator and the discriminator, and the fundamental reason for mode collapse. We also propose a novel generative model, which uses an autoencoder (AE) for manifold learning and OT map for probability distribution transformation. This AE–OT model improves the theoretical rigor and transparency, as well as the computational stability and efficiency; in particular, it eliminates the mode collapse. The experimental results validate our hypothesis, and demonstrate the advantages of our proposed model.
  • video photo
    Bei Wang
    In this talk, we will give two examples of how topology could be used as a knob for machine learning. In the first example, topology is used as an interior knob for dimensionality reduction. In particular, via a method called the H-Isomap, topological information is used in combination with landmark Isomap to obtain homology-preserving embeddings of high-dimensional point clouds. In the second example, topology is used as an exterior knob for probing deep learning models. Specifically, we probe a trained deep neural network by studying neuron activations (combinations of neuron firings) with a large set of input images. Using topological summaries, we study the organizational principle behind neuron activations, and how these activations are related within a layer and across layers. Using an interactive tool called TopoAct, we present visual exploration scenarios that provide valuable insights towards learned representations of an image classifier.
  • video photo
    Jessi Cisewski
    Data exhibiting complicated spatial structures are common in many areas of science (e.g., cosmology, biology), but can be difficult to analyze. Persistent homology is an approach within the area of Topological Data Analysis (TDA) that offers a framework to represent, visualize, and interpret complex data by extracting topological features which may be used to infer properties of the underlying structures. For example, TDA is a beneficial technique for analyzing intricate and spatially complex web-like data such as fibrin or the large-scale structure (LSS) of the Universe. The output from persistent homology, called persistence diagrams, summarizes the different order holes in the data (e.g., connected components, loops, voids). I will present a framework for inference or prediction using functional transformations of persistence diagrams and discuss how persistent homology can be used to locate cosmological voids and filament loops in the LSS of the Universe.
  • video photo
    Pavan Turaga
    In this talk, we present an overview of our work in the area of activity analysis from a diverse set of modalities such as video, motion capture, wearables, and more. We show how topological descriptors derived from the time-series of human activity can lead to robust and invariant representations for a variety of problems. We also discuss our recent work in new topological featurization approaches using perturbation methods, and fusion with deep-learning. Our applications include recognition of human activities as well as uncovering underlying qualities of human activity for gait and balance-related health interventions.
  • video photo
    Hongteng Xu
    Graph representation is significant for many real-world applications, e.g., network analysis, molecule clustering, and classification, etc. In this talk, I will introduce a new nonlinear factorization model, representing graphs that are with topological structures, and optionally, node attributes. This model is based on a pseudo-metric called Gromov-Wasserstein (GW) discrepancy, which compares graphs in a relational way. It achieves a novel and flexible factorization mechanism under the GW discrepancy, which estimates observed graphs as GW barycenters constructed by a set of atoms with different weights. The atoms and the weights associated with the observed graphs are learned by minimizing the GW discrepancy between the observed graphs and the barycenters of the atoms. I will show that this GW factorization model represents graphs with different sizes as vectorized permutation-invariant features. The learning algorithms of this model, its extensions, and potential applications will be discussed in-depth.
  • video photo
    Sinho Chewi
    We study first order methods to compute the barycenter of a probability distribution P over the space of probability measures with finite second moment. We develop a framework to derive global rates of convergence for both gradient descent and stochastic gradient descent despite the fact that the barycenter functional is not geodesically convex. Our analysis overcomes this technical hurdle by employing a Polyak-Lojasiewicz (PL) inequality and relies on tools from optimal transport and metric geometry. In turn, we establish a PL inequality when P is supported on the Bures-Wasserstein manifold of Gaussian probability measures. It leads to the first global rates of convergence for first order methods in this context.
  • video photo
    Christoph Weitkamp
    In this talk, we present a statistical theory for object matching based on the Gromov-Wasserstein distance. To this end, we model general objects as metric measure spaces. Based on this, we propose a simple and efficiently computable symptotic statistical test for pose invariant object discrimination. This is based on an empirical version of a ß- trimmed lower bound of the Gromov-Wasserstein distance. We derive for ß € [0; 1=2) distributional limits of this test statistic. To this end, we introduce a novel U-type process indexed in ß and show its weak convergence. Finally, the theory developed is investigated in Monte Carlo simulations and applied to structural protein comparisons.
  • video photo
    Chao Chen
    Existing generative models (GAN or VAE) focus on generating realistic images based on CNN-derived image features, but fail to preserve the structural properties of real images. This can be fatal in applications where the underlying structure (e.g., neurons, vessels, membranes, and road networks) of the image carries crucial semantic meaning. We propose a novel GAN model that learns the topology of real images, i.e., connectedness and loopy-ness. In particular, we introduce a topological GAN loss that bridges the gap between synthetic image distribution and real image distribution in the topological feature space. By optimizing this loss, the generator produces images with the same structural topology as real images. We also propose new GAN evaluation metrics that measure the topological realism of the synthetic images. We show in experiments that our method generates synthetic images with realistic topology. We also highlight the increased performance that our method brings to downstream tasks such as segmentation.
  • video photo
    Brittany Terese Fasy
    The persistence diagram is a topological summary that is gaining traction as a (directional) descriptor of shapes in Euclidean space. Recent work has shown that well-chosen (finite) sets of diagrams can differentiate between geometric simplicial complexes, providing a method for representing shapes using a finite set of topological descriptors. A related inverse problem is the following: given an oracle we can query for persistence diagrams, what is underlying geometric simplicial complex? This talk will explore the representation of simplicial complexes by parameterized families of persistence diagrams, along with the inverse problem of how to recover the initial simplicial complex.
  • video photo
    Guido Montufar
    We study the problem of minimizing the Wasserstein distance between a probability distribution and an algebraic variety. We consider the setting of finite state spaces and describe the solution depending on the choice of the ground metric and the given distribution. The Wasserstein distance between the distribution and the variety is the minimum of a linear functional over a union of transportation polytopes. We obtain a description in terms of the solutions of a finite number of systems of polynomial equations. A detailed analysis is given for the two bit independence model.
  • video photo
    Justin Solomon
    Sampling provides a common means of access to probability measures in Bayesian inference, streaming data processing, and other applications. In this setting, it is often impossible to access the distribution function or to make assumptions about the supports of the measures involved. In this talk, I will summarize some efforts in our research group to estimate optimal transport distances and solve derived optimization problems (e.g., barycenter estimation and barycentric regression) given only sample access to the input measures.
  • video photo
    Robert McCann
    Flocking and swarming models which seek to explain pattern formation in mathematical biology often assume that organisms interact through a force which is attractive over large distances yet repulsive at short distances. Suppose this force is given as a difference of power laws and normalized so that its unique minimum occurs at unit separation. For a range of exponents corresponding to mild repulsion and strong attraction, we show that the minimum energy configuration is uniquely attained | apart from translations and rotations | by equidistributing the organisms over the vertices of a regular top-dimensional simplex (i.e. an equilateral triangle in two dimensions and regular tetrahedron in three).

    If the attraction is not assumed to be strong, we show these configurations are at least local energy minimizers in the relevant d1 metric from optimal transportation, as are all of the other uncountably many unbalanced configurations with the same support. These therefore form stable attractors for the associated rst- and second-order dynamics. We infer the existence of phase transitions.

    An ingredient from the proof with independent interest is the establishment of a simple isodiametric variance bound which characterizes regular simplices: it shows that among probability measures on Rn whose supports have at most unit diameter, the variance around the mean is maximized precisely by those measures which assign mass 1=(n + 1) to each vertex of a (unit-diameter) regular simplex. Based on preprint with Tongseok Lim at https://arxiv.org/abs/1907.13593
  • video photo
    Theo Lacombe
    Persistence diagrams (PD) are routinely used in topological data analysis as descriptors to encode the topological properties of some object. These diagrams can be compared with a partial matching metric, sometimes called the "Wasserstein distance between persistence diagrams" due to its important similarities with the metrics used in optimal transport literature, although an explicit connection between these two formalisms was yet to come. By considering the space of persistence diagrams as a measure space, we reformulate its metrics as optimal partial transport problems and introduce a generalization of persistence diagrams, namely Radon measures supported on the upper half plane. Such measures naturally appear in topological data analysis when considering continuous representations of persistence diagrams (e.g. persistence surfaces) but also as expectations of probability distributions on the space of persistence diagrams. We will showcase the strength of this optimal-transport-based formalism on two problems arising in topological data analysis. First, we provide a characterization of convergence in the space of persistence diagrams (with respect to their standard metrics) in terms of vague convergence of measures. This result provides a powerful tool to study continuity properties in this space; in particular it gives an exhaustive characterization of continuous linear representations of persistence diagrams, a common tool used when incorporating persistence diagrams in machine learning pipelines. Second, this formalism allows us to prove new results regarding persistence diagrams in a random setting, as it enables to manipulate some limit objects such as expected persistence diagrams (that are not standard persistence diagrams) and to prove convergence rates and stability results in this context.
  • video photo
    Henry Adams
    Given a sample of points X from a metric space M, the Vietoris-Rips simplicial complex VR(X;r) at scale r>0 is a standard construction to attempt to recover M from X, up to homotopy type. A deficiency is that VR(X;r) is not metrizable if it is not locally finite, and thus does not recover metric information about M. We remedy this shortcoming by defining the Vietoris-Rips metric thickening VR^m(X;r) via the theory of optimal transport. Vertices are reinterpreted as Dirac delta masses, points in simplices are reinterpreted as convex combinations of Dirac delta masses, and distances are given by the Wasserstein distance between probability measures. When M is a Riemannian manifold, the Vietoris-Rips thickening satisfies Hausmann's theorem (VR^m(M;r) is homotopy equivalent to M for r sufficiently small) with a simpler proof: homotopy equivalence VR^m(M;r) -> M is now canonically defined as a center of mass map, and its homotopy inverse is the (now continuous) inclusion M -> VR^m(M;r). We discuss Vietoris-Rips thickenings of circles and n-spheres, and relate these constructions to Borsuk-Ulam theorems into higher-dimensional codomains. Joint work with Michal Adamaszek, John Bush, and Florian Frick.
  • video photo
    Lori Ziegelmeier
    A time-varying collection of metric spaces as formed, for example, by a moving school of fish or flock of birds, can contain a vast amount of information. There is sometimes a need to simplify or summarize the dynamic behavior, and recently, topological tools have been applied to this purpose. One such method is a crocker plot, a 2-dimensional image that displays the (non-persistent but varying with scale) topological information at all times simultaneously. We use this method to perform exploratory data analysis and investigate parameter recovery in the collective motion model of D’Orsogna et al. (2006). Then, we use it to choose between unbiased correlated random walk models of Nilsen et al. (2013) that describe motion tracking experiments on pea aphids. Finally, we discuss an extension of the crocker plot that is persistent and equivalent to the information in a vineyard and hence, inherits the nice stability properties of vineyards. For some purposes, the information in a vineyard is more accessible when instead displayed in this new manner.

View Videos By