Optimal transport
Optimal transport
The Optimal transport (OT) problem has attracted a surge of research interest because it expresses the distance between probability distributions, known as the Wasserstein distance. This strong property enables us to apply this distance to diverse machine learning problems such as generative adversarial networks, graph optimal transport, clustering, and domain adaptation.
The Kantorovich relaxation formulation of the optimal transport (OT) problem is explained briefly. Let and be {probabilities} or positive weight vectors as and , respectively. Given two empirical distributions, i.e., discrete measures, , and the ground cost matrix between their supports, the problem can be formulated as
,
where represents the transport matrix, and where the domain is defined as
,
where and are the marginal constraints. Moreover, we present the sum of the two vectors respectively as and , i.e., and . Note that is equal to in the standard OT formulation. The obtained OT matrix brings powerful distances as , which is known as the -th order Wasserstein distance. It is used in various fields according to the value of . Especially, the distance is applied to computer vision when , and clustering when . When is the ground cost matrix and we specifically designate the -th order Wasserstein distance as the OT distance.