This theory is designed to minimize the difference between two distributions. His main process is as follows:
First the two distributions to be compared:
where
Then there is the distance function:
Where
Finally it is the optimization problem of transport theory as:
The subject is understood as
$T1_N = u$ which ensures that the sum of masses at the beginning of the transfer is constant and equal to the pre-transfer distribution, and$T^\top1_M = v$ which ensures that the sum of masses after the transfer is equal to the post-transfer distribution. So what he's trying to say is that I want to control that both the entrance and exit of this probability are error-free, which is commonly understood to mean that what departs from one place correctly follows the intended distribution to the other place. Such a design ensures the completeness and consistency of the probability, which in math we call keeping the marginal distribution.
In addition are some variants of OT optimization which include fast optimization using Sinkhorn distance:
Here
Here
$T^*$ This is designed to reduce the amount of transmission while satisfying the edge distribution.
The specific form of
$h(T)$ should be$h(T) = -\sum_{i,j}T_{i,j}logT_{i,j}$ , and this form of entropy measures the dispersion and smoothness of$T$ , with higher entropy values indicating a uniform distribution of$T$ elements, and lower entropy indicating a distribution by a larger number of 0 elements or an uneven distribution.
Since
$u^{(t)}, v^{(t)}$ are alternately updated and both have analytic expressions, their optimization process is indeed fast and converges in only a few iterations.
This code is partly based on the open-source implementations from PLOT.