Hostname: page-component-cd9895bd7-7cvxr Total loading time: 0 Render date: 2024-12-22T23:21:55.362Z Has data issue: false hasContentIssue false

Discrete-to-continuum limits of optimal transport with linear growth on periodic graphs

Published online by Cambridge University Press:  20 December 2024

Lorenzo Portinale*
Affiliation:
Institut für angewandte Mathematik, Universität Bonn, Bonn, Germany
Filippo Quattrocchi
Affiliation:
Institute of Science and Technology Austria (ISTA), Klosterneuburg, Austria
*
Corresponding author: Lorenzo Portinale; Email: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

We prove discrete-to-continuum convergence for dynamical optimal transport on $\mathbb{Z}^d$-periodic graphs with cost functional having linear growth at infinity. This result provides an answer to a problem left open by Gladbach, Kopfer, Maas, and Portinale (Calc Var Partial Differential Equations 62(5), 2023), where the convergence behaviour of discrete boundary-value dynamical transport problems is proved under the stronger assumption of superlinear growth. Our result extends the known literature to some important classes of examples, such as scaling limits of $1$-Wasserstein transport problems. Similarly to what happens in the quadratic case, the geometry of the graph plays a crucial role in the structure of the limit cost function, as we discuss in the final part of this work, which includes some visual representations.

Type
Papers
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press

1. Introduction

In the Euclidean setting, the Benamou–Brenier [Reference Benamou and Brenier5] formulation of the distance on the space $\mathscr{P}_2(\mathbb{R}^d)$ known as 2-Wasserstein or Kantorovich–Rubinstein distance is given by the minimisation problem

(1.1) \begin{equation} \mathbb{W}_2 (\mu _0, \mu _1)^2 = \inf \left \{ \int _0^1 \int _{\mathbb{R}^d} \frac{|\nu _t|^2}{\mu _t}{\, \mathrm{d}} x{\, \mathrm{d}} t \, : \, \partial _t \mu _t + \nabla \cdot \nu _t = 0, \quad \mu _{t=0}=\mu _0,\quad \mu _{t=1}=\mu _1 \right \}, \end{equation}

for every $\mu _0,\mu _1 \in \mathscr{P}_2(\mathbb{R}^d)$ . The Partial Differential Equation (PDE) constraint is called continuity equation (we write $(\boldsymbol{\mu },\boldsymbol{\nu }) \in \mathsf{CE}$ when $(\boldsymbol{\mu },\boldsymbol{\nu })$ is a solution). Over the years, the Benamou–Brenier formula (1.1) has revealed significant connections between the theory of optimal transport and different fields of mathematics, including partial differential equations [Reference Jordan, Kinderlehrer and Otto29], functional inequalities [Reference Otto and Villani35], and the novel notion of Lott–Sturm–Villani’s synthetic Ricci curvature bounds for metric measure spaces [Reference Lott and Villani30, Reference Lott and Villani31, Reference Sturm36, Reference Sturm37]. Inspired by the dynamical formulation (1.1), in independent works, Maas [Reference Maas32] (in the setting of Markov chains) and Mielke [Reference Mielke33] (in the context of reaction-diffusion systems) introduced a notion of optimal transport in discrete settings structured as a dynamical formulation à la Benamou–Brenier as in (1.1). One of the features of this discretisation procedure is the replacement of the continuity equation with a discrete counterpart: when working on a (finite) graph $(\mathcal{X},\mathcal{E})$ (resp. vertices and edges), the discrete continuity equation reads

\begin{align*} \partial _t m_t(x) + \sum _{y\sim x}J_t(x,y) = 0, \quad \forall x \in \mathcal{X}, \quad \big (\text{we write } ({\boldsymbol{m}}, \boldsymbol{J}) \in \mathcal{CE}_{\mathcal{X}} \big ) \end{align*}

where $(m_t,J_t)$ corresponds to discrete masses and fluxes (s.t. $J_t(x,y) = -J_t(y,x)$ ). Maas’ proposed distance $\mathcal{W}$ [Reference Maas32] is obtained by minimising, under the above constraint, a discrete analogue of the Benamou–Brenier action functional with reference measure $\pi \in \mathscr{P}(\mathcal{X})$ and weight function $\omega \in \mathbb{R}^{\mathcal{E}}_+$ , of the form

\begin{align*} \int _0^1 \frac{1}{2} \sum _{(x,y) \in \mathcal{E}} \hspace{-1mm}\frac{|J_t(x,y)|^2}{\widehat{r}_t(x,y)} \omega (x,y){\, \mathrm{d}} t, \quad \text{where} \quad \widehat{r}_t(x,y) \,:\!=\, \theta _{\log }(r_t(x), r_t(y)), \quad r_t(x)\,:\!=\, \frac{m_t(x)}{\pi (x)}, \end{align*}

and where $\theta _{\log }(a,b) \,:\!=\, \int _0^1 a^s b^{1-s}{\, \mathrm{d}} s$ denotes the $1$ -homogeneous, positive mean called logarithmic mean. With this particular choice of the mean, it was proved [Reference Maas32], [Reference Mielke33] (see also [Reference Chow, Huang, Li and Zhou6]) that the discrete heat flow coincides with the gradient flow of the relative entropy with respect to the discrete distance $\mathcal{W}$ . In discrete settings, the equivalence between static and dynamical optimal transport breaks down, and the latter stands out in applications to evolution equations, discrete Ricci curvature and functional inequalities [Reference Erbar, Henderson, Menz and Tetali9, Reference Erbar and Maas15, Reference Mielke34, Reference Erbar and Maas16, Reference Erbar, Maas and Tetali11, Reference Fathi and Maas18, Reference Erbar and Fathi14]. Subsequently, several contributions have been devoted to the study of the scaling behaviour of discrete transport problems, in the setting of discrete-to-continuum approximation problems. The first convergence results were obtained in [Reference Gigli and Maas21] for symmetric grids on a $d$ -dimensional torus, and by [Reference Trillos20] in a stochastic setting. In both cases, the authors obtained convergence of the discrete distances towards $\mathbb{W}_2$ in the limit of the discretisation getting finer and finer.

Nonetheless, it turned out that the geometry of the graph plays a crucial role in the game. A general result was obtained in [Reference Gladbach, Kopfer and Maas24], where it is proved that the convergence of discrete distances associated with finite-volume partitions with vanishing size to the $2$ -Wasserstein space is substantially equivalent to an asymptotic isotropy condition on the mesh. The first complete characterisation of limits of transport costs on periodic graphs (Figure 1) in arbitrary dimension for general action functionals (not necessarily quadratic) was established in [Reference Gladbach, Kopfer, Maas and Portinale22, Reference Gladbach, Kopfer, Maas and Portinale23]: in this setting, the limit action functional (more precisely, the energy density) can be explicitly characterised in terms of a cell formula, which is a finite-dimensional constrained minimisation problem depending on the initial graph and the cost function at the discrete level. The action functionals considered in [Reference Gladbach, Kopfer, Maas and Portinale23] are of the form

(1.2) \begin{align} (\boldsymbol{\mu }, \boldsymbol{\nu }) \in \mathsf{CE} \quad \mapsto \quad{\mathbb{A}}(\boldsymbol{\mu },\boldsymbol{\nu })\,:\!=\, \int _{(0,1) \times{\mathbb{T}^d}} f \big ( \rho, j \big ){\, \mathrm{d}} \mathscr{L}^{\,d+1} + \int _{ (0,1) \times{\mathbb{T}^d}} f^\infty \big ( \rho ^\perp, j^\perp \big ){\, \mathrm{d}} \boldsymbol{\sigma }, \end{align}

where we used the Lebesgue decomposition

\begin{align*} \boldsymbol{\mu } = \rho \mathscr{L}^{\,d+1} + \boldsymbol{\mu }^\perp, \quad \boldsymbol{\nu } = j \mathscr{L}^{\,d+1} + \boldsymbol{\nu }^\perp, \quad \text{ and }\quad \boldsymbol{\mu }^\perp = \rho ^\perp \boldsymbol{\sigma }, \quad \boldsymbol{\nu }^\perp = j^\perp \boldsymbol{\sigma }, \quad \big ( \boldsymbol{\sigma } \perp \mathscr{L}^{\,d+1} \big ) \end{align*}

and where the energy density $f: \mathbb{R}_+ \times \mathbb{R}^d \to \mathbb{R} \cup \{+\infty \}$ is some given convex, lower semicontinuous function with at least linear growth, i.e. satisfying

(1.3) \begin{align} f(\rho, j) \geq c |j| - C (\rho + 1), \qquad \forall \rho \in \mathbb{R}_+ \quad \text{ and }\quad j \in \mathbb{R}^d, \end{align}

whereas $f^\infty$ denotes its recession function (see (2.2) for the precise definition). The choice $f(\rho, j) \,:\!=\, |j|^2/\rho$ corresponds to the $\mathbb{W}_2$ distance. At the discrete level, on a locally finite connected graph $(\mathcal{X}, \mathcal{E})$ embedded in $\mathbb{R}^d$ , the natural counterpart is represented by action functionals of the form

(1.4) \begin{align} ({\boldsymbol{m}}, \boldsymbol{J}) \in \mathcal{CE}_{\mathcal{X}} \quad \mapsto \quad \mathcal{A}({\boldsymbol{m}},\boldsymbol{J})\,:\!=\, \int _0^1 F(m,J){\, \mathrm{d}} t, \end{align}

for a given lower semicontinuous, convex, and local cost function $F$ , which also has at least linear growth with respect to the second variable (see (2.8) for the precise definition).

The main result in [Reference Gladbach, Kopfer, Maas and Portinale23] is the $\Gamma$ -convergence for constrained functionals as in (1.4), after a suitable rescaling of the graph $\mathcal{X}_{\varepsilon } \,:\!=\, \varepsilon \mathcal{X}$ , $\mathcal{E}_{\varepsilon } \,:\!=\, \varepsilon \mathcal{E}$ , and of the cost $F_{\varepsilon }$ (and associated action $\mathcal{A}_{\varepsilon }$ ), in the framework of $\mathbb{Z}^d$ -periodic graphs. In particular, the limit action is of the form (1.2), where the energy density $f=f_{\mathrm{hom}}$ is given in terms of a cell formula, explicitly reading

\begin{align*} f_{\mathrm{hom}}(\rho, j) \,:\!=\, \inf \left \{ F(m,J) \, : \, (m,J) \in{\mathsf{Rep}}(\rho, j) \right \}, \qquad \rho \in \mathbb{R}_+, \ \ j \in \mathbb{R}^d, \end{align*}

where ${\mathsf{Rep}}(\rho, j)$ denotes the set of discrete representatives of $\rho$ and $j$ , given by all $\mathbb{Z}^d$ -periodic functions $m: \mathcal{X} \to \mathbb{R}_+$ with

\begin{align*} \sum _{x \in \mathcal{X} \cap [0,1)^d} m(x) = \rho \end{align*}

and all $\mathbb{Z}^d$ -periodic anti-symmetric discrete vector fields $J: \mathcal{E} \to \mathbb{R}$ with zero discrete divergence and with effective flux equal to $j$ , i.e.,

(1.5) \begin{align} \mathsf{div}\, J(x) \,:\!=\, \sum _{y \sim x} J(x,y) = 0 \quad \forall x \in \mathcal{X} \quad \text{ and }\quad \mathsf{Eff} (J) \,:\!=\, \frac 12 \sum _{\substack{(x,y) \in \mathcal{E} \\ x \in [0,1)^d} } J(x,y) (y-x) = j . \quad \end{align}

The result covers several examples, both for what concerns the geometric properties of the graph (such as isotropic meshes of $\mathbb T^d$ , or the simple nearest-neighbour interaction on the symmetric grid) as well as the choice of the cost functionals (including discretisation of $p$ -Wasserstein distances in arbitrary dimension and flow-based models, i.e. when $F$ – or $f$ – does not depend on the first variable).

Figure 1. Example of $\mathbb{Z}^d$ -periodic graph embedded in $\mathbb{R}^d$ .

As a consequence of this $\Gamma$ -convergence (in time-space) and a compactness result for curves of measures with bounded action [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.9], one obtains as a corollary [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.10] that, under the stronger assumption of superlinear growth on $F$ , also the corresponding discrete boundary-value problems (i.e. the associated squared distances, in the case of the quadratic Wasserstein problems) $\Gamma$ -converge to the corresponding continuous one, namely ${\mathcal{MA}}_{\varepsilon } \stackrel{\Gamma }{\to }{\mathbb{MA}}_{\mathrm{hom}}$ (with respect to the weak topology), where

\begin{align*}{\mathcal{MA}}_{\varepsilon }(m_0,m_1) &\,:\!=\, \inf \left \{ \mathcal{A}_{\varepsilon }({\boldsymbol{m}}, \boldsymbol{J}) \, : \, ({\boldsymbol{m}},\boldsymbol{J}) \in \mathcal{CE}_{\mathcal{X}_{\varepsilon }} \quad \text{ and }\quad{\boldsymbol{m}}_{t=0}=m_0, \, \,{\boldsymbol{m}}_{t=1}=m_1 \right \}, \\{\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1) &\,:\!=\, \inf \left \{{\mathbb{A}}_{\mathrm{hom}}(\boldsymbol{\mu },\boldsymbol{\nu }) \, : \, (\boldsymbol{\mu }, \boldsymbol{\nu }) \in \mathsf{CE} \quad \text{ and }\quad \boldsymbol{\mu }_{t=0}=\mu _0, \, \, \boldsymbol{\mu }_{t=1}=\mu _1 \right \} \, \end{align*}

are the minimal discrete and homogenised action functionals, respectively. The superlinear growth condition, at the continuous level, is a reinforcement of the condition (1.3) and assumes the existence of a function $\theta :[0,\infty ) \to [0,\infty )$ with $\lim _{t \to \infty } \frac{\theta (t)}{t} = \infty$ and a constant $C \in \mathbb{R}$ such that

\begin{align*} f(\rho, j) \geq (\rho +1) \theta \left ( \frac{\lvert j\rvert }{\rho +1} \right ) - C (\rho +1), \qquad \forall \rho \in \mathbb{R}_+, \ \ j \in \mathbb{R}^d . \end{align*}

In particular, this forces every $(\boldsymbol{\mu },\boldsymbol{\nu }) \in \mathsf{CE}$ with finite action to satisfy $\boldsymbol{\nu } \ll \boldsymbol{\mu } + \mathscr{L}^{\,d+1}$ [[Reference Gladbach, Kopfer, Maas and Portinale23], Remark 6.1], and it ensures compactness in $\mathcal{C}_{\mathrm{KR}} \bigl ([0,1];\, \mathcal{M}_+({\mathbb{T}}^d) \bigr )$ [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.9], i.e. with respect to the time-uniform convergence in the Kantorovich–Rubinstein norm (recall that the KR norm metrises weak convergence on $\mathcal{M}_+({\mathbb{T}}^d)$ , see [[Reference Gladbach, Kopfer, Maas and Portinale23], Appendix A]). This compactness property makes the proof of the convergence ${\mathcal{MA}}_{\varepsilon } \stackrel{\Gamma }{\to }{\mathbb{MA}}_{\mathrm{hom}}$ an easy corollary of the convergence of the time-space energies.

Without the assumption of superlinear growth the situation is much more subtle: in particular, the lower semicontinuity of $\mathbb{MA}$ obtained minimising the functional $\mathbb{A}$ associated to a function $f$ satisfying only (1.3) is not trivial. This is due to the fact that, in this framework, being a solution to $\mathsf{CE}$ with bounded action only ensures bounds for $\boldsymbol{\mu } \in \mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ , which does not suffice to pass to the limit in the constraint given by the boundary conditions: jumps may occur at $t \in \{ 0,1\}$ in the limit. Therefore, when the cost $F$ grows linearly (linear bounds from both below and above), the scaling behaviour of the discrete boundary-value problems ${\mathcal{MA}}_{\varepsilon }$ , as well as the lower semicontinuity of $\mathbb{MA}$ , cannot be understood with the techniques utilised in [Reference Gladbach, Kopfer, Maas and Portinale23]. The main goal of this work is, thus, to provide discrete-to-continuum results for ${\mathcal{MA}}_{\varepsilon }$ for cost functionals with linear growth, as well as for every flow-based type of cost, i.e. $F(m,J) = F(J)$ . With similar arguments, we can also show the lower semicontinuity of $\mathbb{MA}$ for a general energy density $f$ under the same assumptions, see Section 3.3.

Theorem 1.1 (Main result). Assume that either $F$ satisfies the linear growth condition, i.e.

\begin{align*} F(m,J) \leq C \Bigg ( 1 + \sum _{ \substack{ (x,y) \in \mathcal{E} \\ |x| \leq R } } |J(x,y)| + \sum _{ \substack{ x\in \mathcal{X}\\ |x| \leq R }} m(x) \Bigg ) \end{align*}

for some constant $C\lt \infty$ and some $R \gt 0$ , or that $F$ does not depend on the $m$ -variable (flow-based type). Then, as $\varepsilon \to 0$ , the discrete functionals ${\mathcal{MA}}_{\varepsilon }$ $\Gamma$ -converge to the continuous functional ${\mathbb{MA}}_{\mathrm{hom}}$ with respect to weak convergence.

The contribution of this paper is twofold. On one side, thanks to our main result, we can now include important examples, such as the $\mathbb{W}_1$ distance and related approximations, see in particular Section 4 for some explicit computations of the cell formula, including the equivalence between static and dynamical formulations (4.3), as well as some simulations. As typical in this discrete-to-continuum framework, also for $\mathbb{W}_1$ -type problems, the geometry of the graph plays an important role in the homogenised norm obtained in the limit, giving rise to a whole class of crystalline norms, see Proposition 4.4 as well as Figure 2. On the other hand, this work provides ideas and techniques on how to handle the presence of singularities/jumps in the framework of curves of measures which are only of bounded variation, which is of independent interest.

Related literature. In their seminal work [Reference Jordan, Kinderlehrer and Otto29], Jordan, Kinderlehrer, and Otto showed that the heat flow in $\mathbb{R}^d$ can be seen as the gradient flow of the relative entropy with respect to the $2$ -Wasserstein distance. In the same spirit, a discrete counterpart was proved in [Reference Maas32] and [Reference Mielke33], independently, for the discrete heat flow and discrete relative entropy on Markov chains. In [Reference Forkert, Maas and Portinale19], the authors proved the evolutionary $\Gamma$ -convergence of the discrete gradient-flow structures associated with finite-volume partitions and discrete Fokker–Planck equations, generalising a previous result obtained in [Reference Disser and Liero8] in the setting of isotropic, one-dimensional meshes. Similar results were later obtained in [Reference Hraivoronska and Tse26], [Reference Hraivoronska, Schlichting and Tse27] for the study of the limiting behaviour of random walks on tessellations in the diffusive limit. Generalised gradient-flow structures associated to jump processes and approximation results of nonlocal- and local-interaction equations have been studied in a series of works [Reference Esposito, Patacchini, Schlichting and Slepcev12], [Reference Esposito, Patacchini and Schlichting13], [Reference Esposito, Heinze, Pietschmann and Schlichting10]. Recently, [Reference Esposito and Mikolás17] considered the more general setting where the graph also depends on time.

2. General framework: continuous and discrete transport problems

In this section, we first introduce the general class of problems at the continuous level we are interested in, discussing main properties and known results. We then move to the discrete, periodic framework in the spirit of [Reference Gladbach, Kopfer, Maas and Portinale23], summarise the known convergence results and discuss the open problems we want to treat in this work. In contrast with [Reference Gladbach, Kopfer, Maas and Portinale23], for the sake of the exposition we restrict our analysis to the time interval $\mathcal{I} \,:\!=\, (0,1)$ . Nonetheless, our main results easily extend to a general bounded, open interval $\mathcal{I} \subset \mathbb{R}$ .

2.1. The continuous setting: transport problems on the torus

We start by recalling the definition of solutions to the continuity equation on $\mathbb{T}^d$ .

Definition 2.1 (Continuity equation). A pair of measures $(\boldsymbol{\mu }, \boldsymbol{\nu }) \in \mathcal{M}_+\big ((0,1) \times{\mathbb{T}}^d\big ) \times \mathcal{M}^d\big ((0,1) \times{\mathbb{T}}^d\big )$ is said to be a solution to the continuity equation

\begin{align*} \partial _t \boldsymbol{\mu } + \nabla \cdot \boldsymbol{\nu } = 0 \end{align*}

if, for all functions $\varphi \in \mathcal{C}_c^1\big ((0,1)\times{\mathbb{T}^d}\big )$ , the identity

\begin{align*} \int _{(0,1) \times{\mathbb{T}^d}} \partial _t \varphi{\, \mathrm{d}} \boldsymbol{\mu } + \int _{(0,1) \times{\mathbb{T}^d}} \nabla \varphi \cdot{\, \mathrm{d}} \boldsymbol{\nu } = 0 \end{align*}

holds. We use the notation $(\boldsymbol{\mu }, \boldsymbol{\nu }) \in \mathsf{CE}$ .

Throughout the whole paper, we consider energy densities $f$ with the following properties.

Assumption 2.2. Let $f : \mathbb{R}_+ \times \mathbb{R}^d \to \mathbb{R} \cup \{+\infty \}$ be a lower semicontinuous and convex function, whose domain $\mathsf{D}(f)$ has nonempty interior. We assume that there exist constants $c \gt 0$ and $C \lt \infty$ such that the (at least) linear growth condition

(2.1) \begin{align} f(\rho, j) \geq c |j| - C (\rho + 1) \end{align}

holds for all $\rho \in \mathbb{R}_+$ and $j \in \mathbb{R}^d$ .

The corresponding recession function $f^\infty : \mathbb{R}_+ \times \mathbb{R}^d \to \mathbb{R} \cup \{+\infty \}$ is defined by

(2.2) \begin{align} f^\infty (\rho, j) \,:\!=\, \lim _{t \to + \infty } \frac{f(\rho _0 + t \rho, j_0 + t j)}{t}, \end{align}

for every $(\rho _0, j_0) \in \mathsf{D}(f)$ . It is well established that the function $f^\infty$ is lower semicontinuous, convex, and it satisfies the inequality

(2.3) \begin{align} f^\infty (\rho, j) \geq c|j| - C\rho, \qquad \rho \in \mathbb{R}_+, \; j \in \mathbb{R}^d, \end{align}

see [Reference Ambrosio, Fusco and Pallara1, Section 2.6].

Let $\mathscr{L}^{\,d+1}$ denote the Lebesgue measure on $(0,1) \times{\mathbb{T}}^d$ . For $\boldsymbol{\mu } \in \mathcal{M}_+\big ( (0,1) \times{\mathbb{T}}^d\big )$ and $\boldsymbol{\nu } \in \mathcal{M}^d\big ( (0,1) \times{\mathbb{T}}^d\big )$ , we write their Lebesgue decompositions as

\begin{align*} \boldsymbol{\mu } = \rho \mathscr{L}^{\,d+1} + \boldsymbol{\mu }^\perp, \qquad \boldsymbol{\nu } = j \mathscr{L}^{\,d+1} + \boldsymbol{\nu }^\perp, \end{align*}

for some $\rho \in L_+^1\big ( (0,1) \times{\mathbb{T}}^d \big )$ and $j \in L^1\big ( (0,1) \times{\mathbb{T}}^d; \mathbb{R}^d\big )$ . Given these decompositions, there always exists a measure $\boldsymbol{\sigma } \in \mathcal{M}_+\bigl ( (0,1) \times{\mathbb{T}}^d\bigr )$ such that

\begin{align*} \boldsymbol{\mu }^\perp = \rho ^\perp \boldsymbol{\sigma }, \qquad \boldsymbol{\nu }^\perp = j^\perp \boldsymbol{\sigma }, \end{align*}

for some $\rho ^\perp \in L_+^1(\boldsymbol{\sigma })$ and $j^\perp \in L^1(\boldsymbol{\sigma };\mathbb{R}^d)$ (take for example $\boldsymbol{\sigma } \,:\!=\, |\boldsymbol{\mu }^\perp | + |\boldsymbol{\nu }^\perp |$ ).

Definition 2.3 (Action functionals). We define the action functionals by

\begin{align*} &{\mathbb{A}} : \mathcal{M}_+\big ( (0,1) \times{\mathbb{T}^d} \big ) \times \mathcal{M}^d\big ( (0,1) \times{\mathbb{T}}^d\big ) \to \mathbb{R} \cup \{ +\infty \}, \\ &{\mathbb{A}}(\boldsymbol{\mu }, \boldsymbol{\nu }) \,:\!=\, \int _{(0,1) \times{\mathbb{T}^d}} f \big ( \rho, j \big ){\, \mathrm{d}} \mathscr{L}^{\,d+1} + \int _{ (0,1) \times{\mathbb{T}^d}} f^\infty \big ( \rho ^\perp, j^\perp \big ){\, \mathrm{d}} \boldsymbol{\sigma }, \\ &{\mathbb{A}}(\boldsymbol{\mu }) \,:\!=\, \inf _{\boldsymbol{\nu }} \left \{{\mathbb{A}}(\boldsymbol{\mu }, \boldsymbol{\nu }) \, : \, (\boldsymbol{\mu }, \boldsymbol{\nu }) \in \mathsf{CE} \right \} . \end{align*}

Remark 2.4. This definition does not depend on the choice of $\boldsymbol{\sigma }$ , due to the $1$ -homogeneity of $f^\infty$ . As $f(\rho, j) \geq - C(1+\rho )$ and $f^\infty (\rho, j) \geq - C \rho$ from (2.1) and (2.3), the fact that $\boldsymbol{\mu }\bigl ((0,1) \times{\mathbb{T}^d} \bigr ) \lt \infty$ ensures that ${\mathbb{A}}(\boldsymbol{\mu }, \boldsymbol{\nu })$ is well defined in $\mathbb{R} \cup \{ + \infty \}$ .

The natural setting to work in is the space $\mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ of the curves of measures $\boldsymbol{\mu } : (0,1) \to \mathcal{M}_+({\mathbb{T}^d})$ such that the $\mathrm{BV}$ -seminorm $\| \boldsymbol{\mu } \| = \|\boldsymbol{\mu } \|_{\mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )}$ defined by

\begin{align*} \|\boldsymbol{\mu } \| \,:\!=\, \sup \left \{ \int _{(0,1)} \int _{{\mathbb{T}}^d} \partial _t \varphi _t{\, \mathrm{d}} \mu _t{\, \mathrm{d}} t \ : \ \varphi \in \mathcal{C}_c^1\big ((0,1); \mathcal{C}^1({\mathbb{T}^d})\big ), \ \max _{t\in (0,1)} \|\varphi _t\|_{\mathcal{C}^1({\mathbb{T}^d})} \leq 1 \right \} \end{align*}

is finite. Note that, by the trace theorem in BV, curves of measures in $\mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ have a well-defined trace at $t = 0$ and $t=1$ . As shown in [[Reference Gladbach, Kopfer, Maas and Portinale23], Lemma3.13], any solution $(\boldsymbol{\mu }, \boldsymbol{\nu }) \in \mathsf{CE}$ can be disintegrated as ${\, \mathrm{d}} \boldsymbol{\mu }(t,x) ={\, \mathrm{d}} \mu _t(x){\, \mathrm{d}} t$ for some measurable curve $t \mapsto \mu _t \in \mathcal{M}_+({\mathbb{T}}^d)$ with finite constant mass. If ${\mathbb{A}}(\boldsymbol{\mu }) \lt \infty$ , then this curve belongs to $\mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ and

\begin{align*} \| \boldsymbol{\mu } \|_{\mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )} \leq |\boldsymbol{\nu }| \big ( (0,1) \times{\mathbb{T}^d} \big ) . \end{align*}

2.2. Boundary conditions and lower semicontinuity

Define the minimal homogenised action for $\mu _0,\mu _1\in \mathcal{M}_+({\mathbb{T}^d})$ with $\mu _0({\mathbb{T}^d}) = \mu _1({\mathbb{T}^d})$ as

(2.4) \begin{align} {\mathbb{MA}}(\mu _0,\mu _1) \,:\!=\, \inf _{\boldsymbol{\mu } \in \mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )}\left \{{\mathbb{A}}(\boldsymbol{\mu })\,:\,\boldsymbol{\mu }_{t=0} = \mu _0, \boldsymbol{\mu }_{t=1} = \mu _1\right \} . \end{align}

Note that, in general, $\mathbb{MA}$ may be infinite (although the measures have equal masses). Despite the lower semicontinuity property of $\mathbb{A}$ (cfr. [[Reference Gladbach, Kopfer, Maas and Portinale23], Lemma 3.14]), the lower semicontinuity of $\mathbb{MA}$ with respect to the natural weak topology of $\mathcal{M}_+({\mathbb{T}^d}) \times \mathcal{M}_+({\mathbb{T}^d})$ is, in general, nontrivial. More precisely, it is a challenging question to prove (or disprove) that for any two sequences $\mu _0^n$ , $\mu _1^n \in \mathcal{M}_+({\mathbb{T}^d})$ , such that $\mu _i^n \to \mu _i$ weakly in $\mathcal{M}_+({\mathbb{T}^d})$ as $n \to \infty$ for $i=0,1$ , the inequality

(2.5) \begin{align} \liminf _{n \to \infty }{\mathbb{MA}}(\mu _0^n, \mu _1^n) \geq{\mathbb{MA}}(\mu _0,\mu _1) \end{align}

holds. In this work, we provide a positive answer in the case when $f$ has linear growth or it is flow-based (i.e. it does not depend on the first variable), see Remark 3.14 and Proposition 3.15 below. First, we discuss the main challenges and the setup where the lower semicontinuity is already known to hold.

Remark 2.5 (Lack of compatible compactness). We know from [[Reference Gladbach, Kopfer, Maas and Portinale23], Lemma 3.14] that $(\boldsymbol{\mu },\boldsymbol{\nu }) \mapsto{\mathbb{A}}(\boldsymbol{\mu },\boldsymbol{\nu })$ and $\boldsymbol{\mu } \mapsto{\mathbb{A}}(\boldsymbol{\mu })$ are lower semicontinuous w.r.t. the weak topology. Moreover, if $\boldsymbol{\mu }^n$ is a sequence with

\begin{align*} \sup _n{\mathbb{A}}(\boldsymbol{\mu }^n) \lt \infty \quad \text{and} \quad \sup _n \boldsymbol{\mu }^n\bigl ((0,1) \times{\mathbb{T}}^d\bigr ) \lt \infty \end{align*}

then $\boldsymbol{\mu }^n$ is weakly compact and any limit $\boldsymbol{\mu }$ belongs to $\mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ . This can be proved as in [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.4]. Nonetheless, this property does not ensure the lower semicontinuity of $\mathbb{MA}$ because weak convergence does not preserve the boundary conditions (at time $t=0$ and $t=1$ ). For similar issues in the setting of functionals of $\mathbb{R}^d$ -valued curves with bounded variations and their minimisation, see e.g. [Reference Amar and Vitali2].

Remark 2.6 (Superlinear growth). Under the strengthened assumption of superlinear growth on $f$ (with respect to the momentum variable), it is possible to prove the lower semicontinuity property (2.5), in the same way as in the proof of the discrete-to-continuum $\Gamma$ -convergence of boundary-value problems of [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.10]. More precisely, we say that $f$ is of superlinear growth if there exists a function $\theta :[0,\infty ) \to [0,\infty )$ with $\lim _{t \to \infty } \frac{\theta (t)}{t} = \infty$ and a constant $C \in \mathbb{R}$ such that

\begin{align*} f(\rho, j) \geq (\rho +1) \theta \left ( \frac{\lvert j\rvert }{\rho +1} \right ) - C (\rho +1), \qquad \forall \rho \in \mathbb{R}_+, \ \ j \in \mathbb{R}^d . \end{align*}

Arguing as in [[Reference Gladbach, Kopfer, Maas and Portinale23], Remark 5.6], one shows that any function of superlinear growth must satisfy the growth condition given by Assumption 2.2. Moreover, in this case, the recession function satisfies $f^\infty (0,j)=+\infty$ , for every $j \neq 0$ . See [[Reference Gladbach, Kopfer, Maas and Portinale23], Examples 5.7 & 5.8] for some examples belonging to this class. By arguing similarly as in the proof of [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.9], assuming superlinear growth one can show that if $\boldsymbol{\mu }^n$ is a sequence with bounded action ${\mathbb{A}}(\boldsymbol{\mu }^n)$ and bounded total mass $\boldsymbol{\mu }^n \bigl ((0,1) \times{\mathbb{T}}^d \bigr )$ , then, up to a (non-relabelled) subsequence, we have $\boldsymbol{\mu }^n \to \boldsymbol{\mu }$ in $\mathcal{M}_+((0,1)\times{\mathbb{T}^d})$ and $\mu _t^n \to \mu _t$ in $\mathrm{KR}$ norm uniformly in $t \in (0,1)$ , with limit curve $\boldsymbol{\mu } \in W_{\mathrm{KR}}^{1,1}((0,1); \mathcal{M}_+({\mathbb{T}^d}))$ . Using this fact, it is clear that the problem of “jumps” in the limit explained in Remark 2.5 does not occur, and the lower semicontinuity (2.5) directly follows from the lower semicontinuity of $\mathbb{A}$ .

Remark 2.7. (Nonnegativity) Without loss of generality, we can assume that $f \geq 0$ . Indeed, thanks to the linear growth assumption 2.2, we can define a new function

(2.6) \begin{align} \widetilde f(\rho, j) \,:\!=\, f(\rho, j) + C(\rho +1) \geq c|j| \geq 0 \end{align}

which is now nonnegative and with (at least) linear growth. Furthermore, we can compute the recess $\widetilde f^\infty$ and from the definition we see that

(2.7) \begin{align} \widetilde f^\infty (\rho, j) = f^\infty (\rho, j) + C \rho . \end{align}

Denote by $\widetilde{\mathbb{A}}$ the action functional obtained by replacing $f$ with $\widetilde f$ . Thanks to (2.6), (2.7), we have that

\begin{align*} \widetilde{\mathbb{A}}(\boldsymbol{\mu }) \,:\!=\, &\inf _{\boldsymbol{\nu }} \left \{ \widetilde{\mathbb{A}}(\boldsymbol{\mu }, \boldsymbol{\nu }) \, : \, (\boldsymbol{\mu }, \boldsymbol{\nu }) \in \mathsf{CE} \right \}\\ = &\inf _{\boldsymbol{\nu }} \left \{{\mathbb{A}}(\boldsymbol{\mu }, \boldsymbol{\nu }) \, : \, (\boldsymbol{\mu }, \boldsymbol{\nu }) \in \mathsf{CE} \right \} + C(\boldsymbol{\mu }\big ( (0,1) \times{\mathbb{T}}^d\big )+1). \end{align*}

It follows that the corresponding boundary value problems are given by

\begin{align*} \widetilde{\mathbb{MA}}(\mu _0, \mu _1) ={\mathbb{MA}}(\mu _0, \mu _1) + C(\mu _0({\mathbb{T}^d})+1), \quad \text{if } \mu _0({\mathbb{T}^d}) = \mu _1({\mathbb{T}^d}) . \end{align*}

Therefore, the (weak) lower semicontinuity for $\widetilde{\mathbb{MA}}$ is equivalent to that of $\mathbb{MA}$ .

2.3. The discrete framework: transport problems on periodic graphs

We recall the framework of [Reference Gladbach, Kopfer, Maas and Portinale23]: let $(\mathcal{X}, \mathcal{E})$ be a locally finite and $\mathbb{Z}^d$ -periodic connected graph of bounded degree. We encode the set of vertices as $\mathcal{X} = \mathbb{Z}^d \times{V}$ , where $V$ is a finite set, and we use coordinates $x = (x_{\mathsf{z}}, x_{\mathsf{v}}) \in \mathcal{X}$ . The set of edges $\mathcal{E} \subseteq \mathcal{X} \times \mathcal{X}$ is symmetric and $\mathbb{Z}^d$ -periodic, and we use the notation $x \sim y$ whenever $(x,y) \in \mathcal{E}$ . Let $R_0 \,:\!=\, \max _{(x,y) \in \mathcal{E}} | x_{\mathsf{z}} - y_{\mathsf{z}}|_{\infty }$ be the maximal edge length in the supremum norm $|\cdot |_{\infty }$ on $\mathbb{R}^d$ . We use the notation ${\mathcal{X}^Q} \,:\!=\, \{ x \in \mathcal{X} \, : \, x_{\mathsf{z}} = 0 \}$ and $ \mathcal{E}^Q \,:\!=\, \big \{ (x, y) \in \mathcal{E} \, : \, x_{\mathsf{z}} = 0 \big \}.$ For a discussion concerning abstract and embedded graphs, see [[Reference Gladbach, Kopfer, Maas and Portinale23], Remark 2.2].

In what follows, we denote by $\mathbb{R}^{\mathcal{X}}_+$ the set of functions $m \colon \mathcal{X} \to \mathbb{R}_+$ and by $\mathbb{R}^{\mathcal{E}}_a$ the set of anti-symmetric functions $J \colon \mathcal{E} \to \mathbb{R}$ , that is, such that $J(x,y) = -J(y,x)$ . The elements of $\mathbb{R}^{\mathcal{E}}_a$ will often be called (discrete) vector fields.

Assumption 2.8 (Admissible cost function). The function $F : \mathbb{R}_+^{\mathcal{X}} \times \mathbb{R}_a^{\mathcal{E}} \to \mathbb{R} \cup \{+\infty \}$ is assumed to have the following properties:

  1. (a) $F$ is convex and lower semicontinuous.

  2. (b) $F$ is local, meaning that, for some number $R_1 \lt \infty$ , we have $F(m,J) = F(m',J')$ whenever $m, m' \in \mathbb{R}_+^{\mathcal{X}}$ and $J, J' \in \mathbb{R}_a^{\mathcal{E}}$ agree within a ball of radius $R_1$ , i.e.

    \begin{align*} m(x) & = m'(x) && \text{for all } x \in \mathcal{X} \text{ with } |x_{\mathsf{z}}|_{\infty } \leq R_1, \quad and \\ J(x,y) & = J'(x,y) && \text{for all } (x,y) \in \mathcal{E} \text{ with } |x_{\mathsf{z}}|_{\infty }, |y_{\mathsf{z}}|_{\infty } \leq R_1 . \end{align*}
  3. (c) $F$ is of at least linear growth, i.e. there exist $c \gt 0$ and $C \lt \infty$ such that

    (2.8) \begin{align} F(m,J) \geq c \sum _{ (x, y) \in \mathcal{E}^Q } |J(x,y)| - C \Bigg ( 1 + \sum _{\substack{ x\in \mathcal{X}\\ |x_{\mathsf{z}}|_{\infty } \leq R_{\mathrm{max}} }} m(x) \Bigg ) \end{align}
    for any $m \in \mathbb{R}_+^{\mathcal{X}}$ and $J \in \mathbb{R}_a^{\mathcal{E}}$ . Here, $R_{\mathrm{max}} \,:\!=\, \max \{R_0, R_1\}$ .
  4. (d) There exist a $\mathbb{Z}^d$ -periodic function $m^\circ \in \mathbb{R}_+^{\mathcal{X}}$ and a $\mathbb{Z}^d$ -periodic and divergence-free vector field $J^\circ \in \mathbb{R}_a^{\mathcal{E}}$ such that

    \begin{align*} ( m^\circ, J^\circ ) \in \mathsf{D}(F)^\circ . \end{align*}

Remark 2.9. Important examples that satisfy the growth condition (2.8) are of the form

\begin{align*} F(m,J) = \frac 12 \sum _{(x,y) \in \mathcal{E}^Q} \frac{|J(x,y)|^p}{\Lambda \big (q_{xy} m(x), q_{yx} m(y)\big )^{p-1}}, \end{align*}

where $1 \leq p \lt \infty$ , the constants $q_{xy}, q_{yx} \gt 0$ are fixed parameters defined for $(x,y)\in \mathcal{E}^Q$ , and $\Lambda$ is a suitable mean. Functions of this type naturally appear in discretisations of Wasserstein gradient-flow structures [Reference Maas32], [Reference Mielke33], [Reference Chow, Huang, Li and Zhou6], see also [[Reference Gladbach, Kopfer, Maas and Portinale23], Remark2.6].

The rescaled graph. Let ${\mathbb{T}}_{\varepsilon }^d = (\varepsilon \mathbb{Z} / \mathbb{Z})^d$ be the discrete torus of mesh size $\varepsilon \in 1/\mathbb{N}$ . We denote by $[\varepsilon z]$ for $z \in \mathbb{Z}^d$ the corresponding equivalence classes. Equivalently, ${\mathbb{T}}_{\varepsilon }^d = \varepsilon \mathbb{Z}_{\varepsilon }^d$ where $\mathbb{Z}_{\varepsilon }^d = \big (\mathbb{Z} / \tfrac 1\varepsilon \mathbb{Z} \big )^d$ . The rescaled graph $(\mathcal{X}_{\varepsilon }, \mathcal{E}_{\varepsilon })$ is defined as

\begin{align*} \mathcal{X}_{\varepsilon } \,:\!=\,{\mathbb{T}}_{\varepsilon }^d \times{V} \quad \text{and}\quad \mathcal{E}_{\varepsilon } \,:\!=\, \big \{ \big ( T_{\varepsilon }^0 (x), T_{\varepsilon }^0 (y) \big ) \ : \ (x, y) \in \mathcal{E} \big \} \end{align*}

where for $\bar z \in \mathbb{Z}_{\varepsilon }^d$ ,

\begin{align*} T_{\varepsilon }^{\bar z} : \mathcal{X} \to \mathcal{X}_{\varepsilon }, \qquad (z, v) \mapsto \big ( [\varepsilon (\bar z + z)], v \big ) . \end{align*}

For $x = \big ([\varepsilon z], v\big ) \in \mathcal{X}_{\varepsilon }$ , we write

\begin{align*} x_{\mathsf{z}} \,:\!=\, z \in \mathbb{Z}_{\varepsilon }^d, \qquad x_{\mathsf{v}} \,:\!=\, v\in{V} . \end{align*}

The equivalence relation $\sim$ on $\mathcal{X}$ is equivalently defined on $\mathcal{X}_{\varepsilon }$ by means of $\mathcal{E}_{\varepsilon }$ . Hereafter, we always assume that $\varepsilon R_0 \lt \frac 12$ .

The rescaled energies. Let $F : \mathbb{R}_+^{\mathcal{X}} \times \mathbb{R}_a^{\mathcal{E}} \to \mathbb{R} \cup \{ + \infty \}$ be a cost function satisfying Assumption 2.8. For $\varepsilon \gt 0$ satisfying the conditions above, we can define a corresponding energy functional $\mathcal{F}_{\varepsilon }$ in the rescaled periodic setting: following [Reference Gladbach, Kopfer, Maas and Portinale23], for $\bar z \in \mathbb{Z}_{\varepsilon }^d$ , each function $\psi : \mathcal{X}_{\varepsilon } \to \mathbb{R}$ induces a $\frac{1}{\varepsilon }\mathbb{Z}^d$ -periodic function

\begin{align*} \tau _{\varepsilon }^{\bar z} \psi : \mathcal{X} \to \mathbb{R}, \qquad \big (\tau _{\varepsilon }^{\bar z} \psi \big )(x) \,:\!=\, \psi \big ( T_{\varepsilon }^{\bar z}(x) \big ) \quad \text{ for } x \in \mathcal{X} . \end{align*}

Similarly, each function $J : \mathcal{E}_{\varepsilon } \to \mathbb{R}$ induces a $\frac{1}{\varepsilon }\mathbb{Z}^d$ -periodic function

\begin{align*} \tau _{\varepsilon }^{\bar z} J : \mathcal{E} \to \mathbb{R}, \qquad \big ( \tau _{\varepsilon }^{\bar z} J \big )(x,y) \,:\!=\, J \big ( T_{\varepsilon }^{\bar z}(x), T_{\varepsilon }^{\bar z}(y) \big ) \quad \text{ for } (x,y) \in \mathcal{E} . \end{align*}

Definition 2.10 (Discrete energy functional). The rescaled energy is defined by

\begin{align*} \mathcal{F}_{\varepsilon } : \mathbb{R}_+^{\mathcal{X}_{\varepsilon }} \times \mathbb{R}_a^{\mathcal{E}_{\varepsilon }} &\to \mathbb{R} \cup \{ + \infty \}, \qquad (m, J) \stackrel{\mathcal{F}_{\varepsilon }}{\longmapsto } \sum _{z \in \mathbb{Z}_{\varepsilon }^d} \varepsilon ^d F\bigg ( \frac{\tau _{\varepsilon }^z m}{\varepsilon ^d}, \frac{\tau _{\varepsilon }^z J}{\varepsilon ^{d-1}} \bigg ) . \end{align*}

Remark 2.11. As observed in [[Reference Gladbach, Kopfer, Maas and Portinale23], Remark 2.8], the value $\mathcal{F}_{\varepsilon }(m,J)$ is well defined as an element in $\mathbb{R} \cup \{ + \infty \}$ . Indeed, the condition (2.8) yields

\begin{align*} \mathcal{F}_{\varepsilon }( m, J ) = \sum _{z \in \mathbb{Z}_{\varepsilon }^d} \varepsilon ^d F\bigg ( \frac{\tau _{\varepsilon }^z m}{\varepsilon ^d}, \frac{\tau _{\varepsilon }^z J}{\varepsilon ^{d-1}} \bigg ) & \geq - C \sum _{z \in \mathbb{Z}_{\varepsilon }^d} \varepsilon ^d \Bigg ( 1 + \sum _{\substack{ x\in \mathcal{X}\\ |x_{\mathsf{z}}|_{\infty } \leq R_{\mathrm{max}} }} \frac{\tau _{\varepsilon }^z m(x)}{\varepsilon ^d} \Bigg ) \\& \geq - C \bigg ( 1 + (2R_{\mathrm{max}} + 1)^d \sum _{x \in \mathcal{X}_{\varepsilon }} m(x) \bigg ) \gt - \infty . \end{align*}

Definition 2.12 (Discrete continuity equation). A pair $({\boldsymbol{m}}, \boldsymbol{J})$ is said to be a solution to the discrete continuity equation if ${\boldsymbol{m}} : (0,1) \to \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ is continuous, $\boldsymbol{J} : (0,1) \to \mathbb{R}_a^{\mathcal{E}_{\varepsilon }}$ is Borel measurable, and

(2.9) \begin{align} \partial _t m_t(x) + \sum _{y\sim x}J_t(x,y) = 0 \end{align}

holds for all $x \in \mathcal{X}_{\varepsilon }$ in the sense of distributions. We use the notation

\begin{align*} ({\boldsymbol{m}}, \boldsymbol{J}) \in \mathcal{CE}_{\varepsilon } . \end{align*}

Remark 2.13. We may write (2.9) as $ \partial _t m_t + \mathsf{div} J_t = 0$ using the discrete divergence operator, given by

\begin{align*} \mathsf{div} J \in \mathbb{R}^{\mathcal{X}_{\varepsilon }}, \qquad \mathsf{div} J(x) \,:\!=\, \sum _{y \sim x} J(x,y), \qquad \forall J \in \mathbb{R}_a^{\mathcal{E}_{\varepsilon }} . \end{align*}

The proof of the following lemma can be found in [Reference Gladbach, Kopfer, Maas and Portinale23].

Lemma 2.14 (Mass preservation). Let $({\boldsymbol{m}}, \boldsymbol{J}) \in \mathcal{CE}_{\varepsilon }$ . Then we have $m_s(\mathcal{X}_{\varepsilon }) = m_t(\mathcal{X}_{\varepsilon })$ for all $s, t \in (0,1)$ .

We are now ready to define one of the main objects in this paper.

Definition 2.15 (Discrete action functional). For any continuous function ${\boldsymbol{m}}: (0,1) \to \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ such that $t \mapsto \sum _{x \in \mathcal{X}_{\varepsilon }} m_t(x) \in L^1\bigl ((0,1)\bigr )$ and any Borel measurable function $\boldsymbol{J}: (0,1) \to \mathbb{R}_a^{\mathcal{E}_{\varepsilon }}$ , we define

\begin{align*} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}, \boldsymbol{J}) & \,:\!=\, \int _0^1 \mathcal{F}_{\varepsilon }( m_t, J_t ){\, \mathrm{d}} t \in \mathbb{R} \cup \{ + \infty \} . \end{align*}

Furthermore, we set

\begin{align*} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}) & \,:\!=\, \inf _{\boldsymbol{J}} \Big \{ \mathcal{A}_{\varepsilon }({\boldsymbol{m}}, \boldsymbol{J}) \ : \ ({\boldsymbol{m}}, \boldsymbol{J}) \in \mathcal{CE}_{\varepsilon } \Big \} . \end{align*}

Arguing as in Remark 2.11, one can show [[Reference Gladbach, Kopfer, Maas and Portinale23], Remark2.13] that $\mathcal{A}_{\varepsilon }({\boldsymbol{m}}, \boldsymbol{J})$ is well defined as an element in $\mathbb{R} \cup \{ + \infty \}$ , as a consequence of the growth condition (2.8).

Definition 2.16 (Minimal discrete action functional). For any pair of boundary data $m_0$ , $m_1 \in \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ , we define the associated discrete boundary value problem as

\begin{align*}{\mathcal{MA}}_{\varepsilon }(m_0,m_1) & \,:\!=\, \inf \left \{ \mathcal{A}_{\varepsilon }({\boldsymbol{m}} ) \, : \,{\boldsymbol{m}}: (0,1) \to \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}, \quad{\boldsymbol{m}}_{t=0} = m_0 \quad \text{ and }\quad{\boldsymbol{m}}_{t=1} = m_1 \right \} . \end{align*}

The aim of this work is to study the asymptotic behaviour of the energies ${\mathcal{MA}}_{\varepsilon }$ as $\varepsilon \to 0$ under the Assumption 2.8.

3. Statement and proof of the main result

In this paper, we extend the $\Gamma$ -convergence result for the functionals ${\mathcal{MA}}_{\varepsilon }$ towards ${\mathbb{MA}}_{\mathrm{hom}}$ , proved in [Reference Gladbach, Kopfer, Maas and Portinale23] for superlinear cost functionals, to two cases: under the assumption of linear growth (bound both from below and above) and when the function $F$ does not depend on $\rho$ .

Assumption 3.1 (Linear growth). We say that a function $F : \mathbb{R}_+^{\mathcal{X}} \times \mathbb{R}_a^{\mathcal{E}} \to \mathbb{R} \cup \{+\infty \}$ has linear growth if it satisfies

\begin{align*} F(m,J) \leq C \Bigg ( 1 + \sum _{ \substack{ (x,y) \in \mathcal{E} \\ |x_{\mathsf{z}}|_{\infty } \leq R } } |J(x,y)| + \sum _{ \substack{ x\in \mathcal{X}\\ |x_{\mathsf{z}}|_{\infty } \leq R }} m(x) \Bigg ) \end{align*}

for some constant $C\lt \infty$ and some $R \gt 0$ .

Assumption 3.2 (Flow-based). We say that a function $F : \mathbb{R}_+^{\mathcal{X}} \times \mathbb{R}_a^{\mathcal{E}} \to \mathbb{R} \cup \{+\infty \}$ is of flow-based type if it depends only on the the second variable, i.e. (with a slight abuse of notation) $F(m,J) = F(J)$ , for some $F: \mathbb{R}_a^{\mathcal{E}} \to \mathbb{R} \cup \{ +\infty \}$ .

Similarly, we say that $f:\mathbb{R}_+ \times \mathbb{R}^d \to \mathbb{R}$ is of flow-based type if it does not depend on the $\rho$ variable, i.e. $f(\rho, j) = f(j)$ . In this case, the problem simplifies significantly, and the dynamical variational problem described in (2.4) admits an equivalent, static formulation (see (3.23)).

Remark 3.3 (Linear growth vs Lipschitz). While working with convex functions, to assume a linear growth condition (from above) is essentially equivalent to require Lipschitz continuity with respect to the second variable.

Lemma 3.4 (Lipschitz continuity). Let $f: \mathbb{R}_+ \times \mathbb{R}^d \to \mathbb{R}$ be a function, convex in the second variable. Let $C \gt 0$ . Then the following are equivalent:

  1. 1. For every $\rho \in \mathbb{R}_+$ and $j \in \mathbb{R}^d$ the inequality $f(\rho, j) \le C(1+ \rho + \lvert j\rvert )$ holds.

  2. 2. For every $\rho \in \mathbb{R}_+$ , the function $f(\rho, \cdot ) : \mathbb{R}^d \to \mathbb{R}_+$ is Lipschitz continuous (uniformly in $\rho$ ) with constant $C$ , and the inequality $f(\rho, 0) \le C(1+ \rho )$ holds.

In the very same spirit, one can show the analogous result at the discrete level.

Lemma 3.5 (Lipschitz continuity II). Let $F : \mathbb{R}_+^{\mathcal{X}} \times \mathbb{R}_a^{\mathcal{E}} \to \mathbb{R} \cup \{+\infty \}$ be convex in the second variable. Let $C,R \gt 0$ . Then the following are equivalent:

  1. 1. $F$ is of linear growth, in the sense of Assumption 3.1 , with the same constants $C$ and $R$ .

  2. 2. For every $m \in \mathbb{R}_+^{\mathcal{X}}$ , we have that

    \begin{align*} F(m,0) \leq C \bigg ( 1 + \sum _{ \substack{ x\in \mathcal{X}\\ |x_{\mathsf{z}}|_{\infty } \leq R }} m(x) \bigg ), \end{align*}
    as well as that $F$ is Lipschitz continuous with constant $C$ in the second variable, in the sense that
    (3.1) \begin{align} \left | F(m,J_1) - F(m,J_2) \right | \leq C \sum _{ \substack{ (x,y) \in \mathcal{E} \\ |x_{\mathsf{z}}|_{\infty } \leq R } } |J_1(x,y)- J_2(x,y)|, \end{align}
    for every $J_1, J_2 \in \mathbb{R}_a^{\mathcal{E}}$ .

Proof of Lemma 3.4. Let us assume the first condition and fix $\rho \in \mathbb{R}_+$ as well as $j_1,j_2 \in \mathbb{R}^d$ . It follows from the convexity in the second variable that the function

\begin{align*} \mathbb{R} \ni t \mapsto f(\rho, j_1 + t(j_2-j_1)) \end{align*}

is convex. In particular, the inequalities

\begin{align*} f(\rho, j_2) - f(\rho, j_1) \le \frac{f(\rho, j_1 + t(j_2-j_1)) - f(\rho, j_1)}{t} \le \frac{ C \big ( 1+\rho + \big | j_1+t(j_2-j_1) \big | \big ) - f(\rho, j_1)}{t} \end{align*}

hold for every $t \ge 1$ . Letting $t \to \infty$ , we thus find

\begin{align*} f(\rho, j_2) - f(\rho, j_1) \le C \lvert j_2-j_1\rvert \end{align*}

and, by arbitrariness of the arguments, the claimed Lipschitz continuity. The fact that $f(\rho, 0) \le C(1+ \rho )$ trivially follows from the first condition.

Conversely, if the second condition holds, we necessarily have

\begin{align*} f(\rho, j) \le C \lvert j\rvert + f(\rho, 0) \le C(1+ \rho + \lvert j\rvert ), \end{align*}

for every $\rho \in \mathbb{R}_+$ and $j \in \mathbb{R}^d$ , which is precisely the first condition in the statement.

Let us recall the homogenised energy density $f_{\mathrm{hom}}$ , which describes the limit energy and is given by a cell formula. For given $\rho \geq 0$ and $j \in \mathbb{R}^d$ , $f_{\mathrm{hom}}(\rho, j)$ is obtained by minimising over the unit cube the cost among functions $m$ and vector fields $J$ representing $\rho$ and $j$ . More precisely, the function $f_{\mathrm{hom}}:\mathbb{R}_+ \times \mathbb{R}^d \to \mathbb{R}_+$ is given by

\begin{align*} f_{\mathrm{hom}}(\rho, j) \,:\!=\, \inf _{m,J} \bigl \{ F(m,J) \, : \, (m,J) \in{\mathsf{Rep}}(\rho, j) \bigr \}, \end{align*}

where the set of representatives ${\mathsf{Rep}}(\rho, j)$ consists of all $\mathbb{Z}^d$ -periodic functions $m: \mathcal{X} \to \mathbb{R}_+$ and all $\mathbb{Z}^d$ -periodic anti-symmetric discrete vector fields $J : \mathcal{E} \to \mathbb{R}$ satisfying

(3.2) \begin{align} \sum _{x \in \mathcal{X}^Q} m(x) = \rho, \quad \mathsf{div} J = 0, \quad \text{ and }\quad \mathsf{Eff} (J) \,:\!=\, \frac 12 \sum _{(x,y) \in \mathcal{E}^Q} J(x,y) (y_{\mathsf{z}}-x_{\mathsf{z}}) = j . \end{align}

The set of representatives is nonempty for every choice of $\rho$ and $j$ by [[Reference Gladbach, Kopfer, Maas and Portinale23], Lemma 4.5 (iv)]. In the case of embedded graphs, the definition of effective flux coincides with the one provided in the introduction (cfr. (3.2)), see [[Reference Gladbach, Kopfer, Maas and Portinale23], Proposition 9.1].

Remark 3.6. It is not hard to show that if $F$ is of linear growth, then $f_{\mathrm{hom}}$ is also of linear growth (and therefore, in view of Lemma 3.4, it is Lipschitz in the second variable uniformly w.r.t. the first one), see e.g. [Reference Gladbach, Maas and Portinale25].

We denote by ${\mathbb{A}}_{\mathrm{hom}}$ and ${\mathbb{MA}}_{\mathrm{hom}}$ the action functionals corresponding to the choice $f=f_{\mathrm{hom}}$ . In order to talk about $\Gamma$ -convergence, we need to specificy which type of discrete-to-continuum topology/convergence we adopt (in the same spirit of [Reference Gladbach, Kopfer, Maas and Portinale23]).

Definition 3.7 (Embedding). For $\varepsilon \gt 0$ and $z \in \mathbb{R}^d$ , let $Q_{\varepsilon }^z \,:\!=\, \varepsilon z + [0, \varepsilon )^d \subseteq{\mathbb{T}}^d$ denote the projection of the cube with side length $\varepsilon$ based at $\varepsilon z$ to the quotient ${\mathbb{T}}^d = \mathbb{R}^d/\mathbb{Z}^d$ . For $m \in \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ and $J \in \mathbb{R}_a^{\mathcal{E}_{\varepsilon }}$ , we define $\iota _{\varepsilon } m \in \mathcal{M}_+({\mathbb{T}^d})$ and $\iota _{\varepsilon } J \in \mathcal{M}^d({\mathbb{T}}^d)$ by

\begin{align*} \iota _{\varepsilon } m &\,:\!=\, \varepsilon ^{-d} \sum _{x \in \mathcal{X}_{\varepsilon }} m(x) \mathscr{L}^d|_{Q_{\varepsilon }^{x_{\mathsf{z}}}}, \end{align*}
\begin{align*} \iota _{\varepsilon } J & \,:\!=\, \varepsilon ^{-d+1} \sum _{(x,y) \in \mathcal{E}_{\varepsilon }} \frac{J(x,y)}{2} \bigg (\int _0^1 \mathscr{L}^d|_{ Q_{\varepsilon }^{(1-s)x_{\mathsf{z}} + s y_{\mathsf{z}}}}{\, \mathrm{d}} s\bigg ) (y_{\mathsf{z}} - x_{\mathsf{z}}) . \end{align*}

With a slight abuse of notation, given ${\boldsymbol{m}}: (0,1) \to \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ we also write $\iota _{\varepsilon }{\boldsymbol{m}} \in \mathcal{M}_+\bigl ((0,1) \times{\mathbb{T}^d} \bigr )$ for the continuous space-time measure with time disintegration given by $t \mapsto \iota _{\varepsilon } m_t$ . Moreover, for a given sequence of nonnegative discrete measures $m^{\varepsilon } \in \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ , we write

\begin{align*} m_{\varepsilon } \to \mu \in \mathcal{M}_+({\mathbb{T}^d}) \quad \text{weakly} \qquad \text{iff} \qquad \iota _{\varepsilon } m^{\varepsilon } \to \mu \quad \text{weakly in } \mathcal{M}_+({\mathbb{T}^d}) . \end{align*}

Similarly, for ${\boldsymbol{m}}^{\varepsilon }: (0,1) \to \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ we write ${\boldsymbol{m}}^{\varepsilon } \to \boldsymbol{\mu } \in \mathcal{M}_+\bigl ((0,1) \times{\mathbb{T}^d} \bigr )$ with an analogous meaning. Similar notation is used for (Borel, possibly discontinuous) curves of fluxes $\boldsymbol{J} : (0,1) \to \mathbb{R}^{\mathcal{E}_{\varepsilon }}_a$ and convergent sequences of (curves of) fluxes.

Remark 3.8 (Preservation of the continuity equation). The definition of this embedding for masses and fluxes ensures that solutions to the discrete continuity equation are mapped to solutions of $\mathsf{CE}$ , cfr. [[Reference Gladbach, Kopfer, Maas and Portinale23], Lemma 4.9].

We are ready to state our main result.

Theorem 3.9 (Main result). Let $(\mathcal{X},\mathcal{E},F)$ be as described in Section 2.2 and Assumption 2.8 . Assume that $F$ is either of flow-based type (Assumption 3.2) or with linear growth (Assumption 3.1). Then, in either case, the functionals ${\mathcal{MA}}_{\varepsilon }$ $\Gamma$ -convergence to ${\mathbb{MA}}_{\mathrm{hom}}$ as $\varepsilon \rightarrow 0$ with respect to the weak topology of $\mathcal{M}_+({\mathbb{T}^d}) \times \mathcal{M}_+({\mathbb{T}^d})$ . More precisely, we have

  1. (1) Liminf inequality : For any sequences $m_0^{\varepsilon },m_1^{\varepsilon } \in \mathcal{M}_+(\mathcal{X}_{\varepsilon })$ such that $m_i^{\varepsilon } \to \mu _i$ weakly in $\mathcal{M}_+({\mathbb{T}^d})$ for $i=\,0,1$ , we have that

    \begin{align*} \liminf _{\varepsilon \rightarrow 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon }) \geq{\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1) . \end{align*}
  2. (2) Limsup inequality : For any $\mu _0,\mu _1 \in \mathcal{M}_+({\mathbb{T}}^d)$ , there exist sequences $m_0^{\varepsilon }, m_1^{\varepsilon } \in \mathcal{M}_+(\mathcal{X}_{\varepsilon })$ such that $m_i^{\varepsilon } \to \mu _i$ weakly in $\mathcal{M}_+({\mathbb{T}^d})$ for $i=0,1$ , and

    \begin{align*} \limsup _{\varepsilon \rightarrow 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon }, m_1^{\varepsilon }) \leq{\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1) . \end{align*}

Remark 3.10 (Convergence of the actions and superlinear regime). The $\Gamma$ -convergence of the energies $\mathcal{A}_{\varepsilon }$ towards ${\mathbb{A}}_{\mathrm{hom}}$ under Assumption 2.8 is the main result of [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.1]. Related to it, similarly as discussed in Remark 2.6 , the superlinear case [[Reference Gladbach, Kopfer, Maas and Portinale23], Assumption 5.5], not included in the statement, has already been proved in [Reference Gladbach, Kopfer, Maas and Portinale23], and it follows directly from the aforementioned convergence $\mathcal{A}_{\varepsilon } \xrightarrow{\Gamma }{\mathbb{A}}_{\mathrm{hom}}$ and a strong compactness result which holds in such a framework, see in particular [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorems 5.9 & 5.10]. Without the superlinear growth assumption, the proof is much more involved and requires extra work and new ideas, which are the main contribution of this paper.

Remark 3.11 (Compactness under linear growth from below). Just assuming Assumption 2.8 , the following compactness result for sequences of bounded action was proved in [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.4]: if ${\boldsymbol{m}}^{\varepsilon }:(0,1) \to \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ is such that

\begin{align*} \sup _{\varepsilon \gt 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }) \lt \infty \quad \text{ and }\quad \sup _{\varepsilon \gt 0}{\boldsymbol{m}}^{\varepsilon } \big ( (0,1) \times \mathcal{X}_{\varepsilon } \big ) \lt \infty, \end{align*}

then there exists a curve $\boldsymbol{\mu }= \mu _t({\, \mathrm{d}} x){\, \mathrm{d}} t \in \mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ such that, up to a (non-relabelled) subsequence, we have

\begin{align*}{\boldsymbol{m}}^{\varepsilon } \to \boldsymbol{\mu } \quad \text{weakly in } \mathcal{M}_+ \big ( (0,1) \times{\mathbb{T}}^d \big ) \quad \text{ and }\quad m_t^{\varepsilon } \to \mu _t \quad \text{weakly in } \mathcal{M}_+({\mathbb{T}^d}), \end{align*}

for a.e. $t \in (0,1)$ . This is going to be an important tool in the proof of our main result.

3.1. Proof of the limsup inequality

In this section, we prove the limsup inequality in Theorem3.9. This proof does not require Assumption 3.1 or Assumption 3.2, but rather a weaker hypothesis, which is satisfied under either of the two assumptions.

Proposition 3.12 ( $\Gamma$ -limsup). Let $\mu _0, \mu _1$ be nonnegative measures on ${\mathbb{T}}^d$ . Assume that there exists a $\mathbb{Z}^d$ -periodic and divergence-free vector field $\bar J \in \mathbb{R}_a^{\mathcal{X}}$ such that

(3.3) \begin{align} F(m,\bar J) \le C \Biggl (1 + \sum _{\substack{x \in \mathcal X \\ \lvert x\rvert _{\infty } \le R}} m(x)\Biggr ), \qquad m \in \mathbb{R}^{\mathcal{X}}_+, \end{align}

for some finite constants $C$ and $R$ . Then there exist two sequences $(m_0^{\varepsilon })_{\varepsilon \gt 0}$ and $(m_1^{\varepsilon })_{\varepsilon \gt 0}$ in $\mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ such that $m_i^{\varepsilon } \to \mu _i$ weakly in $\mathcal{M}_+({\mathbb{T}}^d)$ for $i=0,1$ and

(3.4) \begin{equation} \limsup _{\varepsilon \to 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon }) \le{\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1) . \end{equation}

Proof. We may and will assume that ${\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1) \lt \infty$ . We also claim that it suffices to prove the statement with ${\mathbb{MA}}(\mu _0,\mu _1)+1/k$ in place of the right-hand side of (3.4) for every $k \in \mathbb{N}_1$ . Indeed, assume that we know of the existence of sequences $(m_i^{\varepsilon, k})_{\varepsilon }$ such that $m_i^{\varepsilon, k} \to \mu _i$ and

\begin{align*} \limsup _{\varepsilon \to 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon, k},m_1^{\varepsilon, k}) \le{\mathbb{MA}}(\mu _0,\mu _1)+1/k, \end{align*}

for every $k \in \mathbb{N}_1$ . Since ${\mathbb{T}}^d$ is compact, the weak convergence is equivalent to convergence in the Kantorovich–Rubinstein norm. Hence, for every $k$ , we can find $\varepsilon _k$ such that when $\varepsilon \le \varepsilon _k$ ,

\begin{align*}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon, k},m_1^{\varepsilon, k}) \le{\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1)+2/k \quad \text{ and }\quad \max _{i=0,1} \|{\iota _{\varepsilon } m_i^{\varepsilon, k}-\mu _i}\|_{\mathrm{KR}} \le 1/k . \end{align*}

We can also assume that $\varepsilon _{k+1} \le \frac{\varepsilon _k}{2}$ , for every $k$ . It now suffices to set

\begin{align*} k_{\varepsilon } \,:\!=\, \max \{ k \in \mathbb{N}_1 \, : \, \varepsilon _k \ge \varepsilon \} \quad \text{ and }\quad m_i^{\varepsilon } \,:\!=\, m_i^{\varepsilon, k_{\varepsilon }}, \end{align*}

for every $\varepsilon$ and $i=0,1$ to get

\begin{align*} \limsup _{\varepsilon \to 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon }) \le{\mathbb{MA}}_{\mathrm{hom}} (\mu _0,\mu _1)+ \limsup _{\varepsilon \to 0} \frac{2}{k_{\varepsilon }} \end{align*}

as well as

\begin{align*} \limsup _{\varepsilon \to 0} \max _{i=0,1} \|{\iota _{\varepsilon } m_i^{\varepsilon } -\mu _i}\|_{\mathrm{KR}} \le \limsup _{\varepsilon \to 0} \frac{1}{k_{\varepsilon }} . \end{align*}

The claim is proved, since $k_{\varepsilon } \to _{\varepsilon } \infty$ , as can be readily verified.

Thus, let us now choose $k$ and keep it fixed. By definition of ${\mathbb{MA}}_{\mathrm{hom}}$ , there exists $\boldsymbol{\mu } = \mu _t({\, \mathrm{d}} x){\, \mathrm{d}} t \in \mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ with $\boldsymbol{\mu }_{t=0} = \mu _0, \boldsymbol{\mu }_{t=1} = \mu _1$ and such that

\begin{align*}{\mathbb{A}}_{\mathrm{hom}}(\boldsymbol{\mu }) \le{\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1)+1/k . \end{align*}

Recall from Remark 3.10 that $\mathcal{A}_{\varepsilon } \xrightarrow{\Gamma }{\mathbb{A}}_{\mathrm{hom}}$ ; in particular, there exists a recovery sequence $({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) \in \mathcal{CE}_{\varepsilon }$ such that ${\boldsymbol{m}}^{\varepsilon } \to \boldsymbol{\mu }$ weakly and

\begin{align*} \limsup _{\varepsilon \to 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) \le{\mathbb{A}}_{\mathrm{hom}}(\boldsymbol{\mu }) . \end{align*}

We shall prove that $\|{\iota _{\varepsilon } m_t^{\varepsilon } - \mu _t}\|_{\mathrm{KR}({\mathbb{T}}^d)} \to 0$ in ( $\mathscr{L}^1$ -)measure or, equivalently, that

(3.5) \begin{equation} \lim _{\varepsilon \to 0} \int _0^1 \min \left \{\|{\iota _{\varepsilon } m_t^{\varepsilon } - \mu _t}\|_{\mathrm{KR}({\mathbb{T}}^d)}, 1 \right \}{\, \mathrm{d}} t = 0 . \end{equation}

In order to do this, assume by contradiction that there exists a subsequence such that

\begin{align*} \int _0^1 \min \left \{\|{\iota _{\varepsilon _n} m_t^{\varepsilon _n} - \mu _t}\|_{\mathrm{KR}({\mathbb{T}}^d)}, 1 \right \}{\, \mathrm{d}} t \gt \delta, \qquad n \in \mathbb{N}, \end{align*}

for some $\delta \gt 0$ . Up to possibly extracting a further subsequence, it can be easily checked that the hypotheses of [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.4] are satisfied (cfr. Remark 3.11); hence, there exists a further (not relabelled) subsequence such that, for almost every $t \in (0,1)$ , $m_t^{\varepsilon _n} \to \mu _t$ weakly and thus $\|{\iota _{\varepsilon _n}m_t^{\varepsilon _n} - \mu _t}\|_{\mathrm{KR}({\mathbb{T}}^d)} \to 0$ . The dominated convergence theorem yields an absurd.

From (3.5) we deduce that for every $T \in (0,1/2)$ there exists a sequence of times $(a_{\varepsilon }^T)_{\varepsilon } \subseteq (0,T)$ such that

\begin{align*} \lim _{\varepsilon \to 0} \|{\iota _{\varepsilon } m_{a_{\varepsilon }^T}^{\varepsilon } - \mu _{a_{\varepsilon }^T}}\|_{\mathrm{KR}({\mathbb{T}}^d)} = 0 . \end{align*}

With a diagonal argument, we find a sequence $(a_{\varepsilon })_{\varepsilon } \subseteq (0,1/2)$ such that

\begin{align*} \lim _{\varepsilon \to 0} a_{\varepsilon } = 0 \quad \text{ and }\quad \lim _{\varepsilon \to 0} \|{\iota _{\varepsilon } m_{a_{\varepsilon }}^{\varepsilon } - \mu _{a_{\varepsilon }}}\|_{\mathrm{KR}({\mathbb{T}}^d)} = \lim _{\varepsilon \to 0} \|{\iota _{\varepsilon } m_{a_{\varepsilon }}^{\varepsilon } - \mu _0}\|_{\mathrm{KR}({\mathbb{T}}^d)} = 0 . \end{align*}

Similarly, we can find another sequence $(b_{\varepsilon })_{\varepsilon } \subseteq (1/2,1)$ such that

\begin{align*} \lim _{\varepsilon \to 0} b_{\varepsilon } = 1 \quad \text{and} \quad \lim _{\varepsilon \to 0} \|{\iota _{\varepsilon } m_{b_{\varepsilon }}^{\varepsilon } - \mu _1}\|_{\mathrm{KR}({\mathbb{T}}^d)}=0 . \end{align*}

We claim the sought recovery sequences is provided by $m_0^{\varepsilon } \,:\!=\, m^{\varepsilon }_{a_{\varepsilon }}$ and $m_1^{\varepsilon } \,:\!=\, m_{b_{\varepsilon }}^{\varepsilon }$ . In order to show this, let us define $\widehat J^{\varepsilon } : \mathcal{E}_{\varepsilon } \to \mathbb{R}$ via the formulaFootnote 1 (recall the assumption (3.3))

\begin{align*} \frac{\tau _{\varepsilon }^z \widehat J^{\varepsilon }}{\varepsilon ^{d-1}} \,:\!=\, \bar J, \qquad z \in \mathbb{Z}^d_{\varepsilon }, \end{align*}

so that $\widehat J^{\varepsilon }$ is divergence-free. Now define

\begin{align*} \widetilde m^{\varepsilon }_t \,:\!=\, \begin{cases} m^{\varepsilon }_{a_{\varepsilon }} &\text{if } t \in [0,a_{\varepsilon }) \\ m^{\varepsilon }_t &\text{if } t \in [a_{\varepsilon },b_{\varepsilon }] \\ m^{\varepsilon }_{b_{\varepsilon }} &\text{if } t \in (b_{\varepsilon },1] \end{cases} \quad \text{ and }\quad \widetilde J^{\varepsilon }_t \,:\!=\, \begin{cases} \widehat J^{\varepsilon } &\text{if } t \in [0,a_{\varepsilon }) \\ J^{\varepsilon }_t &\text{if } t \in [a_{\varepsilon },b_{\varepsilon }] \\ \widehat J^{\varepsilon } &\text{if } t \in (b_{\varepsilon },1] \end{cases} . \end{align*}

It is readily verified that $(\widetilde{\boldsymbol{m}}^{\varepsilon }, \widetilde{\boldsymbol{J}}^{\varepsilon })$ solves the continuity equation for every $\varepsilon$ . Therefore,

(3.6) \begin{align} {\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon }) &\leq \mathcal{A}_{\varepsilon }(\widetilde{\boldsymbol{m}}^{\varepsilon },\widetilde{\boldsymbol{J}}^{\varepsilon }) = \int _0^1 \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \varepsilon ^d F\left ( \frac{\tau ^z_{\varepsilon } \widetilde m_t^{\varepsilon }}{\varepsilon ^d}, \frac{\tau ^z_{\varepsilon } \widetilde J^{\varepsilon }_t}{\varepsilon ^{d-1}} \right ) \,{\, \mathrm{d}} t\\ &= \int _0^{a_{\varepsilon }} \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \varepsilon ^d F\left ( \frac{\tau ^z_{\varepsilon } m_{a_{\varepsilon }}^{\varepsilon }}{\varepsilon ^d}, \bar J \right ) \,{\, \mathrm{d}} t + \int _{a_{\varepsilon }}^{b_{\varepsilon }} \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \varepsilon ^d F\left ( \frac{\tau ^z_{\varepsilon } m_t^{\varepsilon }}{\varepsilon ^d}, \frac{\tau ^z_{\varepsilon } J^{\varepsilon }_t}{\varepsilon ^{d-1}} \right ) \,{\, \mathrm{d}} t\nonumber \\ &\quad + \int _{b_{\varepsilon }}^1 \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \varepsilon ^d F\left ( \frac{\tau ^z_{\varepsilon } m_{b_{\varepsilon }}^{\varepsilon }}{\varepsilon ^d}, \bar J \right ) \,{\, \mathrm{d}} t\nonumber \\ & =: I_1 + I_2 + I_3.\nonumber \end{align}

The first and last integral can be estimated using the assumption (3.3). Indeed,

\begin{align*} I_1+I_3 &\le C \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \Biggl ((a_{\varepsilon }+1-b_{\varepsilon }) \varepsilon ^d + \sum _{\substack{x \in \mathcal X \\ |x_{\mathsf{z}}|_{\infty } \le R}} \left ( a_{\varepsilon } (\tau ^z_{\varepsilon } m_{a_{\varepsilon }}^{\varepsilon })(x) + (1-b_{\varepsilon }) (\tau ^z_{\varepsilon } m_{b_{\varepsilon }}^{\varepsilon })(x) \right )\Biggr ) \\ &\le C \left ( (a_{\varepsilon }+1-b_{\varepsilon }) + (2R+1)^d \sum _{x \in \mathcal{X}_{\varepsilon }} \left (a_{\varepsilon } m_{a_{\varepsilon }}^{\varepsilon }(x) + (1-b_{\varepsilon }) m_{b_{\varepsilon }}^{\varepsilon }(x)\right ) \right ) \\ &= C \left ( (a_{\varepsilon }+1-b_{\varepsilon }) + (2R+1)^d \left (a_{\varepsilon } \iota _{\varepsilon } m_{a_{\varepsilon }}^{\varepsilon }({\mathbb{T}}^d) + (1-b_{\varepsilon })\iota _{\varepsilon } m_{b_{\varepsilon }}^{\varepsilon }({\mathbb{T}}^d)\right ) \right ), \end{align*}

and in the limit we find

(3.7) \begin{align} \limsup _{\varepsilon \to 0} I_1+I_3 \le C\left (0 + (2R+1)^d(0\cdot \mu _0({\mathbb{T}}^d) + 0 \cdot \mu _1({\mathbb{T}}^d))\right ) = 0 . \end{align}

As for the second integral, thanks to Assumption 2.8(c) we have that

(3.8) \begin{align} I_2 - \mathcal A_{\varepsilon }({\boldsymbol{m}}^{\varepsilon },\boldsymbol{J}^{\varepsilon }) &= - \int _{(0,a_{\varepsilon })\cup (b_{\varepsilon },1)} \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \varepsilon ^d F\left ( \frac{\tau ^z_{\varepsilon } m_t^{\varepsilon }}{\varepsilon ^d}, \frac{\tau ^z_{\varepsilon } J^{\varepsilon }_t}{\varepsilon ^{d-1}} \right ) \,{\, \mathrm{d}} t\\ &\le C' \left ((a_{\varepsilon }+1-b_{\varepsilon })+(2R_{\mathrm{max}}+1)^d \iota _{\varepsilon }{\boldsymbol{m}}^{\varepsilon }\left (\left ((0,a_{\varepsilon })\cup (b_{\varepsilon },1)\right ) \times{\mathbb{T}}^d\right ) \right ) .\nonumber \end{align}

Since $(\iota _{\varepsilon }{\boldsymbol{m}}^{\varepsilon })_{\varepsilon }$ converges weakly, for every $a, b \in (0,1)$ , we have that

\begin{align*} \limsup _{\varepsilon \to 0} \iota _{\varepsilon }{\boldsymbol{m}}^{\varepsilon } \left ( \left ( (0,a_{\varepsilon })\cup (b_{\varepsilon },1) \right ) \times{\mathbb{T}}^d \right ) &\leq \limsup _{\varepsilon \to 0} \iota _{\varepsilon }{\boldsymbol{m}}^{\varepsilon } \left ( \big ( (0,a] \cup [b,1) \big ) \times{\mathbb{T}^d} \right ) \\ &\leq \boldsymbol{\mu } \left ( \big ( (0,a] \cup [b,1) \big ) \times{\mathbb{T}^d} \right ) . \end{align*}

Using the fact that the previous estimate holds for every $a,b \in (0,1)$ , we obtain that

\begin{align*} \limsup _{\varepsilon \to 0} \iota _{\varepsilon }{\boldsymbol{m}}^{\varepsilon } \left ( \left ( (0,a_{\varepsilon })\cup (b_{\varepsilon },1) \right ) \times{\mathbb{T}}^d \right ) = 0 . \end{align*}

This, together with the estimate obtained in (3.8), gives us the inequality

(3.9) \begin{align} \limsup _{\varepsilon \to 0} I_2 \le \limsup _{\varepsilon \to 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) . \end{align}

In conclusion, from (3.6), (3.7), and (3.9), we find

\begin{align*} \limsup _{\varepsilon \to 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon }) \le \limsup _{\varepsilon \to 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) \le{\mathbb{A}}(\boldsymbol{\mu }) \le{\mathbb{MA}}(\mu _0,\mu _1) + 1/k, \end{align*}

which is sought upper bound.

3.2. Proof of the liminf inequality

In this section, we provide the proof of the liminf inequality in Theorem3.9. Let $m_0^{\varepsilon }$ , $m_1^{\varepsilon }$ be sequences of measures weakly converging to $\mu _0$ , $\mu _1$ , respectively. We want to show that

(3.10) \begin{align} \liminf _{\varepsilon \to 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon }) \geq{\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1) . \end{align}

Without loss of generality, we will assume that the limit inferior in the latter is a true finite limit, and that $m_0^{\varepsilon }(\mathcal{X}_{\varepsilon })=m_1^{\varepsilon }(\mathcal{X}_{\varepsilon })$ for every $\varepsilon \gt 0$ .

We split the proof into two parts: first for $F$ with linear growth and then for $F$ of flow-based type, respectively Assumption 3.1 and Assumption 3.2.

3.2.1. Case 1: F with linear growth

Assume that $F$ satisfies Assumption 3.1. Recall that, as a consequence of Lemma 3.5, $F$ is Lipschitz continuous as well, in the sense of (3.1).

Proof of the liminf inequality (linear growth). With a very similar argument as the one provided by Remark 2.7 in the continuous setting, we can with no loss of generality assume that $F$ is nonnegative. Let $({\boldsymbol{m}}^{\varepsilon },\boldsymbol{J}^{\varepsilon }) \in \mathcal{CE}_{\varepsilon }$ be approximate optimal solutions associated to ${\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon })$ , i.e. such that

(3.11) \begin{align} \lim _{\varepsilon \to 0} \bigl ( \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon },\boldsymbol{J}^{\varepsilon }) -{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon }) \bigr ) = 0 . \end{align}

As usual, we write ${\, \mathrm{d}}{\boldsymbol{m}}^{\varepsilon } (t,x) = m_t^{\varepsilon }({\, \mathrm{d}} x){\, \mathrm{d}} t$ for some measurable curve $t \mapsto m_t^{\varepsilon } \in \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ of constant, finite mass. By compactness (Remark 3.11), we know that up to a further non-relabelled subsequence, ${\boldsymbol{m}}^{\varepsilon } \to \boldsymbol{\mu }$ weakly in $\mathcal{M}_+\big ((0,1)\times{\mathbb{T}^d} \big )$ with $\boldsymbol{\mu } \in \mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ . Due to the lack of continuity of the trace operators in $\mathrm{BV}$ , a priori we cannot conclude that $\boldsymbol{\mu }_{t=0} = \mu _0$ and $\boldsymbol{\mu }_{t=1} = \mu _1$ . In other words, there might be a “jump” in the limit as $\varepsilon \to 0$ at the boundary of $(0,1)$ . In order to take care of this problem, we rescale our measures ${\boldsymbol{m}}^{\varepsilon }$ in time, so as to be able to “see” the jump in the interior of $(0,1)$ .

To this purpose, for $\delta \in (0,1/2)$ , we define $\mathcal{I}_\delta \,:\!=\, (\delta, 1-\delta )$ and ${\boldsymbol{m}}^{\varepsilon, \delta } \in \mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ as

(3.12) \begin{align} m_t^{\varepsilon, \delta } \,:\!=\, \begin{cases} m_0^{\varepsilon } &\text{if } t \in (0,\delta ] \\ m_{\frac{t-\delta }{1-2\delta }}^{\varepsilon } &\text{if } t \in \mathcal{I}_\delta \\ m_1^{\varepsilon } &\text{if } t \in [1-\delta, 1) \end{cases}, \quad{\, \mathrm{d}}{\boldsymbol{m}}^{\varepsilon, \delta }(t,x) \,:\!=\, m_t^{\varepsilon, \delta }({\, \mathrm{d}} x){\, \mathrm{d}} t . \end{align}

By construction, the convergence of the boundary data, and the fact that, by assumption, ${\boldsymbol{m}}^{\varepsilon } \to \boldsymbol{\mu }$ weakly, it is straightforward to see that ${\boldsymbol{m}}^{\varepsilon, \delta } \to \widehat{\boldsymbol{\mu }}^\delta$ weakly, where

(3.13) \begin{align} \widehat \mu _t^\delta \,:\!=\, \begin{cases} \mu _0 &\text{if } t \in (0,\delta ] \\ \mu _{\frac{t-\delta }{1-2\delta }} &\text{if } t \in \mathcal{I}_\delta \\ \mu _1 &\text{if } t \in [1-\delta, 1) \end{cases}, \quad{\, \mathrm{d}} \widehat{\boldsymbol{\mu }}^\delta (t,x) \,:\!=\, \widehat \mu _t^\delta ({\, \mathrm{d}} x){\, \mathrm{d}} t . \end{align}

Note that the rescaled curve $t \mapsto \widehat \mu _t^\delta$ might have discontinuities at $t=\delta$ and $t= 1-\delta$ , which correspond to the possible jumps in the limit as $\varepsilon \to 0$ for ${\boldsymbol{m}}^{\varepsilon }$ at $\{0,1\}$ . Nevertheless, $\widehat{\boldsymbol{\mu }}^\delta$ is a competitor for ${\mathbb{MA}}(\mu _0,\mu _1)$ , which, by the $\Gamma$ -convergence of $\mathcal{A}_{\varepsilon }$ towards ${\mathbb{A}}_{\mathrm{hom}}$ (Remark 3.10), ensures that

(3.14) \begin{align} \liminf _{\varepsilon \to 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon, \delta }) \geq{\mathbb{A}}_{\mathrm{hom}}(\widehat{\boldsymbol{\mu }}^\delta ) \geq{\mathbb{MA}}_{\mathrm{hom}}(\mu _0,\mu _1) . \end{align}

We are left with estimating from above the left-hand side of the latter displayed equation. To do so, we seek a suitable curve of discrete vector fields $\boldsymbol{J}^{\varepsilon, \delta }$ with $({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta }) \in \mathcal{CE}_{\varepsilon }$ and having an action $\mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon, \delta },\boldsymbol{J}^{\varepsilon, \delta })$ comparable with $\mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon },\boldsymbol{J}^{\varepsilon })$ for small $\delta \gt 0$ . We set

\begin{align*} J_t^{\varepsilon, \delta } \,:\!=\, \begin{cases} 0 &\text{if } t \in (0,\delta ] \\ \frac{1}{1-2\delta } J_{\frac{t-\delta }{1-2\delta }}^{\varepsilon } &\text{if } t \in \mathcal{I}_\delta \\ 0 &\text{if } t \in [1-\delta, 1) \end{cases}, \quad{\, \mathrm{d}} \boldsymbol{J}^{\varepsilon, \delta }(t,x) \,:\!=\, J_t^{\varepsilon, \delta }({\, \mathrm{d}} x){\, \mathrm{d}} t . \end{align*}

We claim that $ ({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta } ) \in \mathcal{CE}_{\varepsilon }$ and

(3.15) \begin{align} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta }) \leq \bigl (1+C(F)\delta \bigr ) \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) + C(F) \delta \Big ( 1 + \iota _{\varepsilon }{\boldsymbol{m}}^{\varepsilon } ((0,1) \times{\mathbb{T}^d}) \Big ), \end{align}

where $C(F) \in \mathbb{R}_+$ only depends on $F$ (specifically on the constants in Assumption 2.8 and Assumption 3.1). This would suffice to conclude the proof of the sought liminf inequality. Indeed, from (3.11) and (3.15), we infer

\begin{align*} \liminf _{\varepsilon \to 0} &{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon }, m_1^{\varepsilon }) = \liminf _{\varepsilon \to 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) \\ &\geq \frac{1}{1+C(F)\delta } \liminf _{\varepsilon \to 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta }) - \frac{C(F) \delta }{1+C(F)\delta } \Big ( 1 + \iota _{\varepsilon }{\boldsymbol{m}}^{\varepsilon } ((0,1) \times{\mathbb{T}^d}) \Big ) \end{align*}

which, combined with (3.14), yields

\begin{equation*} \liminf _{\varepsilon \to 0} {\mathcal {MA}}_{\varepsilon }(m_0^{\varepsilon }, m_1^{\varepsilon }) \geq \frac {{\mathbb {MA}}_{\mathrm {hom}}(\mu _0,\mu _1)}{1+C(F)\delta } - \frac {C(F) \delta }{1+C(F)\delta } \Big ( 1 + \mu _0({\mathbb {T}^d}) \Big ) \end{equation*}

for any $\delta \in (0,1/2)$ . We conclude by letting $\delta \to 0$ .

We are left with the proof of $ ({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta } ) \in \mathcal{CE}_{\varepsilon }$ and of the claim (3.15).

Proof of $ ({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta } ) \in \mathcal{CE}_{\varepsilon }$ . Let us fix $x \in \mathcal{X}_{\varepsilon }$ and $\varphi \in C^1_c \bigl ((0,1)\bigr )$ . Set $\widetilde \varphi : = \varphi \circ r_\delta$ , with $r_\delta (s) \,:\!=\,(1-2\delta )s+\delta$ . We have

(3.16) \begin{align} \int _0^1 \partial _t \varphi m_t^{\varepsilon, \delta }(x){\, \mathrm{d}} t &= \int _0^\delta \partial _t \varphi \, m_0^{\varepsilon }(x){\, \mathrm{d}} t + \int _{1-\delta }^1 \partial _t \varphi \, m_1^{\varepsilon }(x){\, \mathrm{d}} t + \int _{\mathcal{I}_\delta } \partial _t \varphi \, m_{r_\delta ^{-1}(t)}^{\varepsilon }(x){\, \mathrm{d}} t \\ &= \varphi (\delta ) \, m_0^{\varepsilon }(x) - \varphi (1-\delta ) \, m_1^{\varepsilon }(x) + (1-2\delta )\int _0^1 (\partial _t \varphi ) \circ r_\delta \, m_s^{\varepsilon }(x){\, \mathrm{d}} s \nonumber \\ &= \widetilde \varphi (0) \, m_0^{\varepsilon }(x) - \widetilde \varphi (1) \, m_1^{\varepsilon }(x) + \int _0^1 \partial _s \widetilde \varphi \, m_s^{\varepsilon }(x){\, \mathrm{d}} s \nonumber \\ &= \int _0^1 \widetilde \varphi \, \sum _{y \sim x} J_s^{\varepsilon }(x,y){\, \mathrm{d}} s = \frac{1}{1-2\delta } \int _{\mathcal{I}_\delta } \varphi \, \sum _{y \sim x} J^{\varepsilon }_{r_\delta ^{-1}(t)}(x,y){\, \mathrm{d}} t \nonumber \\ &= \int _0^1 \varphi \sum _{y \sim x} J^{\varepsilon, \delta }_t(x,y){\, \mathrm{d}} t, \nonumber \end{align}

where, in the fourth equality, we used that $({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) \in \mathcal{CE}_{\varepsilon }$ .

Proof of the action estimate. Define $r_\delta (s) \,:\!=\,(1-2\delta )s+\delta$ . Note that, by construction, for $(t,(x,y)) \in \mathcal{I}_\delta \times \mathcal{E}_{\varepsilon }$ ,

\begin{align*} m_t^{\varepsilon, \delta }(x) = m_{r_\delta ^{-1}(t)}^{\varepsilon }(x), \qquad J_t^{\varepsilon, \delta }(x,y) = \frac 1{1-2\delta } J_{ r_\delta ^{-1}(t)}^{\varepsilon }(x,y) . \end{align*}

On the other hand, for $(t,(x,y)) \in \big ( (0,\delta ] \cup [1-\delta, 1) \big ) \times \mathcal{E}_{\varepsilon }$ , we have that

\begin{align*} m_t^{\varepsilon, \delta }(x) = \begin{cases} m_0^{\varepsilon }(x) &\text{if } t \in (0,\delta ] \\ m_1^{\varepsilon }(x) &\text{if } t \in [1-\delta, 1) \end{cases} \quad \text{ and }\quad J_t^{\varepsilon, \delta }(x,y) = 0 . \end{align*}

It follows that the action of $({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta })$ is given by

(3.17) \begin{align} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta }) = \int _0^1 \mathcal{F}_{\varepsilon }(m_t^{\varepsilon, \delta }, J_t^{\varepsilon, \delta }){\, \mathrm{d}} t = \mathcal{A}_{\varepsilon }^{\mathcal{I}_\delta } ({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta }) + \delta \sum _{i=0,1} \mathcal{F}_{\varepsilon }(m_i^{\varepsilon },0), \end{align}

where we used the notation

\begin{align*} \mathcal{A}_{\varepsilon }^{\mathcal{I}_\delta }({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta }) &\,:\!=\, \int _{\mathcal{I}_\delta } \mathcal{F}_{\varepsilon }(m_t^{\varepsilon, \delta }, J_t^{\varepsilon, \delta }){\, \mathrm{d}} t = (1-2\delta ) \int _0^1 \mathcal{F}_{\varepsilon }\Big ( m_t^{\varepsilon }, \frac 1{1-2\delta }J_t^{\varepsilon } \Big ){\, \mathrm{d}} t . \end{align*}

Using Assumption 3.1, we see that, for $i=0,1$ ,

\begin{align*} \mathcal{F}_{\varepsilon }(m_i^{\varepsilon },0) \leq C (m_i^{\varepsilon }(\mathcal{X}_{\varepsilon }) +1) = C \bigl ( \iota _{\varepsilon }{\boldsymbol{m}}^{\varepsilon } \bigl ((0,1)\times{\mathbb{T}^d}\bigl ) + 1\bigr ) \end{align*}

and, by the Lipschitz continuity exhibited in Lemma 3.5, we also infer that

\begin{align*} \mathcal{A}_{\varepsilon }^{\mathcal{I}_\delta }({\boldsymbol{m}}^{\varepsilon, \delta }, \boldsymbol{J}^{\varepsilon, \delta }) &\leq (1-2\delta ) \left (\mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) + C \Big (\frac 1{1-2\delta }-1 \Big ) \sum _{z \in \mathbb{Z}_{\varepsilon }^d} \varepsilon ^d \sum _{ \substack{ (x,y) \in \mathcal{E} \\ |x_{\mathsf{z}}|_{\infty } \leq R } } \int _0^1 \frac{\lvert \tau _{\varepsilon }^z J^{\varepsilon }_t(x,y)\rvert }{\varepsilon ^{d-1}}{\, \mathrm{d}} t \right ) \\ &\le (1-2\delta ) \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) + 2 \delta C (2R+1)^d \sum _{z \in \mathbb{Z}_{\varepsilon }^d} \varepsilon ^d \sum _{ (x,y) \in \mathcal{E}^Q } \int _0^1 \frac{\lvert \tau _{\varepsilon }^z J^{\varepsilon }_t(x,y)\rvert }{\varepsilon ^{d-1}}{\, \mathrm{d}} t . \end{align*}

Since we assumed $F$ to be nonnegative, we can estimate

\begin{equation*} (1-2\delta ) \mathcal {A}_{\varepsilon }({\boldsymbol {m}}^{\varepsilon }, \boldsymbol {J}^{\varepsilon }) \le \mathcal {A}_{\varepsilon }({\boldsymbol {m}}^{\varepsilon }, \boldsymbol {J}^{\varepsilon }) \end{equation*}

and, using (2.8),

\begin{equation*} \sum _{z \in \mathbb {Z}_{\varepsilon }^d} \varepsilon ^d \sum _{ (x,y) \in \mathcal {E}^Q } \int _0^1 \frac {\lvert \tau _{\varepsilon }^z J^{\varepsilon }_t(x,y)\rvert }{\varepsilon ^{d-1}} {\, \mathrm {d}} t \le \frac {1}{c} \mathcal {A}_{\varepsilon }({\boldsymbol {m}}^{\varepsilon }, \boldsymbol {J}^{\varepsilon }) + \frac {C}{c} \left ( 1 + (1+2R_{\mathrm {max}})^d \| {\iota _{\varepsilon } {\boldsymbol {m}}^{\varepsilon }}\|_{\mathrm {TV}} \right ) . \end{equation*}

Combining these estimates with (3.17), we find (3.15).

3.2.2. Case 2: F is flow-based

In this section, we show (3.10) in the case where $F$ (and hence $f_{\mathrm{hom}}$ ) is of flow-based type, i.e. it satisfies Assumption 3.2. We start by observing that, in this particular setting, both the discrete and the continuous formulations of the boundary-value problems admit an equivalent, static formulation.

Let $(\boldsymbol{\mu },\boldsymbol{\nu }) \in \mathsf{CE}$ and consider the Lebesgue decomposition

\begin{align*} \boldsymbol{\mu } = \rho \mathscr{L}^{\,d+1} + \rho ^\perp \boldsymbol{\sigma }, \qquad \boldsymbol{\nu } = j \mathscr{L}^{\,d+1} + j^\perp \boldsymbol{\sigma } . \end{align*}

We know that every solution to the continuity equation can be disintegrated in the form $\boldsymbol{\mu }({\, \mathrm{d}} t,{\, \mathrm{d}} x) = \mu _t({\, \mathrm{d}} x){\, \mathrm{d}} t$ for some measurable curve $t \mapsto \mu _t \in \mathcal{M}_+({\mathbb{T}^d})$ of constant, finite mass. If $f$ is a function as in Assumption 2.2 that further does not depend on $\rho$ , then Jensen’s inequality yields

(3.18) \begin{align} \int _0^1 \int _{{\mathbb{T}}^d} f(j_t){\, \mathrm{d}} x{\, \mathrm{d}} t \ge \int _{{\mathbb{T}}^d} f\left (\int _0^1 j_t{\, \mathrm{d}} t\right ){\, \mathrm{d}} x . \end{align}

In order to take care of the singular part, consider the disintegration of $\boldsymbol{\sigma }$ with respect to the projection map $\pi :(t,x) \mapsto x$ , in the form

\begin{align*} \boldsymbol{\sigma }({\, \mathrm{d}} t,{\, \mathrm{d}} x) = \sigma ^x({\, \mathrm{d}} t) (\pi _{\#}\boldsymbol{\sigma })({\, \mathrm{d}} x), \end{align*}

for some measurable $x \mapsto \sigma ^x \in \mathscr{P}\big ( (0,1) \big )$ . Due to the convexity of $f^\infty$ , by Jensen’s inequality we also obtain

(3.19) \begin{align} \int _{(0,1) \times{\mathbb{T}^d}} f^\infty (j^\perp ){\, \mathrm{d}} \boldsymbol{\sigma } \ge \int _{{\mathbb{T}}^d} f^\infty \left (\int j^\perp{\, \mathrm{d}} \sigma ^x \right ){\, \mathrm{d}} \pi _{\#} \boldsymbol{\sigma } (x) . \end{align}

Now, we define the new space-time measures

(3.20) \begin{align} \begin{split} \widetilde{\boldsymbol{\mu }} \,:\!=\, \widetilde \mu _t({\, \mathrm{d}} x){\, \mathrm{d}} t \quad \text{and} \quad \widetilde{\boldsymbol{\nu }} \,:\!=\, \widehat j \mathscr{L}^{\,d+1} + \widehat j^\perp{\, \mathrm{d}} t \otimes \pi _{\#}\boldsymbol{\sigma }, \quad \text{where} \\ \widetilde \mu _t\,:\!=\, \mu _0 + t (\mu _1 - \mu _0), \quad \widehat j(x) \,:\!=\, \int _0^1 j_t(x){\, \mathrm{d}} t, \quad \text{and} \ \ \widehat j^\perp (x) \,:\!=\, \int j^\perp{\, \mathrm{d}} \sigma ^x, \end{split} \end{align}

and note that $(\widetilde{\boldsymbol{\mu }}, \widetilde{\boldsymbol{\nu }}) \in \mathsf{CE}$ . By (3.18) and (3.19), we, therefore, have

(3.21) \begin{align} {\mathbb{A}}(\boldsymbol{\mu }, \boldsymbol{\nu }) \geq \int _{{\mathbb{T}}^d} f(\widehat j){\, \mathrm{d}} x + \int _{{\mathbb{T}}^d} f^\infty (\widehat j^\perp ){\, \mathrm{d}} \pi _{\#}\boldsymbol{\sigma }(x) . \end{align}

We need to be careful here: the decomposition of $\widetilde{\boldsymbol{\nu }}$ in (3.20) may not be a Lebesgue decomposition, in the sense that ${\, \mathrm{d}} t \otimes \pi _\# \boldsymbol{\sigma }$ can have a nonzero absolutely continuous part. Let $\widetilde \sigma \in \mathcal{M}_+({\mathbb{T}^d})$ be singular w.r.t. $\mathscr{L}^d$ and such that $\mu _0,\mu _1, \pi _\# \boldsymbol{\sigma } \ll \mathscr{L}^d + \widetilde \sigma$ . We can write the Lebesgue decompositions

\begin{equation*} \widetilde {\boldsymbol {\mu }} = \widetilde \rho \mathscr {L}^{d+1} + \widetilde \rho ^\perp {\, \mathrm {d}} t \otimes \widetilde \sigma, \qquad \widetilde {\boldsymbol {\nu }} = \widetilde j \mathscr {L}^{d+1} + \widetilde j^\perp {\, \mathrm {d}} t \otimes \widetilde \sigma . \end{equation*}

If we write

\begin{equation*} \pi _\# \boldsymbol {\sigma } = \alpha \mathscr {L}^d + \beta \widetilde \sigma \end{equation*}

for some functions $\alpha, \beta :{\mathbb{T}^d} \to \mathbb{R}_+$ , then

\begin{equation*} \widetilde j = \widehat j + \alpha \widehat j^\perp \quad \text {and} \quad \widetilde j^\perp = \beta \widehat j^\perp . \end{equation*}

The inequality (3.21) becomes, recalling that $f^\infty$ is $1$ -homogeneous,

(3.22) \begin{equation} {\mathbb{A}}(\boldsymbol{\mu }, \boldsymbol{\nu }) \geq \int _{{\mathbb{T}^d}} \bigl ( f(\widehat j) + f^\infty (\alpha \widehat j^\perp ) \bigr ){\, \mathrm{d}} x + \int _{{\mathbb{T}^d}} f^\infty (\beta \widehat j^\perp ){\, \mathrm{d}} \widetilde \sigma . \end{equation}

At this point, we need a lemma.

Lemma 3.13. For every $j_1,j_2 \in \mathbb{R}^d$ , we have that $ f(j_1+j_2) \le f(j_1) + f^\infty (j_2)$ .

Proof. Let $g \le f$ be a convex and Lipschitz continuous function. By convexity, for every $\epsilon \in (0,1)$ , we have

\begin{equation*} g(j_1+j_2) = g\left ((1-\epsilon )\frac {j_1}{1-\epsilon } + \epsilon \frac {j_2}{\epsilon }\right ) \le (1-\epsilon ) g\left ( \frac {j_1}{1-\epsilon } \right ) + \epsilon g\left (\frac {j_2}{\epsilon } \right ) . \end{equation*}

Let $j_0 \in \mathsf{D}(f)$ . By the Lipschitz continuity of $g$ ,

\begin{equation*} g(j_1+j_2) \le (1-\epsilon ) \left ( g(j_1) + (\mathrm {Lip} g) \left ( \frac {1}{1-\epsilon } - 1 \right ) \lvert j_1\rvert \right ) + \epsilon g\left ( \frac {j_2}{\epsilon } + j_0 \right ) + \epsilon (\mathrm {Lip} g) \lvert j_0\rvert \end{equation*}

and, since $g \le f$ ,

\begin{equation*} g(j_1 + j_2) \le (1-\epsilon ) f(j_1) + \epsilon f \left ( \frac {j_2}{\epsilon } + j_0 \right )+ \epsilon (\mathrm {Lip} g) \bigl ( \lvert j_0\rvert +\lvert j_1\rvert \bigr ) . \end{equation*}

As we let $\epsilon \to 0$ , we find

\begin{equation*} g(j_1 + j_2) \le f(j_1) + f^\infty (j_2) . \end{equation*}

Since $f$ is convex and lower semicontinuous, we conclude by an approximation argument.

Applying this lemma with $j_1 = \widehat j(x)$ and $j_2 = \alpha \widehat j^\perp (x)$ for every $x \in{\mathbb{T}^d}$ , (3.22) finally becomes

\begin{equation*} {\mathbb {A}}(\boldsymbol {\mu }, \boldsymbol {\nu }) \ge \int _{{\mathbb {T}^d}} f(\widetilde j) {\, \mathrm {d}} x + \int _{{\mathbb {T}^d}} f^\infty (\widetilde j^\perp ) {\, \mathrm {d}} \widetilde \sigma = {\mathbb {A}}(\widetilde {\boldsymbol {\mu }}, \widetilde {\boldsymbol {\nu }}) . \end{equation*}

In other words, we have shown that an optimal curve $\boldsymbol{\mu }$ between two given boundary data is always given by the affine interpolation (and a constant-in-time flux). We conclude that

(3.23) \begin{align} &{\mathbb{MA}}(\mu _0,\mu _1) ={\mathbb{A}}(\boldsymbol{\widetilde \mu })\\ &\qquad = \inf _\nu \left \{ \int _{{\mathbb{T}^d}} f(\widetilde j){\, \mathrm{d}} x + \int _{{\mathbb{T}^d}} f^\infty (\widetilde j^\perp ){\, \mathrm{d}} \widetilde \sigma \, : \, \nu = \widetilde j \mathscr{L}^d + \widetilde j^\perp \widetilde \sigma, \ \mathscr{L}^d \perp \widetilde \sigma \ \ \text{and} \ \ \nabla \cdot \nu = \mu _0 - \mu _1 \right \}\nonumber . \end{align}

We refer to the latter expression as the static formulation of the boundary value problem described by ${\mathbb{MA}}(\mu _0,\mu _1)$ (in the case when $f$ is of flow-based type).

Remark 3.14. Using this equivalence, the lower semicontinuity of $\mathbb{MA}$ directly follows from the semicontinuity of $\mathbb{A}$ given by [[Reference Gladbach, Kopfer, Maas and Portinale23], Lemma 3.14].

Arguing in a similar way (in fact, via an even simpler argument, due to the lack of singularities), we obtain a static formulation of the discrete transport problem in terms of a discrete divergence equation, when $F(m,J)=F(J)$ . Precisely, in this case we obtain

\begin{align*}{\mathcal{MA}}_{\varepsilon }(m_0,m_1) = \inf \left \{ \mathcal{F}_{\varepsilon }(J) \, : \, J \in \mathbb{R}_a^{\mathcal{E}_{\varepsilon }}, \quad \mathsf{div} J = m_0 - m_1 \right \} . \end{align*}

The sought $\Gamma$ -liminf inequality easily follows from such static formulations.

Proof of the liminf inequality (flow-based type). Let $m_0^{\varepsilon }$ , $m_1^{\varepsilon } \in \mathbb{R}_+^{\mathcal{X}_{\varepsilon }}$ be sequences of discrete nonnegative measures that converge weakly (via $\iota _{\varepsilon }$ in the usual sense) to $\mu _0$ , $\mu _1$ , and such that $m_0^{\varepsilon }(\mathcal{X}_{\varepsilon }) = m_1^{\varepsilon }(\mathcal{X}_{\varepsilon })$ for every $\varepsilon \gt 0$ . Let $({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) \in \mathcal{CE}_{\varepsilon }$ be (almost-)optimal solutions associated with ${\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon },m_1^{\varepsilon })$ , namely

(3.24) \begin{align} \liminf _{\varepsilon \to 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon }, m_1^{\varepsilon }) = \liminf _{\varepsilon \to 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }) = \liminf _{\varepsilon \to 0} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) . \end{align}

Consider the discrete equivalent of the measure constructed in (3.20), namely

\begin{align*} \widetilde m_t^{\varepsilon } \,:\!=\, m_0^{\varepsilon } + t (m_1^{\varepsilon } - m_0^{\varepsilon }) \quad \text{ and }\quad \widetilde J^{\varepsilon }_t \equiv \widetilde J^{\varepsilon }\,:\!=\, \int _0^1 J_s^{\varepsilon }{\, \mathrm{d}} s, \end{align*}

which still solves the continuity equation. By applying Jensen’s inequality, the convexity of $F$ ensures that

(3.25) \begin{align} \mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) \geq \mathcal{A}_{\varepsilon }(\widetilde{\boldsymbol{m}}^{\varepsilon }, \widetilde{\boldsymbol{J}}^{\varepsilon }) = \mathcal{F}_{\varepsilon }(\widetilde J^{\varepsilon }) \quad \text{ and }\quad (\widetilde{\boldsymbol{m}}^{\varepsilon }, \widetilde{\boldsymbol{J}}^{\varepsilon }) \in \mathcal{CE}_{\varepsilon } . \end{align}

Thus $\mathcal{A}_{\varepsilon }({\boldsymbol{m}}^{\varepsilon }, \boldsymbol{J}^{\varepsilon }) \ge \mathcal{A}_{\varepsilon }(\widetilde{\boldsymbol{m}}^{\varepsilon })$ . Note that, by construction, $\widetilde{\boldsymbol{m}}^{\varepsilon } \to \widetilde{\boldsymbol{\mu }}$ weakly, where

\begin{align*} \widetilde{\boldsymbol{\mu }} \,:\!=\, \widetilde \mu _t({\, \mathrm{d}} x){\, \mathrm{d}} t \quad \text{with} \quad \widetilde \mu _t\,:\!=\, \mu _0 + t (\mu _1 - \mu _0). \end{align*}

Hence, from (3.24), (3.25), and the $\Gamma$ -convergence of $\mathcal{A}_{\varepsilon }$ to ${\mathbb{A}}_{\mathrm{hom}}$ (cfr. [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.1]), we infer that

\begin{align*} \liminf _{\varepsilon \to 0}{\mathcal{MA}}_{\varepsilon }(m_0^{\varepsilon }, m_1^{\varepsilon }) \geq{\mathbb{A}}_{\mathrm{hom}}(\widetilde{\boldsymbol{\mu }}) \geq{\mathbb{MA}}_{\mathrm{hom}}( \mu _0, \mu _1 ), \end{align*}

which concludes the proof of the liminf inequality.

3.3. About the lower semicontinuity of $\mathbb{MA}$

In view of our main result, whenever $F$ satisfies either Assumption 3.1 or Assumption 3.2, the limit boundary-value problem ${\mathbb{MA}}_{\mathrm{hom}}(\cdot, \cdot )$ is necessarily jointly lower semicontinuous with respect to the weak topology on $\mathcal{M}_+({\mathbb{T}^d}) \times \mathcal{M}_+({\mathbb{T}^d})$ . This indeed follows from the general fact that any $\Gamma$ -limit with respect to a given topology is always lower semicontinuous with respect to that same topology. Using a very similar proof to that of the $\Gamma$ -liminf inequality, we can actually show that, if $f$ is with linear growth or it is of flow-based type, then the associated $\mathbb{MA}$ is always lower semicontinuous (even if, a priori, $f$ is not of the form $f=f_{\mathrm{hom}}$ ), thus providing a positive answer in this framework to the validity of (2.5). In the flow-based setting, this fact has been observed in Remark 3.14.

Proposition 3.15. Assume that $f$ is with linear growth, namely it satisfies one of the two equivalent conditions appearing in Lemma 3.4 , and assume that $(\mu _0^n$ , $\mu _1^n) \to (\mu _0$ , $\mu _1) \in \mathcal{M}_+({\mathbb{T}^d}) \times \mathcal{M}_+({\mathbb{T}^d})$ weakly. Then:

\begin{align*} \liminf _{n \to \infty }{\mathbb{MA}}(\mu _0^n, \mu _1^n) \geq{\mathbb{MA}}(\mu _0,\mu _1) . \end{align*}

The proof goes along the same line of the proof of the $\Gamma$ -liminf inequality for discrete energies $F$ with linear growth. We sketch it here and add details whenever we encounter nontrivial differences between the two proofs.

Proof. Let $(\boldsymbol{\mu }^n,\boldsymbol{\nu }^n) \in \mathsf{CE}$ be (almost-)optimal solutions associated to ${\mathbb{MA}}(\mu _0^n,\mu _1^n)$ , i.e.

(3.26) \begin{align} \liminf _{n \to \infty }{\mathbb{MA}}(\mu _0^n,\mu _1^n) = \liminf _{n \to \infty }{\mathbb{A}}(\boldsymbol{\mu }^n,\boldsymbol{\nu }^n) . \end{align}

With no loss of generality, we can assume $\sup _n{\mathbb{A}}(\boldsymbol{\mu }^n, \boldsymbol{\nu }^n) \lt \infty$ and that the limits inferior are true limits. By the compactness of Remark 2.5, we know that, up to a non-relabelled subsequence, $\boldsymbol{\mu }^n \to \boldsymbol{\mu }$ weakly in $\mathcal{M}_+\bigl ( (0,1) \times{\mathbb{T}^d} \bigr )$ . Moreover, we also have ${\, \mathrm{d}} \boldsymbol{\mu }(t,x) = \mu _t({\, \mathrm{d}} x){\, \mathrm{d}} t \in \mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ for some measurable curve $t \mapsto \mu _t \in \mathcal{M}_+({\mathbb{T}^d})$ of constant, finite mass. Once again, due to the lack of continuity of the trace operators in $\mathrm{BV}$ , we cannot ensure that $\boldsymbol{\mu }_{t=0} = \mu _0$ and $\boldsymbol{\mu }_{t=1} = \mu _1$ . To solve this issue, we rescale our measures $\boldsymbol{\mu }^n$ in time in the same spirit as in (3.12). For a given $\delta \gt 0$ , we define $\mathcal{I}_\delta \,:\!=\, (\delta, 1-\delta )$ and $\boldsymbol{\mu }^{n,\delta } \in \mathrm{BV}_{\mathrm{KR}}\bigl ((0,1); \mathcal{M}_+({\mathbb{T}^d})\bigr )$ as

\begin{align*} \mu _t^{n,\delta } \,:\!=\, \begin{cases} \mu _0^n &\text{if } t \in (0,\delta ] \\ \mu _{\frac{t-\delta }{1-2\delta }}^n &\text{if } t \in \mathcal{I}_\delta \\ \mu _1^n &\text{if } t \in [1-\delta, 1) \end{cases}, \quad{\, \mathrm{d}} \boldsymbol{\mu }^{n,\delta }(t,x) \,:\!=\, \mu _t^{n,\delta }({\, \mathrm{d}} x){\, \mathrm{d}} t . \end{align*}

By construction, it is not hard to see that $\boldsymbol{\mu }^{n,\delta } \to \widehat{\boldsymbol{\mu }}^\delta$ weakly, where

\begin{align*} \widehat \mu _t^\delta \,:\!=\, \begin{cases} \mu _0 &\text{if } t \in (0,\delta ] \\ \mu _{\frac{t-\delta }{1-2\delta }} &\text{if } t \in \mathcal{I}_\delta \\ \mu _1 &\text{if } t \in [1-\delta, 1) \end{cases}, \quad{\, \mathrm{d}} \widehat{\boldsymbol{\mu }}^\delta (t,x) \,:\!=\, \widehat \mu _t^\delta ({\, \mathrm{d}} x){\, \mathrm{d}} t . \end{align*}

We stress that, as in (3.13), the rescaled curve $t \mapsto \widehat \mu _t^\delta$ could have discontinuities at $t=\delta$ and $t= 1-\delta$ , corresponding to the possible jumps in the limit as $n \to \infty$ for $\boldsymbol{\mu }^n$ at $\{0,1\}$ . Nevertheless, $\widehat{\boldsymbol{\mu }}^\delta$ is a competitor for ${\mathbb{MA}}(\mu _0,\mu _1)$ , which, by lower semicontinuity of $\mathbb{A}$ , ensures that

(3.27) \begin{align} \liminf _{n \to \infty }{\mathbb{A}}(\boldsymbol{\mu }^{n,\delta }) \geq{\mathbb{A}}(\widehat{\boldsymbol{\mu }}^\delta ) \geq{\mathbb{MA}}(\mu _0,\mu _1) . \end{align}

In order to estimate the left-hand side of the latter displayed equation, we seek a suitable vector meausure $\boldsymbol{\nu }^{n,\delta }$ so that $(\boldsymbol{\mu }^{n,\delta }, \boldsymbol{\nu }^{n,\delta }) \in \mathsf{CE}$ and whose action ${\mathbb{A}}(\boldsymbol{\mu }^{n,\delta },\boldsymbol{\nu }^{n,\delta })$ is comparable with ${\mathbb{A}}(\boldsymbol{\mu }^n,\boldsymbol{\nu }^n)$ for small $\delta \gt 0$ .

It is useful to introduce the following notation: for $\delta \in (0,1/2)$ ,

\begin{align*} r_\delta : (0,1) \to \mathcal{I}_\delta, \quad r_\delta (s) \,:\!=\, (1-2\delta )s + \delta, \end{align*}
\begin{align*} R_\delta :(0,1) \times{\mathbb{T}^d} \to \mathcal{I}_\delta \times{\mathbb{T}^d}, \quad R_\delta (s,x) \,:\!=\, (r_\delta (s), x) . \end{align*}

Define $\widehat \iota _\delta : \mathcal{M}^d(\mathcal{I}_\delta \times{\mathbb{T}^d}) \to \mathcal{M}^d\bigl ((0,1)\times{\mathbb{T}^d}\bigr )$ as the natural embedding obtained by extending to $0$ any measure outside $\mathcal{I}_\delta$ , and set

\begin{align*} \boldsymbol{\nu }^{n,\delta } \,:\!=\, \widehat \iota _\delta \big [ (R_\delta )_{\#} \boldsymbol{\nu }^n \big ] \in \mathcal{M}^d((0,1)\times{\mathbb{T}^d}) . \end{align*}

The proof that $( \boldsymbol{\mu }^{n,\delta }, \boldsymbol{\nu }^{n,\delta } ) \in \mathsf{CE}$ works as in (3.16). In the same spirit as in (3.15), we claim that

(3.28) \begin{align} {\mathbb{A}}(\boldsymbol{\mu }^{n,\delta }, \boldsymbol{\nu }^{n,\delta }) \leq \bigl (1+C(f)\delta \bigr ){\mathbb{A}}(\boldsymbol{\mu }^n, \boldsymbol{\nu }^n) + C(f) \delta \Big ( 1 + \boldsymbol{\mu }^n ((0,1) \times{\mathbb{T}^d}) \Big ), \end{align}

where $C(f) \in \mathbb{R}_+$ only depends on $f$ . The combination of (3.26), (3.27) and (3.28), and the arbitrariness of $\delta$ would then suffice to conclude the proof.

We are left with the proof of the claim (3.28), which is a bit more involved, compared to that of (3.15), due to the presence of the singular part at the continuous level. We need the following.

Lemma 3.16. Let $\boldsymbol{\sigma } \in \mathcal{M}_+\bigl ( (0,1) \times{\mathbb{T}}^d \bigr )$ be a singular measure with respect to $\mathscr{L}^{\,d+1}$ . Then, the measure $(R_\delta )_{\#} \boldsymbol{\sigma } \in \mathcal{M}_+ (\mathcal{I}_\delta \times{\mathbb{T}^d})$ is also singular with respect to $\mathscr{L}^{\,d+1}$ . Moreover, for every measure $\boldsymbol{\xi } = f \mathscr{L}^{\,d+1} + f^\perp \boldsymbol{\sigma } \in \mathcal{M}^n((0,1) \times{\mathbb{T}^d})$ , we have the decomposition

\begin{align*} (R_\delta )_{\#} \boldsymbol{\xi } = f^\delta \mathscr{L}^{\,d+1} + f^{\delta, \perp } (R_\delta )_{\#} \boldsymbol{\sigma }, \end{align*}

where the respective densities are given by the formulas

\begin{align*} f^\delta (t,x) = \frac 1{1-2\delta } f\bigl (r_\delta ^{-1}(t), x\bigr ) \quad \text{ and }\quad f^{\delta, \perp }(t,x) = f^\perp \bigl (r_\delta ^{-1}(t), x\bigr ) . \end{align*}

Proof. By assumption, $\boldsymbol{\sigma }$ is singular with respect to $\mathscr{L}^{\,d+1}$ , which means there exists a set $A \subset (0,1) \times{\mathbb{T}^d}$ such that $\mathscr{L}^{\,d+1}(A) = 0 = \boldsymbol{\sigma }(A^c)$ . By the very definition of push-forward and the bijectivity of $R_\delta$ , we then have that

\begin{align*} (R_\delta )_{\#} \boldsymbol{\sigma } \big ((R_\delta (A))^c \big ) = \boldsymbol{\sigma } \Big ( R_\delta ^{-1} \big (R_\delta (A^c) \big ) \Big ) = \boldsymbol{\sigma }(A^c) = 0, \end{align*}

whereas, by the scaling properties of the Lebesgue measure, we have that $\mathscr{L}^{\,d+1}(R_\delta (A)) = (1-2\delta ) \mathscr{L}^{\,d+1}(A) = 0$ , which shows the claimed singularity. The second part of the lemma follows from the fact that $(R_\delta )_{\#} \mathscr{L}^{\,d+1} = (1-2\delta )^{-1} \mathscr{L}^{\,d+1}$ and the following statement: for every $\boldsymbol{\xi }' = f' \boldsymbol{\sigma }'$ with $\boldsymbol{\sigma }' \in \mathcal{M}_+((0,1) \times{\mathbb{T}^d})$ , we claim that

(3.29) \begin{align} \frac{{\, \mathrm{d}} (R_\delta )_{\#} \boldsymbol{\xi }'}{{\, \mathrm{d}} (R_\delta )_{\#} \boldsymbol{\sigma }'} (t,x) = f'(R_\delta ^{-1}(t,x)), \quad \forall (t,x) \in \mathcal{I}_\delta \times{\mathbb{T}^d} . \end{align}

Indeed, by definition of push-forward, we have for every test function $\varphi \in C_b$

\begin{align*} \int \varphi{\, \mathrm{d}} (R_\delta )_{\#} \boldsymbol{\xi }' = \int (\varphi \circ R_\delta ){\, \mathrm{d}} \boldsymbol{\xi }' = \int (\varphi \circ R_\delta ) f'{\, \mathrm{d}} \boldsymbol{\sigma }' = \int \varphi \cdot (f' \circ R_\delta ^{-1}){\, \mathrm{d}} (R_\delta )_{\#} \boldsymbol{\sigma }', \end{align*}

which indeed shows (3.29).

Let

\begin{align*} \boldsymbol{\mu }^n = \rho ^n{\, \mathrm{d}} \mathscr{L}^{\,d+1} + \rho ^{n,\perp }{\, \mathrm{d}} \boldsymbol{\sigma } \quad \text{ and }\quad \boldsymbol{\nu }^n = j^n{\, \mathrm{d}} \mathscr{L}^{\,d+1} + j^{n,\perp }{\, \mathrm{d}} \boldsymbol{\sigma } \end{align*}

be Lebesgue decompositions. We apply Lemma 3.16 to both $\boldsymbol{\mu }^n$ and $\boldsymbol{\nu }^n$ and find that, on $\mathcal{I}_\delta \times{\mathbb{T}^d}$ , we have

\begin{gather*} \boldsymbol{\mu }^{n,\delta } = \rho ^{n,\delta }{\, \mathrm{d}} \mathscr{L}^{\,d+1} + \rho ^{n, \delta, \perp }{\, \mathrm{d}} (R_\delta )_{\#} \boldsymbol{\sigma } \quad \text{ and }\quad \boldsymbol{\nu }^{n,\delta } = j^{n,\delta }{\, \mathrm{d}} \mathscr{L}^{\,d+1} + j^{n, \delta, \perp }{\, \mathrm{d}} (R_\delta )_{\#} \boldsymbol{\sigma }, \end{gather*}

with $(R_\delta )_{\#} \boldsymbol{\sigma }$ singular with respect to $\mathscr{L}^{\,d+1}$ and

(3.30) \begin{align} \rho ^{n,\delta }(t,x) &= \big ( \rho ^n \circ R_\delta ^{-1} \big ) (t,x), \qquad &&\rho ^{n, \delta, \perp }(t,x) = (1-2\delta ) \big ( \rho ^{n,\perp } \circ R_\delta ^{-1} \big ) (t,x), \\ j^{n,\delta }(t,x) &= \frac 1{1-2\delta } \big ( j^n \circ R_\delta ^{-1} \big ) (t,x), \qquad &&j^{n, \delta, \perp }(t,x) = \big ( j^{n,\perp } \circ R_\delta ^{-1} \big ) .\nonumber \end{align}

Further consider the Lebesgue decompositions

\begin{equation*} \mu _i^n = \rho _i^n {\, \mathrm {d}} \mathscr {L}^d + \rho _i^{n,\perp } {\, \mathrm {d}} \sigma _i, \qquad i \in \{0,1\} \end{equation*}

for some $\sigma _1,\sigma _2 \in \mathcal{M}_+({\mathbb{T}}^d)$ singular w.r.t. $\mathscr{L}^d$ . The action of $(\boldsymbol{\mu }^{n,\delta }, \boldsymbol{\nu }^{n,\delta })$ is given byFootnote 2

\begin{align*}{\mathbb{A}}(\boldsymbol{\mu }^{n,\delta }, \boldsymbol{\nu }^{n,\delta }) &={\mathbb{A}}^{\mathcal{I}_\delta } (\boldsymbol{\mu }^{n,\delta }, \boldsymbol{\nu }^{n,\delta }) + \sum _{i=0,1} \delta \Big ( \int _{{\mathbb{T}^d}} f(\rho _i^n,0){\, \mathrm{d}} \mathscr{L}^d + \int _{{\mathbb{T}^d}} f^\infty (\rho _i^{n,\perp },0){\, \mathrm{d}} \sigma _i \Big ), \end{align*}

where we used the notation

\begin{equation*} {\mathbb {A}}^{\mathcal {I}_\delta }(\boldsymbol {\mu }^{n,\delta }, \boldsymbol {\nu }^{n,\delta }) \,:\!=\, \int _{\mathcal {I}_\delta \times {\mathbb {T}^d}} f(\rho ^{n,\delta }, j^{n,\delta }) {\, \mathrm {d}} \mathscr {L}^{d+1} + \int _{ \mathcal {I}_\delta \times {\mathbb {T}^d}} f^\infty ( \rho ^{n,\delta, \perp }, j^{n,\delta, \perp } ) {\, \mathrm {d}} (R_\delta )_{\#} \boldsymbol {\sigma } . \end{equation*}

Making use of the formulas (3.30) and the homogeneity of $f^\infty$ , we find

(3.31) \begin{align} {\mathbb{A}}^{\mathcal{I}_\delta }&(\boldsymbol{\mu }^{n,\delta }, \boldsymbol{\nu }^{n,\delta }) = (1-2\delta ) \int _{(0,1) \times{\mathbb{T}^d}} f\Big ( \rho ^n, \frac{j^n}{1-2\delta } \Big ){\, \mathrm{d}} \mathscr{L}^{\,d+1} + \int _{(0,1) \times{\mathbb{T}^d}} f^\infty \big ( (1-2\delta ) \rho ^{n,\perp }, j^{n,\perp } \big ){\, \mathrm{d}} \boldsymbol{\sigma } \\ &\qquad = (1-2\delta ) \bigg ( \int _{(0,1) \times{\mathbb{T}^d}} f\Big ( \rho ^n, \frac{j^n}{1-2\delta } \Big ){\, \mathrm{d}} \mathscr{L}^{\,d+1} + \int _{(0,1) \times{\mathbb{T}^d}} f^\infty \Big ( \rho ^{n,\perp }, \frac{j^{n,\perp }}{1-2\delta } \Big ){\, \mathrm{d}} \boldsymbol{\sigma } \bigg ) .\nonumber \end{align}

Furthermore, it follows from the linear growth assumption that, for $i=0,1$ ,

\begin{align*} \int _{{\mathbb{T}^d}} f(\rho _i^n,0){\, \mathrm{d}} \mathscr{L}^d + \int _{{\mathbb{T}^d}} f^\infty (\rho _i^{n,\perp },0){\, \mathrm{d}} \sigma _i \leq C (\mu _i^n({\mathbb{T}^d}) +1) = C \bigl ( \boldsymbol{\mu }^n ((0,1) \times{\mathbb{T}^d}) + 1 \bigr ) \end{align*}

as well as, by (3.31), the nonnegativity of $f$ , and Assumption 2.2,

\begin{align*}{\mathbb{A}}^{\mathcal{I}_\delta }(\boldsymbol{\mu }^{n,\delta }, \boldsymbol{\nu }^{n,\delta }) &\leq{\mathbb{A}}(\boldsymbol{\mu }^n, \boldsymbol{\nu }^n) + 2\delta (\mathrm{Lip} f ) \bigg ( \int _{(0,1) \times{\mathbb{T}^d}} | j^n |{\, \mathrm{d}} \mathscr{L}^{\,d+1} + \int _{(0,1) \times{\mathbb{T}^d}} | j^{n,\perp } |{\, \mathrm{d}} \boldsymbol{\sigma } \bigg ) \\ &\leq{\mathbb{A}}(\boldsymbol{\mu }^n, \boldsymbol{\nu }^n) + \frac{2 \delta (\mathrm{Lip} f)}{c} \bigg ({\mathbb{A}}(\boldsymbol{\mu }^n,\boldsymbol{\nu }^n) + C(1+\|\boldsymbol{\mu }^n\|_{\mathrm{TV}}) \bigg ) . \end{align*}

We thus conclude (3.28).

4. Analysis of the cell problem with examples

This section is devoted to the characterisation and illustration of $f_{\mathrm{hom}}$ in the case where the function $F$ is of the form

(4.1) \begin{align} F(m,J) = F(J) = \sum _{ (x, y) \in \mathcal{E}^Q } \alpha _{xy} \lvert J(x,y)\rvert \end{align}

for some strictly positive function $\alpha : \mathcal{E}^Q \ni (x,y) \mapsto \alpha _{xy}\gt 0$ . A natural problem of interest is to determine whether/when the $\Gamma$ -limit ${\mathbb{MA}}_{\mathrm{hom}}$ can be the $W_1$ -distance. The analogous problem for the $W_2$ -distance has been extensively studied in [Reference Gladbach, Kopfer and Maas24] and [Reference Gladbach, Kopfer, Maas and Portinale23] in the case where the graph stucture is associated with finite-volume partitions.

4.1. Discrete $1$ -Wasserstein distance

We start the analysis of this special setting by observing that, in this case, the discrete functional ${\mathcal{MA}}_{\varepsilon }$ actually coincides with the $\mathbb{W}_1$ distance associated to a natural induced metric structure. In order to prove this fact, we first define $\widetilde \alpha ^{\varepsilon } : \mathcal{E}_{\varepsilon } \to \mathbb{R}_+$ as the unique function such that

\begin{equation*} \frac {\tau _{\varepsilon }^z \widetilde \alpha ^{\varepsilon }}{\varepsilon } \Big |_{\mathcal {E}^Q} \,:\!=\, \alpha \, \qquad z \in \mathbb {Z}_{\varepsilon }^d . \end{equation*}

It is easy to check that this definition is well-posed and determines the value of $\widetilde \alpha _{xy}$ for every $(x,y) \in \mathcal{E}_{\varepsilon }$ . Further let

\begin{equation*} \alpha _{xy}^{\varepsilon } = \frac {\widetilde \alpha _{xy}^{\varepsilon } + \widetilde \alpha _{yx}^{\varepsilon }}{2}, \qquad (x,y) \in \mathcal {E}_{\varepsilon } \end{equation*}

be the symmetrisation of $\widetilde \alpha ^{\varepsilon }$ . Given $J \in \mathbb{R}^{\mathcal{E}_{\varepsilon }}_a$ , we can write $\mathcal{F}_{\varepsilon }(J)$ in terms of $\alpha ^{\varepsilon }$ . Precisely,

\begin{align*} \mathcal{F}_{\varepsilon }(J) &= \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \varepsilon ^d F \left ( \frac{\tau _{\varepsilon }^z J }{\varepsilon ^{d-1}} \right ) = \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \varepsilon ^d \sum _{(\widehat x, \widehat y) \in \mathcal{E}^Q } \alpha _{\widehat x \widehat y} \frac{ \lvert \tau ^z_{\varepsilon } J(\widehat x, \widehat y)\rvert }{\varepsilon ^{d-1}} \\ &= \sum _{z \in \mathbb{Z}^d_{\varepsilon }} \sum _{(\widehat x, \widehat y) \in \mathcal{E}^Q} \tau ^z_{\varepsilon } \widetilde \alpha ^{\varepsilon }(\widehat x, \widehat y) \lvert \tau _{\varepsilon }^z J(\widehat x, \widehat y)\rvert = \sum _{(x,y) \in \mathcal{E}_{\varepsilon }} \widetilde \alpha ^{\varepsilon }_{xy} \lvert J(x, y)\rvert \\ &= \sum _{(x,y) \in \mathcal{E}_{\varepsilon }} \alpha ^{\varepsilon }_{xy} \lvert J(x, y)\rvert, \end{align*}

where in the last passage we used that $\lvert J\rvert$ is symmetric. We define a distance on $\mathcal{X}_{\varepsilon }$ given by

\begin{align*} d_{\varepsilon }(x,y)\,:\!=\,{\mathcal{MA}}_{\varepsilon }(\delta _x, \delta _y), \qquad \forall x,y \in \mathcal{X}_{\varepsilon } . \end{align*}

One can easily show that $d_{\varepsilon }$ indeed defines a metric on $\mathcal{X}_{\varepsilon }$ . In fact, $d_{\varepsilon }$ can be seen as a weighted graph distance.

Proposition 4.1. For every $x,y \in \mathcal{X}_{\varepsilon }$ , we have

\begin{equation*} d_{\varepsilon }(x,y) = \inf \left \{ \sum _{i=0}^{k-1} 2 \alpha _{x_i x_{i+1}}^{\varepsilon } \, : \, x_0 = x, \, x_k = y, \, (x_i,x_{i+1}) \in \mathcal {E}_{\varepsilon } \ \, \, \forall i, \, k \in \mathbb {N} \right \} . \end{equation*}

Proof. The inequality $\le$ directly follows by choosing unit fluxes along admissible paths: let $x_0=x,x_1,\dots, x_{k-1},x_k=y$ be a path, i.e. $(x_i,x_{i+1}) \in \mathcal{E}_{\varepsilon }$ for every $i = 0,1,\dots, k$ , and consider

(4.2) \begin{equation} J^P \,:\!=\, \sum _{i=0}^{k-1} \bigl (\delta _{(x_i,x_{i+1})}-\delta _{(x_{i+1},x_i)} \bigr ), \end{equation}

which has divergence equal to $\delta _x-\delta _y$ . Then,

\begin{align*} d_{\varepsilon }(x,y) &={\mathcal{MA}}_{\varepsilon }(\delta _x, \delta _y) = \inf \left \{ \mathcal{F}_{\varepsilon }(J) \, : \, \mathsf{div} J = \delta _x-\delta _y \right \} \\ &= \inf \left \{ \sum _{(x,y) \in \mathcal{E}_{\varepsilon }} \alpha ^{\varepsilon }_{xy} \lvert J(x, y)\rvert \, : \, \mathsf{div} J = \delta _x-\delta _y \right \} \\ &\le \sum _{(x,y) \in \mathcal{E}_{\varepsilon }} \alpha ^{\varepsilon }_{xy} \lvert J^P(x,y)\rvert \le \sum _{i=0}^{k-1} 2 \alpha _{x_i x_{i+1}}^{\varepsilon }, \end{align*}

where in the last inequality we used that $\alpha ^{\varepsilon }$ is symmetric.

To prove the converse, let $\bar J \in \mathbb{R}^{\mathcal{E}_{\varepsilon }}_a$ be an optimal flux for ${\mathcal{MA}}_{\varepsilon }(\delta _x,\delta _y)$ , that is,

\begin{equation*} \mathsf {div} \bar J = \delta _x-\delta _y \quad \text { and }\quad {\mathcal {MA}}_{\varepsilon }(\delta _x,\delta _y) = \sum _{ (x, y) \in \mathcal {E}_{\varepsilon } } \alpha _{xy}^{\varepsilon } \lvert \bar J(x,y)\rvert . \end{equation*}

Since the graph $\mathcal{E}_{\varepsilon }$ is finite, in order for $\bar J$ to satisfy the divergence condition, there must exist a simple path $x_0 = x, x_1,\dots, x_k = y$ such that $(x_i,x_{i+1}) \in \mathcal{E}_{\varepsilon }$ and $\bar J(x_i,x_{i+1}) \gt 0$ for every $i$ . Let $J^P$ be the associated vector field as in (4.2). Note that, for every $\lambda \in \mathbb{R}$ , we have $\mathsf{div} \bigl ((1-\lambda ) \bar J + \lambda J^P\bigr ) = \delta _x - \delta _y$ . Furthermore, the function

\begin{equation*} \lambda \mapsto \mathcal {F}_{\varepsilon }((1-\lambda ) \bar J + \lambda J^P\bigr ) = \sum _{(x,y) \in \mathcal {E}_{\varepsilon }} \alpha _{xy}^{\varepsilon } \lvert (1-\lambda )\bar J(x,y) + \lambda J^P(x,y)\rvert \end{equation*}

is differentiable at $\lambda = 0$ , since $(\bar J(x,y) = 0) \Rightarrow (J^P(x,y) = 0)$ . By optimality, the derivative at $\lambda = 0$ must equal $0$ , i.e.

\begin{equation*} 0= \sum _{ (x, y) \in \mathcal {E}_{\varepsilon } } \alpha _{xy}^{\varepsilon } \bigl ( J^P(x,y) - \bar J(x,y) \bigr ) {\textrm {sgn}} \bar J(x,y) = \sum _{ (x, y) \in \mathcal {E}_{\varepsilon } } \alpha _{xy}^{\varepsilon } J^P(x,y) {\textrm {sgn}} \bar J(x,y) - d_{\varepsilon }(x,y), \end{equation*}

and, since $(J^P(x,y) \neq 0) \Rightarrow ({\textrm{sgn}}\, J^P(x,y) ={\textrm{sgn}}\, \bar J(x,y))$ , we have

\begin{equation*} d_{\varepsilon }(x,y) = \sum _{(x,y) \in \mathcal {E}_{\varepsilon }} \alpha _{xy}^{\varepsilon } \lvert J^P(x,y)\rvert = 2\sum _{i=1}^k \alpha _{x_i x_{i+1}}^{\varepsilon }, \end{equation*}

where, in the last equality, we used that the path is simple (and the symmetry of $\alpha ^{\varepsilon }$ ). This shows the inequality $\geq$ and concludes the proof.

Consider the $1$ -Wasserstein distance associted to $d_{\varepsilon }$ , that is,

\begin{align*} \mathbb{W}_{1,\varepsilon }(m_0,m_1) = \inf \left \{ \int _{\mathcal{X}_{\varepsilon } \times \mathcal{X}_{\varepsilon }} d_{\varepsilon }(x,y){\, \mathrm{d}} \pi (x,y) \, : \, (e_0)_{\#} \pi = m_0, \quad (e_1)_{\#} \pi = m_1 \right \}, \end{align*}

as well as, by Kantorovich duality,

\begin{align*} \mathbb{W}_{1,\varepsilon }(m_0,m_1) = \sup \left \{ \int _{\mathcal{X}_{\varepsilon }} \varphi{\, \mathrm{d}} (m_0-m_1) \, : \, \mathrm{Lip}_{d_{\varepsilon }}(\varphi ) \leq 1 \right \}, \end{align*}

for every $m_0,m_1 \in \mathscr{P}(\mathcal{X}_{\varepsilon })$ .

Proposition 4.2. For every $m_0,m_1 \in \mathscr{P}(\mathcal{X}_{\varepsilon })$ , we have

(4.3) \begin{align} {\mathcal{MA}}_{\varepsilon }(m_0,m_1) = \mathbb{W}_{1,\varepsilon }(m_0,m_1) . \end{align}

Proof of ≥. Fix $m_0,m_1 \in \mathscr{P}(\mathcal{X}_{\varepsilon })$ and set $m\,:\!=\, m_0-m_1$ . Let $\bar J \in \mathbb{R}_a^{\mathcal{E}_{\varepsilon }}$ be an optimal flux for ${\mathcal{MA}}_{\varepsilon }(m_0,m_1)$ , that is,

\begin{align*} \mathsf{div} \bar J = m \quad \text{ and }\quad{\mathcal{MA}}_{\varepsilon }(m_0,m_1) = \sum _{ (x, y) \in \mathcal{E}_{\varepsilon } } \alpha _{xy}^{\varepsilon } \lvert \bar J(x,y)\rvert . \end{align*}

Let $\varphi : \mathcal{X}_{\varepsilon } \to \mathbb{R}$ be such that $\mathrm{Lip}_{d_{\varepsilon }} \varphi \leq 1$ , i.e. $ \left | \varphi (y) - \varphi (x) \right | \leq d_{\varepsilon }(x,y)$ for $ x, y \in \mathcal{X}_{\varepsilon } .$ Then,

\begin{align*} \int _{\mathcal{X}_{\varepsilon }} \varphi{\, \mathrm{d}} m &= \int _{\mathcal{X}_{\varepsilon }} \varphi{\, \mathrm{d}} \mathsf{div} \bar J = \sum _{x \in \mathcal{X}_{\varepsilon }} \varphi (x) \sum _{y \sim x} \bar J(x,y) = \frac 12 \sum _{(x,y) \in \mathcal{E}_{\varepsilon }} \varphi (x) \big ( \bar J(x,y) - \bar J(y,x) \big )\\ &= \frac 12 \sum _{(x,y) \in \mathcal{E}_{\varepsilon }} \big ( \varphi (y) - \varphi (x) \big ) \bar J(x,y) \leq \frac 12 \sum _{(x,y) \in \mathcal{E}_{\varepsilon }} d_{\varepsilon }(x,y) \lvert \bar J(x,y)\rvert . \end{align*}

In order to conclude, we make the following crucial observation: as a consequence of the optimality of $\bar J$ , we claim that

(4.4) \begin{equation} \bar J(x,y) \neq 0 \quad \Longrightarrow \quad d_{\varepsilon }(x,y) = 2 \alpha _{xy}^{\varepsilon } . \end{equation}

To this end, assume that $\bar J(x,y) \neq 0$ and consider an optimal $J^{(x,y)}$ for $d_{\varepsilon }(x,y) ={\mathcal{MA}}_{\varepsilon }(\delta _x, \delta _y)$ . Note that, by construction,

\begin{align*} \mathsf{div} \big ( J^{(x,y)} \big ) = \delta _x - \delta _y = \mathsf{div} \widetilde J, \qquad \text{where} \ \widetilde J \,:\!=\, \delta _{(x,y)} - \delta _{(y,x)}, \end{align*}

which in turns also implies that

\begin{align*} \mathsf{div} \big ( \bar J + \bar J(x,y) \big ( J^{(x,y)} - \widetilde J \big ) \big ) = \mathsf{div} \bar J . \end{align*}

By optimality of $J^{(x,y)}$ , we have

(4.5) \begin{equation} \mathcal{F}_{\varepsilon }(\widetilde J) = 2 \alpha _{xy}^{\varepsilon } \geq \mathcal{F}_{\varepsilon } \big ( J^{(x,y)} \big ) = \sum _{(\widetilde x, \widetilde y) \in \mathcal{E}_{\varepsilon }} \alpha _{\widetilde x, \widetilde y}^{\varepsilon } \lvert J^{(x,y)}(\widetilde x, \widetilde y)\rvert, \end{equation}

whereas the optimality of $\bar J$ yields

\begin{align*} \mathcal{F}_{\varepsilon } \bigl (\bar J + \bar J(x,y) \big ( J^{(x,y)} - \widetilde J \big )\bigr ) &= \sum _{(\widetilde x, \widetilde y) \in \mathcal{E}_{\varepsilon } \setminus \{(x,y), (y,x)\} } \alpha _{\widetilde x \widetilde y}^{\varepsilon } \lvert \bar J(\widetilde x, \widetilde y) + \bar J(x,y) J^{(x,y)}(\widetilde x, \widetilde y)\rvert \\ &\quad + \alpha _{xy}^{\varepsilon } \lvert \bar J(x,y) J^{(x,y)}(x,y)\rvert + \alpha _{yx}^{\varepsilon } \lvert \bar J(x,y) J^{(x,y)}(y,x)\rvert \\ &\ge \mathcal{F}_{\varepsilon }(\bar J) = \sum _{(\widetilde x, \widetilde y) \in \mathcal{E}_{\varepsilon }} \alpha _{\widetilde x\widetilde y}^{\varepsilon } \lvert \bar J(\widetilde x, \widetilde y)\rvert . \end{align*}

By applying the triangle inequality and simplifying the latter formula, we find

(4.6) \begin{equation} \sum _{(\widetilde x,\widetilde y) \in \mathcal{E}_{\varepsilon }} \alpha _{\widetilde x\widetilde y}^{\varepsilon } \lvert \bar J(x,y) J^{(x,y)}(\widetilde x,\widetilde y)\rvert \ge 2\alpha _{xy}^{\varepsilon } \lvert \bar J(x,y)\rvert . \end{equation}

The combination of (4.5) and (4.6) implies $d_{\varepsilon }(x,y) = 2\alpha _{xy}^{\varepsilon }$ . With (4.4) at hand, we can write

\begin{equation*} \int _{\mathcal {X}_{\varepsilon }} \varphi {\, \mathrm {d}} m \le \sum _{(x,y) \in \mathcal {E}_{\varepsilon }} \alpha _{xy}^{\varepsilon } \lvert \bar J(x,y)\rvert = {\mathcal {MA}}_{\varepsilon }(m_0,m_1), \end{equation*}

and we conclude by arbitrariness of $\varphi$ .

Proof of ≤. Let $\pi$ be such that $(e_i)_{\#} \pi = m_i$ for $i=0,1$ . Further, for every $x,y \in \mathcal{X}_{\varepsilon }$ , let $J^{(x,y)} \in \mathbb{R}^{\mathcal{E}_{\varepsilon }}_a$ be optimal for ${\mathcal{MA}}_{\varepsilon }(\delta _x, \delta _y)$ . It follows from a direct computation that the divergence of the asymmetric flux

\begin{equation*} J \,:\!=\, \sum _{x,y \in \mathcal {X}_{\varepsilon }} \pi (x,y) J^{(x,y)} \end{equation*}

is equal to $m_0 - m_1$ . Thus,

\begin{equation*} {\mathcal {MA}}_{\varepsilon }(m_0,m_1) \le \sum _{(\widetilde x, \widetilde y) \in \mathcal {E}_{\varepsilon }} \alpha _{\widetilde x \widetilde y}^{\varepsilon } \lvert J(\widetilde x, \widetilde y)\rvert \le \sum _{x,y \in \mathcal {X}_{\varepsilon }} \pi (x,y) \sum _{(\widetilde x,\widetilde y) \in \mathcal {E}_{\varepsilon }} \alpha _{\widetilde x \widetilde y}^{\varepsilon } \lvert J^{(x,y)}(\widetilde x, \widetilde y)\rvert = \int _{\mathcal {X}_{\varepsilon } \times \mathcal {X}_{\varepsilon }} d_{\varepsilon } {\, \mathrm {d}} \pi, \end{equation*}

and we conclude by arbitrariness of $\pi$ .

In view of the equality ${\mathcal{MA}}_{\varepsilon } = \mathbb{W}_{1,\varepsilon }$ , it is worth noting that for cost functions of the form (4.1) there are (at least) two different possible methods to show discrete-to-continuum limits for ${\mathcal{MA}}_{\varepsilon }$ . One such method is provided by the current work and makes use of the $\Gamma$ -convergence of $\mathcal{A}_{\varepsilon }$ to ${\mathbb{A}}_{\mathrm{hom}}$ proved in [[Reference Gladbach, Kopfer, Maas and Portinale23], Theorem 5.4]. The convergence of the “weighted graph distance” $d_{\varepsilon }$ follows a posteriori. Another approach is to study directly the scaling limits of the distance $d_{\varepsilon }$ as $\varepsilon \to 0$ and, from that, infer the convergence of the associated $1$ -Wasserstein distances, in a similar spirit as in [Reference Buttazzo, De Pascale and Fragalà4].

4.2. General properties of $f_{\mathrm{hom}}$

For $j \in \mathbb{R}^d$ , recall that

(4.7) \begin{equation} f_{\mathrm{hom}}(j) \,:\!=\, \inf \left \{ F(J) \, : \, J \in{\mathsf{Rep}}(j) \right \}, \end{equation}

where ${\mathsf{Rep}}(j)$ is the set of all $\mathbb{Z}^d$ -periodic functions $J \in \mathbb{R}^{\mathcal{E}}_a$ such that

\begin{equation*} \mathsf {Eff}(J) \,:\!=\, \frac {1}{2} \sum _{(x,y) \in \mathcal {E}^Q} J(x,y)(y_{\mathsf {z}} - x_{\mathsf {z}}) = j \quad \text {and} \quad \mathsf {div} J \equiv 0 . \end{equation*}

As noted in [[Reference Gladbach, Kopfer, Maas and Portinale23], Lemma 4.7], we may as well write $\min$ in place of $\inf$ in (4.7).

Our first observation is that, indeed, the homogenised density is a norm. This has already been proved in [[Reference Gladbach, Kopfer, Maas and Portinale23], Corollary 5.3]; for the sake of completeness, we provide here a simple proof in our setting.

Proposition 4.3. The function $f_{\mathrm{hom}}$ is a norm.

Proof. Finiteness follows from the nonemptiness of the set of representatives proved in [[Reference Gladbach, Kopfer, Maas and Portinale23], Lemma 4.5]. To prove positiveness, take any $j \in \mathbb{R}^d$ and $J \in{\mathsf{Rep}}(j)$ . For every norm $\|{\cdot }\|$ , we have

(4.8) \begin{equation} \|{j}\| = \|{\mathsf{Eff}(J)}\| \le \frac{1}{2}\sum _{(x,y) \in \mathcal{E}^Q} \lvert J(x,y)\rvert \|{y_{\mathsf{z}} - x_{\mathsf{z}}}\| \le \frac{F(J)}{2} \max _{(x,y) \in \mathcal{E}^Q} \frac{\|{y_{\mathsf{z}} - x_{\mathsf{z}}}\|}{ \alpha _{xy}} . \end{equation}

The constant that multiplies $F(J)$ at the right-hand side is finite because every $\alpha _{xy}$ is strictly positive and the graph $(\mathcal{X},\mathcal{E})$ is locally finite. Absolute homogeneity and the triangle inequality follow from the absolute homogeneity and subadditivity of $F$ and the affinity of the constraints.

Hence, ${\mathbb{MA}}_{\mathrm{hom}}$ is always (i.e. for any choice of $(\alpha _{xy})_{x,y}$ and of the graph $(\mathcal{X},\mathcal{E})$ ) the $W_1$ -distance w.r.t. some norm. However, the norm $f_{\mathrm{hom}}$ can equal the $2$ -norm $\lvert \cdot \rvert _2$ only in dimension $d=1$ . In fact, the unit ball for $f_{\mathrm{hom}}$ has to be a polytope, namely the associated sphere is contained in the union of finitely many hyperplanes. These types of norms are also known as crystalline norms.

Proposition 4.4. The unit ball associated to the norm $f_{\mathrm{hom}}$ , namely

\begin{equation*} B \,:\!=\, \left \{ j \in \mathbb {R}^d \, : \, f_{\mathrm {hom}}(j) \le 1 \right \}, \end{equation*}

is the convex hull of finitely many points. In particular, the associated unit sphere is contained in the union of finitely many hyperplanes, i.e. $f_{\mathrm{hom}}$ is a crystalline norm.

Proof. Let $X$ be the vector space of all $\mathbb{Z}^d$ -periodic functions $J \in \mathbb{R}^{\mathcal{E}}_a$ such that $\mathsf{div} J \equiv 0$ . The sublevel set

\begin{equation*} X_1 \,:\!=\, \left \{ J \in X \, : \, F(J) \le 1 \right \} \end{equation*}

is clearly compact (due to the strict positivity of $(\alpha _{xy})_{x,y}$ ) and can be written as finite intersection of half-spaces, namely

\begin{equation*} X_1 = \bigcap _{r \in \{-1,1\}^{\mathcal {E}^Q} } \left \{ J \in X \, : \, \sum _{(x,y) \in \mathcal {E}^Q} \alpha _{xy} r_{xy} J(x,y) \le 1 \right \}. \end{equation*}

Thus, $X_1$ is the convex hull of some finite set of points $A$ , that is, $ X_1 ={\mathrm{conv}}(A)$ . Since $f_{\mathrm{hom}}$ is defined as a minimum, we have

\begin{equation*} B = \left \{ j \in \mathbb {R}^d \, : \, \exists J \in {\mathsf {Rep}}(j), \, F(J) \le 1 \right \} = \mathsf {Eff}(X_1) = \mathsf {Eff}({\mathrm {conv}}(A)) = {\mathrm {conv}}(\mathsf {Eff}(A)), \end{equation*}

where the last equality is due to the linearity of $\mathsf{Eff}$ .

4.3. Embedded graphs

To visualise some examples, we shall now focus on the case where $(\mathcal{X}, \mathcal{E})$ is embedded, in the sense that $V$ is a subset of $[0,1)^d$ , and we use the identification $(z,v) \equiv z+v$ (see also [[Reference Gladbach, Kopfer, Maas and Portinale23], Remark 2.2]). It has been proved in [[Reference Gladbach, Kopfer, Maas and Portinale23], Proposition 9.1] that, for embedded graphs, the identity

(4.9) \begin{equation} \mathsf{Eff}(J) = \frac{1}{2} \sum _{ (x, y) \in \mathcal{E}^Q } J(x,y)(y-x) \end{equation}

holds for every $\mathbb{Z}^d$ -periodic and divergence-free vector field $J \in \mathbb{R}^{\mathcal{E}}_a$ . In what follows, we also make the choice

\begin{equation*} \alpha _{xy} \,:\!=\, \frac {1}{2}\lvert x-y\rvert _2, \qquad (x,y) \in \mathcal {E}^Q . \end{equation*}

4.3.1. One-dimensional case with nearest-neighbour interaction

Assume $d=1$ , let $x_1 \lt x_2 \lt \cdots \lt x_k$ be an enumeration of $V$ , and set

\begin{equation*} \mathcal {E} \,:\!=\, \{ (x,y) \in \mathcal {X} \times \mathcal {X} \text { s.t., there is no $z \in \mathcal {X}$ strictly between $x$ and $y$} \}. \end{equation*}

In other words, denoting $x_0 = x_k - 1$ and $x_{k+1} = x_1 + 1$ ,

\begin{equation*} \mathcal {E} = \bigcup _{z \in \mathbb {Z}} \bigcup _{i=1}^k \left \{ (x_i,x_{i+1}) \right \} \cup \left \{ (x_i,x_{i-1}) \right \} . \end{equation*}

By rewriting (4.8) using (4.9), and by the definition of $f_{\mathrm{hom}}$ , we find

\begin{equation*} \lvert j\rvert \le f_{\mathrm {hom}}(j), \qquad j \in \mathbb {R}^d . \end{equation*}

On the other hand, given $j \in \mathbb{R}^d$ , choose

\begin{equation*} J(x,y) \,:\!=\, j {\textrm {sgn}}(y-x), \qquad (x,y) \in \mathcal {E} . \end{equation*}

This vector field is in ${\mathsf{Rep}}(j)$ because

\begin{equation*} \mathsf {div} J(x_i) = J(x_i,x_{i+1}) + J(x_i,x_{i-1}) = j - j = 0 \end{equation*}

for every $i$ , and

\begin{align*} \mathsf{Eff}(J) &= \frac{1}{2} \sum _{i=1}^k \bigl ( J(x_i,x_{i+1})(x_{i+1}-x_i) + J(x_i,x_{i-1})(x_{i-1}-x_i) \bigr ) \\ &= \frac{j}{2} \sum _{i=1}^k \bigl ( \lvert x_{i+1} - x_i\rvert + \lvert x_i - x_{i-1}\rvert \bigr ) \\ &= \frac{j}{2} \bigl ( x_{k+1} - x_1 + x_k - x_0 \bigr ) \\ &= j. \end{align*}

A similar computation shows that $F(J) = \lvert j\rvert$ , from which $f_{\mathrm{hom}}(j) = \lvert j\rvert$ .

Figure 2. Examples of graphs in $\mathbb{R}^2$ and corresponding unit balls for $f_{\mathrm{hom}}$ .

4.3.2. Cubic partition

Consider the case where $\mathcal{X} = \mathbb{Z}^d$ and

\begin{equation*} \mathcal {E} \,:\!=\, \left \{ (x,y) \in \mathbb {Z}^d \times \mathbb {Z}^d \, : \, \lvert x-y\rvert _\infty = 1 \right \}. \end{equation*}

It is a result of [[Reference Gladbach, Kopfer, Maas and Portinale23], Section 9.2] that

\begin{equation*} f_{\mathrm {hom}}(j) = \lvert j\rvert _1, \qquad j \in \mathbb {R}^d . \end{equation*}

Notice that, in this case, the $2$ -norm is evaluated only at vectors on the coordinate axes. Therefore, the same result holds when $\alpha _{xy} = \frac{1}{2}\lvert x-y\rvert _p$ , for any $p$ .

4.3.3. Graphs in $\mathbb{R}^2$

A few other examples in dimension $d=2$ are shown in Figure 2: for each one, we display the graph and the unit ball in the corresponding norm $f_{\mathrm{hom}}$ . To algorithmically construct the unit balls, we solve the variational problem (4.7) for every $j$ on a discretisation of the circle $\mathbb S^1$ . In turn, this is achieved with the help of the Python library CVXPY [Reference Diamond and Boyd7], [Reference Agrawal, Verschueren, Diamond and Boyd3]. For visualisation, we make use of the library matplotlib [Reference Hunter28].

Acknowledgements

The authors are very grateful to Jan Maas for his useful feedback on a preliminary version of this work. The authors also thank the anonymous reviewers for the careful reading of the manuscript and for useful suggestions. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper

Financial support

L.P. gratefully acknowledges fundings from the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy – GZ 2047/1, Projekt-ID 390685813. F.Q. gratefully acknowledges support from the Austrian Science Fund (FWF) project 10.55776/F65.

Competing interests

None.

Footnotes

1 The definition is well-posed because $\varepsilon R_0$ is assumed to be smaller than $1/2$ .

2 Note that the definition of the action does not depend on the choice of the measure which is singular with respect to $\mathscr{L}^{\,d+1}$ , therefore we can use $(R_\delta )_{\#} \boldsymbol{\sigma }$ instead of $\boldsymbol{\sigma }$ .

References

Ambrosio, L., Fusco, N. & Pallara, D. (2000) Functions of Bounded Variation and Free Discontinuity Problems. Oxford Mathematical Monographs. The Clarendon Press, Oxford University Press, New York.CrossRefGoogle Scholar
Amar, M. & Vitali, E. (1998) Homogenization of periodic Finsler metrics. J. Convex Anal. 5(1), 171186.Google Scholar
Agrawal, A., Verschueren, R., Diamond, S. & Boyd, S. (2018) A rewriting system for convex optimization problems. J. Control Decis. 5(1), 4260.CrossRefGoogle Scholar
Buttazzo, G., De Pascale, L. & Fragalà, I. (2001) Topological equivalence of some variational problems involving distances. Discrete Contin. Dyn. Syst. 7(2), 247258.CrossRefGoogle Scholar
Benamou, J.-D. & Brenier, Y. (2000) A computational fluid mechanics solution to the Monge–Kantorovich mass transfer problem. Numer. Math. 84(3), 375393.CrossRefGoogle Scholar
Chow, S.-N., Huang, W., Li, Y. & Zhou, H. (2012) Fokker-Planck equations for a free energy functional or Markov process on a graph. Arch. Ration. Mech. Anal. 203(3), 9691008.CrossRefGoogle Scholar
Diamond, S. & Boyd, S. (2016) CVXPY: A Python-embedded modeling language for convex optimization. J Mach Learn Res. 17(83), 15.Google ScholarPubMed
Disser, K. & Liero, M. (2015) On gradient structures for Markov chains and the passage to Wasserstein gradient flows. Netw. Heterog. Media 10(2), 233253.CrossRefGoogle Scholar
Erbar, M., Henderson, C., Menz, G. & Tetali, P. (2017) Ricci curvature bounds for weakly interacting Markov chains. Electron. J. Probab. 22, 23.CrossRefGoogle Scholar
Esposito, A., Heinze, G., Pietschmann, J.-F. & Schlichting, A. (2023) Graph-to-local limit fora multi-species nonlocal cross-interaction system. PAMM, 23(4), e202300094.Google Scholar
Erbar, M., Maas, J. & Tetali, P. (2015) Discrete Ricci curvature bounds for Bernoulli-Laplace and random transposition models. Ann. Fac. Sci. Toulouse Math. 24(4), 781800.CrossRefGoogle Scholar
Esposito, A., Patacchini, F. S., Schlichting, A. & Slepcev, D. (2021) Nonlocal-interaction equation on graphs: Gradient flow structure and continuum limit. Arch. Ration. Mech. Anal. 240(2), 699760.CrossRefGoogle Scholar
Esposito, A., Patacchini, F. S. & Schlichting, A. (2024). On a class of nonlocal continuity equations on graphs. European J. Appl. Math., 35(1), 109126.CrossRefGoogle Scholar
Erbar, M. & Fathi, M. (2018) Poincaré, modified logarithmic Sobolev and isoperimetric inequalities for Markov chains with non-negative Ricci curvature. J. Funct. Anal. 274(11), 30563089.CrossRefGoogle Scholar
Erbar, M. & Maas, J. (2012) Ricci curvature of finite Markov chains via convexity of the entropy. Arch. Ration. Mech. Anal. 206(3), 9971038.CrossRefGoogle Scholar
Erbar, M. & Maas, J. (2014) Gradient flow structures for discrete porous medium equations. Discrete Contin. Dyn. Syst. 34(4), 13551374.CrossRefGoogle Scholar
Esposito, A. & Mikolás, L. (2024). On evolution PDEs on co-evolving graphs. Discrete Contin.Dyn. Syst., 44(9), 2660–2683.CrossRefGoogle Scholar
Fathi, M. & Maas, J. (2016) Entropic Ricci curvature bounds for discrete interacting systems. Ann. Appl. Probab. 26(3), 17741806.CrossRefGoogle Scholar
Forkert, D., Maas, J. & Portinale, L. (2022) Evolutionary Γ-convergence of entropic gradient flow structures for Fokker-Planck equations in multiple dimensions. SIAM J. Math. Anal. 54(4), 42974333.CrossRefGoogle Scholar
Trillos, N. García (2020) Gromov-Hausdorff limit of Wasserstein spaces on point clouds. Calc. Var. Partial Differ. Equ. 59(2), 43.Google Scholar
Gigli, N. & Maas, J. (2013) Gromov-Hausdorff convergence of discrete transportation metrics. SIAM J. Math. Anal. 45(2), 879899.CrossRefGoogle Scholar
Gladbach, P., Kopfer, E., Maas, J. & Portinale, L. (2020) Homogenisation of one-dimensional discrete optimal transport. J. Math. Pures Appl. 139, 204234.CrossRefGoogle Scholar
Gladbach, P., Kopfer, E., Maas, J. & Portinale, L. (2023) Homogenisation of dynamical optimal transport on periodic graphs. Calc. Var. Partial Differential Equations 62(5), 75.CrossRefGoogle ScholarPubMed
Gladbach, P., Kopfer, E. & Maas, J. (2020) Scaling limits of discrete optimal transport. SIAM J. Math. Anal. 52(3), 27592802.CrossRefGoogle Scholar
Gladbach, P., Maas, J. & Portinale, L. (2024) Stochastic Homogenisation of nonlinear minimum-cost flow problems: arXiv preprint arXiv:2412.05217, 2024.Google Scholar
Hraivoronska, A. & Tse, O. (2023) Diffusive limit of random walks on tessellations via generalized gradient flows. SIAM J. Math. Anal. 55(4), 29482995.CrossRefGoogle Scholar
Hraivoronska, A., Schlichting, A. & Tse, O. (2024) Variational convergence of the Scharfetter–Gummel scheme to the aggregation-diffusion equation and vanishing diffusion limit. Numer. Math. 156(6), 2221–2292.CrossRefGoogle Scholar
Hunter, J. D. (2007) Matplotlib: A 2d graphics environment. Comput Sci Eng. 9(3), 9095.CrossRefGoogle Scholar
Jordan, R., Kinderlehrer, D. & Otto, F. (1998) The variational formulation of the Fokker–Planck equation. SIAM J. Math. Anal. 29(1), 117.CrossRefGoogle Scholar
Lott, J. & Villani, C. (2007) Weak curvature conditions and functional inequalities. J Funct Anal. 245(1), 311333.CrossRefGoogle Scholar
Lott, J. & Villani, C. (2009) Ricci curvature for metric-measure spaces via optimal transport. Ann. of Math. 169(3), 903991.CrossRefGoogle Scholar
Maas, J. (2011) Gradient flows of the entropy for finite Markov chains. J. Funct. Anal. 261(8), 22502292.CrossRefGoogle Scholar
Mielke, A. (2011) A gradient structure for reaction-diffusion systems and for energy-drift-diffusion systems. Nonlinearity 24(4), 13291346.CrossRefGoogle Scholar
Mielke, A. (2013) Geodesic convexity of the relative entropy in reversible Markov chains. Calc. Var. Partial Differ. Equ. 48(1-2), 131.CrossRefGoogle Scholar
Otto, F. & Villani, C. (2000) Generalization of an inequality by Talagrand and links with the logarithmic Sobolev inequality. J Funct Anal. 173(2), 361400.CrossRefGoogle Scholar
Sturm, K.-T. (2006) On the geometry of metric measure spaces. I. Acta Math. 196(1), 65131.CrossRefGoogle Scholar
Sturm, K.-T. (2006) On the geometry of metric measure spaces. II. Acta Math. 196(1), 133177.CrossRefGoogle Scholar
Figure 0

Figure 1. Example of $\mathbb{Z}^d$-periodic graph embedded in $\mathbb{R}^d$.

Figure 1

Figure 2. Examples of graphs in $\mathbb{R}^2$ and corresponding unit balls for $f_{\mathrm{hom}}$.